- About this Journal ·
- Abstracting and Indexing ·
- Aims and Scope ·
- Annual Issues ·
- Article Processing Charges ·
- Articles in Press ·
- Author Guidelines ·
- Bibliographic Information ·
- Citations to this Journal ·
- Contact Information ·
- Editorial Board ·
- Editorial Workflow ·
- Free eTOC Alerts ·
- Publication Ethics ·
- Reviewers Acknowledgment ·
- Submit a Manuscript ·
- Subscription Information ·
- Table of Contents
Journal of Applied Mathematics
Volume 2013 (2013), Article ID 841780, 15 pages
A Hybrid Multiobjective Differential Evolution Algorithm and Its Application to the Optimization of Grinding and Classification
1School of Information Science and Engineering, Central South University, Changsha 410083, China
2Department of Mathematics & Statistics, Curtin University, Perth, WA 6845, Australia
Received 19 July 2013; Accepted 5 September 2013
Academic Editor: Dewei Li
Copyright © 2013 Yalin Wang et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
The grinding-classification is the prerequisite process for full recovery of the nonrenewable minerals with both production quality and quantity objectives concerned. Its natural formulation is a constrained multiobjective optimization problem of complex expression since the process is composed of one grinding machine and two classification machines. In this paper, a hybrid differential evolution (DE) algorithm with multi-population is proposed. Some infeasible solutions with better performance are allowed to be saved, and they participate randomly in the evolution. In order to exploit the meaningful infeasible solutions, a functionally partitioned multi-population mechanism is designed to find an optimal solution from all possible directions. Meanwhile, a simplex method for local search is inserted into the evolution process to enhance the searching strategy in the optimization process. Simulation results from the test of some benchmark problems indicate that the proposed algorithm tends to converge quickly and effectively to the Pareto frontier with better distribution. Finally, the proposed algorithm is applied to solve a multiobjective optimization model of a grinding and classification process. Based on the technique for order performance by similarity to ideal solution (TOPSIS), the satisfactory solution is obtained by using a decision-making method for multiple attributes.
Grinding-classification is an important prerequisite process for most mineral processing plants. The grinding process reduces the particle size of raw ores and is usually followed by classification to separate them into different sizes. Grinding-classification operation is required to produce pulp with suitable concentration and fineness for flotation. The pulp quality will directly influence the subsequent flotation efficiency and recovery of valuable metals from tailings. In order to improve economic efficiency and energy consumption, the process optimization objectives include product quality and output yields. Under certain mineral source conditions, the objectives are decided by a series of operation variables such as the solid flow of feed ore to ball mill, the steel ball filling rate, and the flow rates of water added to the first and the second classifier recycles. To solve the optimization model of products’ output and quality in the grinding-classification process is of great significance to improve the technical and economic specifications, and it has been a continuous endeavor of the scientists and engineers [1–3].
Grinding-classification is an energy-intensive process influenced by many interacting factors with mutual restraints. The goals of grinding-classification optimization problem are decided by multiple constrained input control variables of nonlinear relationships. So, the optimization model of grinding-classification operation is a complex constrained multiobjective optimization problem (CMOP). Generally, constrained multiobjective problems are so difficult to be solved that the constraint handling techniques and multiobjective optimization methods need to be combined for optimization.
Multiobjective optimization problems (MOPs), in the case of traditional optimization methods, are often handled by aggregating multiple objectives into a single scalar objective through weighting factors. MOPs have a set of equally good (nondominating) solutions instead of a single one, called a Pareto optimum which was introduced by Edgeworth in 1881  and later generalized by Pareto in 1896 . The practical MOPs are often implicated in series of equations, functions, or procedures with complicated constraints. Therefore, the evolutionary algorithms are attractive approaches for low requirements on mathematical expression . Since the mid 1980s, there has been a growing interest in solving MOPs using evolutionary approaches [7–10]. One of the most successful evolutionary algorithms for the optimization of continuous space functions is the differential evolution (DE) . DE is simple and efficiently converges to the global optimum in most cases [12, 13]. Its efficiency has been proven  in many application fields such as pattern recognition  and mechanical engineering .
There have been many improvements for DE to solve MOPs. Abbass  firstly provided a Pareto DE (PDE) algorithm for MOPs in which DE was employed to create new solutions, and only the nondominated solutions were kept as the basis for the next generation. Madavan  developed a Pareto differential evolution approach (PDEA) in which new solutions were created by DE and kept in an auxiliary population. Xue et al.  introduced multiobjective differential evolution (MODE) and used Pareto-based ranking assignment and crowding distance metric, but in a different manner from PDEA. Robic and Filipi , also adopting Pareto-based ranking assignment and crowding distance metric, developed a DE for multiobjective optimization (DEMO) with a different population update strategy and achieved good results. Huang et al.  extended the self-adaptive DE (SADE) to solve MOPs by a so called multiobjective self-adaptive DE (MOSADE). They further extended MOSADE by using objectivewise learning strategies . Adeyemo and Otieno  provided multiobjective differential evolution algorithm (MDEA). In MDEA, a new solution was generated by DE variant and compared with target solution. If it dominates the target solution, then it was added to the new population; otherwise, a target solution was added.
On the other hand, single-objective constrained optimization problems have been studied intensively in the past years [24–28]. Different constraint handling techniques have been proposed to solve constrained optimization problems. Michalewicz and Schoenauer  divided constraints handling methods used in evolutionary algorithms into four categories: preserving feasibility of solutions, penalty functions, separating the feasible and infeasible solutions, and hybrid methods. The differences among these methods are how to deal with the infeasible individuals throughout the search phases. Currently, the penalty function method is most widely used, and this algorithm strongly depends on the choice of the penalty parameter.
Although the multiobjective optimization and the constraint handling problem have received lots of contribution, respectively, the CMOPs are still difficult to be solved in practice. Coello and Christiansen  proposed a simple approach to solve CMOPs by ignoring any solution that violates any of the assigned constraints. Deb et al.  proposed a constrained multiobjective algorithm based on the concept of constrained domination, which is also known as superiority of the feasible solution. Woldesenbet et al.  introduced a constraint handling technique based on adaptive penalty functions and distance measures by extending the corresponding version for the single-objective constrained optimization.
In the MOP of grinding and classification process, the definitions of Pareto solutions, Pareto frontier, and Pareto dominance are in consistency with the classic definitions. Clearly, the Pareto frontier is a mapping of the Pareto-optimal solutions to the objective space. In the minimization sense, general constrained MOPs can be formulated as follows where is the objective vector, is a parameter vector, is the inequality constraint, and is the equality constraint. and are, respectively, the lower and upper bounds of the decision variable .
In this paper, based on the specific industrial background of continuous bauxite grinding-classification operation, a new hybrid DE algorithm is proposed to solve complex constrained multiobjective optimization problems. Firstly, a hybrid DE algorithm for MOPs with simplex method (SM-DEMO) is designed to overcome the problems of global performance degradation and being trapped in local optimum. Then, for the MOPs with complicated constraints, the proposed algorithm is formed by combining SM-DEMO and functional partitioned multi-population. In this method, the construction of penalty functions is not required, and the meaningful infeasible solutions are fully utilized.
The remainder of the paper is structured as follows. Section 2 describes the SM-DEMO algorithm for unconstrained cases. The proposed algorithm of multipopulation for constrained MOPs is given in Section 3 with verification of performance by benchmark testing results. Section 4 describes the model of products’ output and quality in the grinding-classification process in detail and the application of the proposed algorithm in the optimization model. Finally, the conclusions based on the present study are drawn in Section 5.
2. SM-DEMO Algorithm for Unconstrained MOPs
In order to efficiently solve multiobjective optimization problem and find the approximately complete and nearoptimal Pareto frontier, we proposed a hybrid DE algorithm for unconstrained multiobjective optimization with simplex method.
The differential evolution, with initialization, crossover, and selection as in usual genetic algorithms, uses a perturbation of two members as the mutation operator to produce a new individual. The mutation operator of the DE algorithm is described as follows.
Considering each target individual , in the th generation of size , a mutant individual is defined by where indexes , , and represent mutually different integers that are different from and that are randomly generated over , and is the scaling factor.
The simplex method, proposed by Spendley, Hext, and Himsworth and later refined by Nelder and Mead (NM) , is a derivative-free line-search method that is particularly designed for traditional unconstrained minimization scenarios. Clearly, NM method can be deemed as a direct line-search method of the steepest descent kind. The ingredients of the replacement process consist of four basic operations: reflection, expansion, contraction, and shrinkage. Through these operations, the simplex can improve itself and approximate to a local optimum point sequentially. Furthermore, the simplex can vary its shape, size, and orientation to adapt itself to the local contour of the objective function.
2.1. Main Strategy of SM-DEMO
The SM-DEMO algorithm is improved by the following three points compared with DE.
2.1.1. Modified Selection Operation
After traditional DE evolution, the individual may violate the boundary constraints and . is replaced by new individual being adjusted as follows:
The new population is combined with the existing parent population to form a new set of bigger size than . A nondominated ranking of is performed, and the best individuals are selected. This approach allows a global nondomination checking between both the parent and the new generation rather than only in the new generation as is done in other approaches, whereas it requires additional computational cost in sorting the combined.
2.1.2. Nondominated Ranking Based on Euclidean Distance
The solutions within each nondominated frontier that reside in the less crowded region in the frontier are assigned a higher rank, as the NSGA-II algorithm  developed by Deb et al. indicated. The crowding distance of the th solution in its frontier (marked with solid circles) is the average side length of the cuboids (shown with a dashed box in Figure 1(a)). The crowding-distance computation requires sorting the population according to each objective function value in ascending order of magnitude. As shown in Figure 1, and are two solutions near in the same rank, and is the crowding distance of the th solution, traditionally calculated as follows: where , are the objective vectors. For each objective function, the boundary solutions (solutions with the smallest and the largest function values) are assigned an infinite distance value.
A crowding-distance metric is used to estimate the density of solutions surrounding a particular solution in the population and is obtained from the average distance of the two solutions on either side of the solution along each of the objectives. As shown in Figure 1, , , are the individuals of the generation on the same frontier, and we easily know that the density in Figure 1(a) is better than that in Figure 1(b). If we use (4) to calculate the crowding distance of , we only know that in Figure 1(a) it is better than in Figure 1(b); the crowding distance of in Figures 1(a) and 1(c) is equal, which is against the knowledge.
To distinguish the mentioned situations, we propose an improved crowding-distance metric based on Euclidean distance. is the center point of the line , is the objective vector, and the crowding distance is defined as follows:
The crowded-comparison operator guides the selection process at the various stages of the algorithm toward a uniformly distributed Pareto-optimal frontier. To carry out the comparison, we assume that every individual in the population has two attributes: nondomination rank and crowding distance . Then, a partial order is defined as . If , that is, between two solutions with different nondomination ranks, we prefer the solution with the lower (better) rank, namely, . Otherwise, if both solutions belong to the same frontier, that is, , then we prefer the solution that is located in a lesser crowded region, that is, .
2.1.3. Simplex Method for Local Search
The simplex method for local search is mixed in the evolution process to enhance the searching strategy in the optimization process. The goal of integrating NM simplex method  and DE is to enrich the population diversity and avoid being trapped in local minimum. We apply simplex method operator to the present population when the number of iterations is greater than a particular value (like ). The individuals that achieved the single extreme value in each objective function are marked as the initial vertex points of simplex method. The present population is updated according to simplex method until the terminal conditions are satisfied.
The computation steps of the algorithm are included in Section 3.2.
2.2. Evaluation Criteria
Unlike the single-objective optimization, it is more complicated for solution quality evaluation in the case of multiobjective optimization. Many of the suggested methods can be summarized in two types. One is to evaluate the convergence degree by computing the proximity between the solution frontier and the actual Pareto frontier. The other is to evaluate the distribution degree of the solutions in objective space by computing the distances among the individuals. Here, we choose both methods to evaluate the performance of the SM-DEMO algorithm.
(1) Convergence Evaluation. Deb et al.  proposed this method in 2002. It is described as follows: where is the extent of convergence to a known of Pareto-optimal set, is the obtained nondomination Pareto frontier, is the real nondomination Pareto frontier, is the Euclidean distance of with , and is the number of obtained solutions.
(2) Distribution Degree Evaluation. The nonuniformity in the distribution is measured by as follows: where is the Euclidean distance among consecutive solutions in the obtained nondominated set of solutions and parameter is the average distance.
2.3. Experimental Studies
Four well-known benchmark test functions  are used here to compare the performance of SM-DEMO with NSGA-II, DEMO/Parent. These four problems are called ZDT2, ZDT3, ZDT4, and ZDT6; each has two objective functions. We describe them in Table 1.
The simulation is carried out under the environment of Intel Pentium 4, CPU 3.06 GHz, 512 MB memory, Windows XP Professional, Matlab7.1. Initialization parameters are set as follows: population size , scaling factor , cross rate , maximum evolution generation , and number of SM evolution iterations .
All of the three algorithms are real coded, with equal population size and equal maximum evolution generation. Each algorithm independently runs 20 times for each test function. Because we cannot get the real Pareto-optimal set, we will take 60 Pareto-optimal solutions obtained by the three algorithms as a true Pareto-optimal solution set.
We evaluated the algorithms based on the two performance indexes and . Table 2 shows the mean and variance of and using three algorithms: SM-DEMO, NSGA-II, and DEMO/Parent. We can learn from Table 2, for the ZDT2 function, that all of the three algorithms have a good performance, while the SM-DEMO is slightly better than the other two algorithms. In terms of convergences, for ZDT3, ZDT4 and ZDT6, which are more complex than ZDT2, SM-DEMO is significantly better than DEMO/Parent and NSGA-II.
Figure 2 shows a random running of SM-DEMO algorithm. It is clear that SM-DEMO algorithm can produce a good approximation and a uniform distribution.
3. Proposed Hybrid Algorithm for CMOP
The space of constrained multiobjective optimization problem can be divided into the feasible solution space and the infeasible solution space, as shown in Figure 3, where is the search space, is the feasible solution space, and is the infeasible solution space. is the feasible solution, and is the infeasible solution. Assume that is the global optimal solution and is the closest one to . If the infeasible population is not excluded by the evolution algorithm, it is permitted to explore boundary regions from new directions, where the optimum is likely to be found.
3.1. General Idea of the Proposed Algorithm
Researchers have gradually realized the merit of infeasible solutions in searching for the global optimum in the feasible region. Some infeasible solutions with better performance are allowed to be saved. Farmani et al.  formulated a method to ensure that infeasible solutions with a slight violation become feasible in evolution. Based on the constraints processing approach of multiobjective optimization problems, the proposed hybrid DE algorithm avoids constructing penalty function and deleting meaningful infeasible solutions directly.
Here, the proposed algorithm will produce multiple groups of functional partitions, which include an evolutionary population of size , an intermediate population to save feasible individuals, an intermediate population to save infeasible individuals, a population to save the optimal feasible solution found in the search process and a population to save the optimal infeasible solution. The relationship of multi-population is shown in Figure 4.
With the description of (1), equality constraints are always transformed into inequality constraints as , where and is a positive tolerance value. To evaluate the infeasible solution, the degree of constraint violation of individual on the constraint is calculated as follows:
The final constraint violation of each individual in the population can be obtained by calculating the mean of the normalized constraint violations.
In order to take advantage of the infeasible solutions with better performance, we proposed the following adaptive differential mutation operator to generate individual variation learning from the mutation operators in DE/rand-to-best/1/bin, according to rules defined by Price et al. . Considering each individual vector , a mutant individual is defined by where and represent different integers and also different from , randomly generated over ; is the scaling factor; is randomly generated from , is randomly generated from ; and is the mutation factor as follows: where is the initial value of the variability factor, is a small constant, to ensure that the fractional is meaningful, and is defined as follows:
3.2. Framework of the Proposed Algorithm
The proposed algorithm is described as follows.
Step 1 (initialization). Generate the population , , and of size , , and . Set the value of (crossover probability), (the number of function evaluations), (the iterative number of evolution by simplex method), (the current generation number), and positive control parameter for scaling the difference vectors , . Randomly generate the parent population from the decision space. Set the , and , and let the intermediate populations and be empty.
Step 2 (DE reproduction). By (3) and (9) for mutation, crossover, and selection, an offspring is created. Judge the constraints of all individuals in . In accordance with (8), we first calculate constraint violation degree of all of the individuals. If , the solution is feasible and preserved to the intermediate set ; if , the solution is infeasible and preserved to the intermediate set .
Step 3 (simplex method). Apply NM simplex method operator to the present population if . Update the present population when the number of iterations exceeds maximum iterations.
Step 4 ( construction). Rank chromosomes in based on (5), and generate the elitist population (the size is ) from the ranked population .
Step 5 ( construction). Add the chromosomes in with slight constraint violations to the .
Step 6 (mixing the population). Combine with the existing parent population to form a new set . Remove the duplicate individuals in and the existing parent population.
Step 7 (evolution). Randomly choose chromosomes from , , and . Use the adaptive differential mutation and uniform discrete crossover to obtain the offspring population .
Step 8 (termination). If the stopping criterion is met, stop and output the best solution; else, go to Step 2.
3.3. Experimental Study
For CTP problem, there are the six parameters , , , , , and that must be chosen in a way so that a portion of the unconstrained Pareto-optimal region is infeasible. Each constraint is an implicit non-linear function of decision variables. Thus, it may be difficult to find a number of solutions on a non-linear constraint boundary. We take two sets of values for six parameters in CTP problem, which are determined as CTP1: , , , , and ; CTP2: , , , , , and . The Pareto frontiers, the feasible solution spaces, and the infeasible solution spaces are shown in Figure 5.
The parameters are initialized as follows. The size of population is , size of is , size of is , scaling factors and are randomly generated within , cross rate is , maximum evolution generation is , and number of SM evolution iterations is . All of the proposed algorithms and CNSGA-II are real coded with equal population size and equal maximum evolution generation. Each algorithm runs 20 times independently for each test function.
Figure 6 shows the result of a random running of the proposed algorithm and NSGA-II, the smooth curve “—” represents the Pareto frontier, and “◆” stands for the solution achieved by the proposed algorithm or NSGA-II.
It is obvious that the proposed algorithm returns a better approximation to the true Pareto-optimal frontie and a distribution of higher uniformity. We also evaluated the algorithms based on the two criterions and , as shown in Table 4. It can be observed from the data in Table 4 that the proposed algorithm performs significantly better than the classical CNSGA-II algorithm in convergence and distribution uniformity. The simulation results show that this algorithm can accurately converge to global Pareto solutions and can maintain diversity of population.
4. Optimization of Grinding and Classification Process
4.1. Bauxite Grinding and Classification Process
The grinding and classification process is the key preparation for the bauxite mineral processing. Here, we consider a bauxite grinding process in a certain mineral company with single grinding and two-stage classification, as shown in Figure 7.
The process consists in a grinding ball mill and two spiral classifiers. First classifier recycle will be put back to the ball mill for regrinding, and the first-stage overflow will be put into second spiral classifier after being mixed with water; the second classifier recycle will be prepared for Bayer production as the rough concentrate, and the second-stage overflow will be sent to the next flotation process. The production objectives are composed of the production yields, technically represented by the solid flow of feed ore since the process is nonstorable, and the mineral processing quality, represented by percentage of the small-size fractions of mineral particles in the second-stage overflows.
4.2. Predictive Model of the Grinding and Classification Process
Here, we establish the mathematical predictive model of each unit process in the bauxite grinding and classification process. The notations of the indexes, decision variables, and parameters are listed in Table 5. These notations will be used for the model of the grinding and classification process.
4.2.1. Ball Mill Circuit Model
Here, is the particle percentage of th size fraction in the ball mill overflow, is the particle percentage of th size fraction in feed ore, rate of the first classifier recycle is known, and is the efficiency of the first spiral classifier. According to a technical report of field investigation and study, we have that where is the mean residence time of minerals, is the internal concentration in ball mill, and where () is the solid flow of feed ore, is the water addition of the first classifier recycle, and is the classifier water addition. is the breakage distribution function; is the breakage rate function, and it satisfied the following equation: where is the particle with the th size, it is a unit, when per millimeter is a unit, , , and , , , and are four key parameters to control the breakage rate function.
In a concrete grinding and classification process, the ball mill size is fixed, and the speed of ball mill is constant. Through data acquisition and testing of grinding and classification steady-state loop, the regression model between , , , and condition variables, size fraction distribution can be established. The input variables are ball filling rate , solid flow of feed ore , water addition of the first classifier recycle , and parameters of feed ore size fraction distribution , . The regression model is
The value of can be obtained by the experimental data regression, , can be obtained from feed ore size fraction distribution, and is the cumulative particle percentage less than the th size fraction in feed ore, and it is represented as follows:
4.2.2. Spiral Classifier Model
is the particle percentage of the th size fraction in the first classifier overflow, and is the particle percentage of the th size fraction in ball mill overflow. The spiral classifier model is as follows: where is the efficiency of the first spiral classifier and the mechanism formula of is shown as follows: where is the particle with the th size, and represent maximum and minimum particle sizes, is the particle size fraction after correction separation, is separation accuracy, and is intermix index.
Through data acquisition and testing of grinding and classification steady-state loop, the regression model between classification parameters and condition variables, size fraction distribution can be established. The input variables include the solid flow of feed ore , the classifier water addition , and the parameters of ball mill overflow size fraction distribution , . The regression model is shown as follows: where the value of can be obtained by data regression.
The first-stage overflow calculation formula is as follows:
Similarly, we can get the second spiral classifier model as follows: where is the particle percentage of the th size fraction in the second classifier overflow, is the particle percentage of the th size fraction in the first classifier overflow, and is the efficiency of the second spiral classifier. The spiral classifier model is as follows: where, , , , and are key parameters to the efficiency of the second spiral classifier. Through data acquisition and testing of grinding and classification steady-state loop, the regression model between classification parameters and condition variables, size fraction distribution can be established. The input variables include solid flow of feed ore and parameters of the first-stage classifier overflow size fraction distribution , , which are solved by similar equation to (20). The regression model is shown as follows: where the value of can be obtained by experimental data regression.
4.3. Optimization Model of Grinding and Classification Process
Two objective functions in the process are identified: one is to maximize output , and the other is to maximize the small-size fractions (less than 0.075 mm fractions) in the second-stage overflow . It is also necessary to ensure that the grinding product meets all of other technical requirements and the least disturbance in the following flotation circuit. As the constraints, the feed load of the grinding circuit , the steel ball filling rate , the first and the second overflows and , and the particle percentage of fine size fraction in the first and the second classifier overflows and should be within the user specified bounds.
The operation variables are the solid flow of feed ore , water addition of the first classifier recycle , ball filling rate , and water addition of the second classifier . Based on all of the above, grinding and classification process multiobjective optimization model is as follows:
With the practical process data from a grinding circuit of a mineral plant, the simulation of this hybrid intelligent method adopted the same parameters on the variation in fresh slurry feed velocity, density, particle size distribution, and cyclone feed operating configurations.
The comparison of production data and optimization results in Table 6 is shown in Figure 8, where “◆” represents the proposed algorithm optimization results and “○” represents the original data collected from the field without optimization of raw data. According to the objectives, the data point closer to the upper right edge is more beneficial. Obviously, the proposed optimization result is far better than the original data, indicating the effectiveness of the optimization approach.
4.4. TOPSIS Method for Solution Selection
The resolution of a multiobjective optimization problem does not end when the Pareto-optimal set is found. In practical operational problems, a single solution must be selected. TOPSIS  is a useful technique in dealing with multiattribute or multicriteria decision-making (MADM/MCDM) problems in the real world. The standard TOPSIS method attempts to choose alternatives that simultaneously have the shortest distance from the positive-ideal solution and the farthest distance from the negative-ideal solution. According to the TOPSIS method, the relative closeness coefficient is calculated, and the best solution in Table 6 is the solution number 10 as and . The corresponding decision variables are , , , and .
Promoted by the requirements of engineering optimization in complex practical processes of grinding and classification, we proposed a hybrid multiobjective differential evolution algorithm with a few beneficial features integrated. Firstly, an archiving mechanism for infeasible solutions is established with functional partitioned multi-population, which aims to direct the population to approach or land in the feasible region from different directions during evolution. Secondly, we propose an infeasible constraint violations function to select infeasible population with better performance, so that they are allowed to be saved and to participate in the subsequent evolution. Thirdly, a nondominating ranking strategy is designed to improve the crowding-distance sorting and return uniform distribution of Pareto solutions. Finally, the simplex method is inserted in the differential evolution process to purposefully enrich the diversity without excessive computation cost. The advantage of the proposed algorithm is the exemption from constructing penalty function and the preservation of meaningful infeasible solutions directly. Simulation results on benchmarks indicate that the proposed algorithm can converge quickly and effectively to the true Pareto frontier with better distribution.
Based on the investigated information about grinding circuit process, we established a multiobjective optimal model with equations from mechanism knowledge, parameters recognized by data regression, and constraints of technical requirements. The nonlinear multiobjective optimization model is too complicated to be solved by traditional gradient-based algorithms. The proposed hybrid differential evolution algorithm is applied and tested to achieve a Pareto solution set. It is proven to be valuable for operation decision making in the industrial process and showed superiority to the operation carried out in the production. In fact, many operating parameters in complex processes are highly coupled and conflicting with each other. The optimal operation of the entire production process is very difficult to obtain by manual calculation; let alone the fluctuation situation of process conditions. The application case indicates that the proposed method has good performance and is helpful to inspire further research on evolutionary methods for engineering optimization.
Conflict of Interests
The authors declare that there is no conflict of interests regarding the publication of this paper.
This work is supported by the National Natural Science Foundation of China (61374156, 61273187, and 61134006), the National Key Technology Research and Development Program of the Ministry of Science and Technology of China (2012BAK09B04), the Science Fund for Creative Research Groups of the National Natural Science Foundation of China (61321003), the Fund for Doctor Station of the Ministry of Education (20110162130011 and 20100162120019) and the Hunan Province Science and Technology Plan Project (2012CK4018).
- K. Mitra and R. Gopinath, “Multiobjective optimization of an industrial grinding operation using elitist nondominated sorting genetic algorithm,” Chemical Engineering Science, vol. 59, no. 2, pp. 385–396, 2004.
- G. Yu, T. Y. Chai, and X. C. Luo, “Multiobjective production planning optimization using hybrid evolutionary algorithms for mineral processing,” IEEE Transactions on Evolutionary Computation, vol. 15, no. 4, pp. 487–514, 2011.
- T. Y. Ma, W. H. Gui, Y. L. Wang, and C. H. Yang, “Dynamic optimization control for grinding and classification process,” Control and Decision, vol. 27, no. 2, pp. 286–290, 2012 (Chinese).
- F. Y. Edgeworth, Mathematical Physics: An Essay on the Application of Mathematics to the Moral Sciences, C. Kegan Paul & Company, London, UK, 1881.
- V. Pareto, Cours d'Economie Politique, F. Rouge, Lausanne, Switzerland, 1896.
- X. F. Chen, W. H. Gui, Y. L. Wang, and L. Cen, “Multi-step optimal control of complex process: a genetic programming strategy and its application,” Engineering Applications of Artificial Intelligence, vol. 17, no. 5, pp. 491–500, 2004.
- C. A. C. Coello, “Evolutionary multi-objective optimization: a historical view of the field,” IEEE Computational Intelligence Magazine, vol. 1, no. 1, pp. 28–36, 2006.
- K. Deb, A. Pratap, S. Agarwal, and T. Meyarivan, “A fast and elitist multiobjective genetic algorithm: NSGA-II,” IEEE Transactions on Evolutionary Computation, vol. 6, no. 2, pp. 182–197, 2002.
- C. M. Fonesca and P. J. Fleming, “Genetic algorithms for multi objective optimization: formulation, discussion and generalization,” in Proceedings of the 5th International Conference on Genetic Algorithms, pp. 415–423, Morgan Kaufmann, San Mateo, Calif, USA, 1993.
- S. Sengupta, S. Das, M. Nasir, A. V. Vasilakos, and W. Pedrycz, “Energy-efficient differentiated coverage of dynamic objects using an improved evolutionary multi-objective optimization algorithm with fuzzy-dominance,” in Proceedings of IEEE Congress on Evolutionary Computation, pp. 1–8, IEEE Press, Brisbane, Australia, 2012.
- K. V. Price, R. M. Storn, and J. A. Lampinen, Differential Evolution a Practical Approach to Global Optimization, Springer, Berlin, Germany, 2005.
- W. Y. Gong and Z. H. Cai, “An improved multiobjective differential evolution based on Pareto-adaptive ε-dominance and orthogonal design,” European Journal of Operational Research, vol. 198, no. 2, pp. 576–601, 2009.
- X. H. Zeng, W. K. Wong, and S. Y. Leung, “An operator allocation optimization model for balancing control of the hybrid assembly lines using Pareto utility discrete differential evolution algorithm,” Computers and Operations Research, vol. 39, no. 5, pp. 1145–1159, 2012.
- E. Zitzler and L. Thiele, “Multiobjective evolutionary algorithms: a comparative case study and the strength Pareto approach,” IEEE Transactions on Evolutionary Computation, vol. 3, no. 4, pp. 257–271, 1999.
- A. Al-Ani, A. Alsukker, and R. N. Khushaba, “Feature subset selection using differential evolution and a wheel based search strategy,” Swarm and Evolutionary Computation, vol. 9, pp. 15–26, 2013.
- G. Y. Li and M. G. Liu, “The summary of differential evolution algorithm and its improvements,” in Proceedings of the 3rd International Conference on Advanced Computer Theory and Engineering (ICACTE '10), pp. V3153–V3156, Chengdu, China, August 2010.
- H. A. Abbass, “The self-adaptive Pareto differential evolution algorithm,” in Proceedings of the IEEE Congress on Evolutionary Computation, pp. 831–836, IEEE Press, Honolulu, Hawaii, USA, May 2002.
- N. K. Madavan, “Multiobjective optimization using a Pareto differential evolution approach,” in Proceedings of the IEEE Congress on Evolutionary Computation, pp. 1145–1150, IEEE Press, Honolulu, Hawaii, USA, May 2002.
- F. Xue, A. C. Sanderson, and R. J. Graves, “Pareto-based multi-objective differential evolution,” in Proceedings of the IEEE Congress on Evolutionary Computation,, pp. 862–869, IEEE Press, Piscataway, NJ, USA, December 2003.
- T. Robic and B. Filipi, “DEMO: differential evolution for multiobjective optimization,” in Evolutionary Multi-Criterion Optimization, pp. 520–533, Springer, Berlin, Germany, 2005.
- V. L. Huang, A. K. Qin, P. N. Suganthan, and M. F. Tasgetiren, “Multi-objective optimization based on self-adaptive differential evolution algorithm,” in Proceedings of the IEEE Congress on Evolutionary Computation (CEC '07), pp. 3601–3608, IEEE Press, Singapore, September 2007.
- V. L. Huang, S. Z. Zhao, R. Mallipeddi, and P. N. Suganthan, “Multi-objective optimization using self-adaptive differential evolution algorithm,” in Proceedings of the IEEE Congress on Evolutionary Computation (CEC '09), pp. 190–194, IEEE Press, Trondheim, Norway, May 2009.
- J. A. Adeyemo and F. A. O. Otieno, “Multi-objective differential evolution algorithm for solving engineering problems,” Journal of Applied Sciences, vol. 9, no. 20, pp. 3652–3661, 2009.
- Y. Wang and Z. Cai, “Combining multiobjective optimization with differential evolution to solve constrained optimization problems,” IEEE Transactions on Evolutionary Computation, vol. 16, no. 1, pp. 117–134, 2012.
- A. Homaifar, C. X. Qi, and S. H. Lai, “Constrained optimization via genetic algorithms,” Simulation, vol. 62, no. 4, pp. 242–253, 1994.
- J. A. Joines and C. R. Houck, “On the use of non-stationary penalty functions to solve nonlinear constrained optimization problems with GA's,” in Proceedings of the 1st IEEE Conference on Evolutionary Computation, pp. 579–584, Orlando, Fla, USA, June 1994.
- F. Jiménez and J. L. Verdegay, “Evolutionary techniques for constrained optimization problems,” in Proceedings of the 7th European Congress on Intelligent Techniques and Soft Computing (EUFIT '99), Aachen, Germany, 1999.
- C. Sun, J. Zeng, and J. Pan, “An improved vector particle swarm optimization for constrained optimization problems,” Information Sciences, vol. 181, no. 6, pp. 1153–1163, 2011.
- Z. Michalewicz and M. Schoenauer, “Evolutionary algorithms for constrained parameter optimization problems,” Evolutionary Computation, vol. 4, no. 1, pp. 1–32, 1996.
- C. A. C. Coello and A. D. Christiansen, “Moses: a multiobjective optimization tool for engineering design,” Engineering Optimization, vol. 31, no. 1–3, pp. 337–368, 1999.
- Y. G. Woldesenbet, G. G. Yen, and B. G. Tessema, “Constraint handling in multiobjective evolutionary optimization,” IEEE Transactions on Evolutionary Computation, vol. 13, no. 3, pp. 514–525, 2009.
- J. A. Nelder and R. Mead, “A simplex method for function minimization,” The Computer Journal, vol. 7, no. 4, pp. 308–313, 1965.
- D. A. van Veldhuizen and G. B. Lamont, “Multiobjective evolutionary algorithms: analyzing the state-of-the-art,” Evolutionary Computation, vol. 8, no. 2, pp. 125–147, 2000.
- R. Farmani, D. A. Savic, and G. A. Walters, “Evolutionary multi-objective optimization in water distribution network design,” Engineering Optimization, vol. 37, no. 2, pp. 167–183, 2005.
- K. Deb and T. Goel, Controlled Elitist Non-Dominated Sorting Genetic Algorithms for Better Convergence, Springer, Berlin, Germany, 2001.
- Y. J. Lai, T. Y. Liu, and C. L. Hwang, “Topsis for MODM,” European Journal of Operational Research, vol. 76, no. 3, pp. 486–500, 1994.