Research Article  Open Access
Shaohua Wu, Yong Hu, Wei Wang, Xinyong Feng, Wanneng Shu, "Application of Global Optimization Methods for Feature Selection and Machine Learning", Mathematical Problems in Engineering, vol. 2013, Article ID 241517, 8 pages, 2013. https://doi.org/10.1155/2013/241517
Application of Global Optimization Methods for Feature Selection and Machine Learning
Abstract
The feature selection process constitutes a commonly encountered problem of global combinatorial optimization. The process reduces the number of features by removing irrelevant and redundant data. This paper proposed a novel immune clonal genetic algorithm based on immune clonal algorithm designed to solve the feature selection problem. The proposed algorithm has more exploration and exploitation abilities due to the clonal selection theory, and each antibody in the search space specifies a subset of the possible features. Experimental results show that the proposed algorithm simplifies the feature selection process effectively and obtains higher classification accuracy than other feature selection algorithms.
1. Introduction
With the explosive development of massive data, it is difficult to analyze and extract high level knowledge from data. The increasing trend of highdimensional data collection and problem representation calls for the use of feature selection in many machine learning tasks [1]. Machine learning is the most commonly used technique to address larger and more complex tasks by analyzing the most relevant information already present in databases [2]. Machine learning is programming computers to optimize a performance criterion using example data or past experience. The selection of relevant features and elimination of irrelevant ones are the key problems in machine learning that have become an open issue in the field of machine learning [3]. Feature selection (FS) is frequently used as a preprocessing step to machine learning that chooses a subset of features from the original set of features forming patterns in a training dataset. In recent years, feature selection has been successfully applied in classification problem, such as data mining applications, information retrieval processing, and pattern classification. FS has recently become an area of intense interests and research.
Feature selection is a preprocessing technique for effective data analysis in the emerging field of data mining which is aimed at choosing a subset of original features so that the feature space is optimally reduced according to the predetermined targets [4]. Feature selection is one of the most important means which can influence the classification accuracy rate and improve the predictive accuracy of algorithms by reducing the dimensionality, removing irrelevant features, and reducing the amount of data needed for the learning process [5, 6]. FS has been an important field of research and development since 1970’s and proven to be effective in removing irrelevant features, reducing the cost of feature measurement and dimensionality, increasing classifier efficiency and classification accuracy rate, and enhancing comprehensibility of learned results.
Both theoretical analysis and empirical evidence show that irrelevant and redundant features affecting the speed and accuracy of learning algorithms and thus should be eliminated as well. An efficient and robust feature selection approach including genetic algorithms (GA) and immune clone algorithm (ICA) can eliminate noisy, irrelevant, and redundant data that have been tried out for feature selection.
In order to find a subset of features that are most relevant to the classification task, this paper makes use of FS technique, together with machine learning knowledge, and proposes a novel optimization algorithm for feature selection called immune clonal genetic feature selection algorithm (ICGFSA). We describe the feature selection for selection of optimal subsets in both empirical and theoretical work in machine learning, and we present a general framework that we use to compare different algorithms. Experimental results show that the proposed algorithm simplifies the feature selection process effectively and either obtains higher classification accuracy or uses fewer features than other feature selection algorithms.
The structure of the rest of the paper is organized as follows. A brief survey is given in Section 2. We study the classification accuracy and formalize it as a mathematical optimization model in Section 3. Section 4 explains the details of the ICGFSA. Several experiments conducted to evaluate the effectiveness of the proposed approach are presented in Section 5. Finally, Section 6 concludes the paper and discusses some future research directions.
2. Related Works
In this section, we focus our discussion on the prior research on feature selection and machine learning. There has been substantial work on feature selection for selection of optimal subsets from the original dataset, which are necessary and sufficient for solving the classification problem.
Extreme learning machine (ELM) is a new learning algorithm for Single Layer Feedforward Neural network (SLFN) whose learning speed is faster than traditional feedforward network learning algorithm like back propagation algorithm while obtaining better generalization performance [7]. Support vector machines (SVM) is a very popular machine learning method used in many applications, such as classification. It finds the maximum margin hyperplane between two classes using the training data and applying an optimization technique [8]. SVM has shown good generalization performance on many classification problems.
Genetic algorithm has been proven to be very effective solution in a great variety of approximately optimum search problems. Recently, Huang and Wang proposed a genetic algorithm to simultaneously optimize the parameters and input feature subset of support vector machine (SVM) without loss of accuracy in classification problems [9]. In [10], a hybrid genetic algorithm is adopted to find a subset of features that are most relevant to the classification task. Two stages of optimization are involved. The inner and outer optimizations cooperate with each other and achieve the high global predictive accuracy as well as the high local search efficiency. Reference [11] proposed and investigated the use of a genetic algorithm method for simultaneously aiming at a higher accuracy level for the software effort estimates.
To further settle the feature selection problems, Mr. Liu et al. proposed an improved feature selection (IFS) method by integrating MSPSO, SVM with score method [12]. Reference [13] proposed a new evolutionary algorithm called Intelligent Dynamic Swarm (IDS), that is, a modified Particle Swarm Optimization. To evaluate the classification accuracy of ITIN and remaining four feature selection algorithms, Naive Bayes, SVM, and ELM classifiers are used for ten UCI repository datasets. Deisy et al. proposed ITIN performs better than the existing above algorithms in terms of number of features [14].
The feature selection process constitutes a commonly encountered problem of global combinatorial optimization. Chuang et al. presented a novel optimization algorithm called catfish binary particle swarm optimization, in which the socalled catfish effect is applied to improve the performance of binary particle swarm optimization [15]. Reference [16] proposed a new information gain and divergencebased feature selection method for statistical machine learningbased text categorization without relying on more complex dependence models. Han et al. study employs feature selection (FS) techniques, such as mutualinformationbased filter and genetic algorithmbased wrapper, to help search for the important sensors in data driven chiller FDD applications, so as to improve FDD performance while saving initial sensor cost.
3. Classification Accuracy and Score
In this section, the proposed feature selection model will be discussed. In general, feature selection problem can be described as follows.
Definition 1. Assume that represents a training dataset with features or attributes and instances, denotes the instances, denotes feature space of constructed from features, which gives an optimal performance for the classifier, and represents the set of classes where instances are tagged.
Definition 2. Assume that represents a value vector of features, where is the value of corresponding to the feature , .
The feature selection approaches are used to generate a feature subset based on the relevance and feature interaction of data samples. The main goal of classification learning is to characterize the relationship between and . Assume that is the subset of alreadyselected features, is the subset of unselected features, and , . Therefore, any optimal feature subset obtained by selection algorithms should preserve the existing relationship between and hidden in the dataset.
The best subset of features is selected by evaluating a number of predefined criteria, such as classification accuracy and score. In order to evaluate the classification accuracy rate, the specific equation on classification accuracy is defined as follows.
Definition 3. Assume that is the set of data items to be classified and is the class of the item . If returns the classification accuracy rates of , then classification accuracy can be formulated as
where represents the number of elements in the collection , .
score is an effective approach which measures the discrimination of two sets of real numbers. The larger the score is, the more this feature is discriminative.
Definition 4. Given training vectors . If the number of the dataset is , then the score of the feature is defined as where , are the average of the th feature of the whole dataset and the th dataset, respectively; is the th feature of the th instance in the th dataset; is the number of datasets. and .
4. Heuristic Feature Selection Algorithm
In this section, we focus our discussion on algorithms that explicitly attempt to select an optimal feature subset. Finding an optimal feature subset is usually difficult, and feature selection for selection of optimal subsets has been shown to be NPhard. Therefore, a number of heuristic algorithms have been used to perform feature selection of training and testing data, such as genetic algorithms, particle swarm optimization, neural networks, and simulated annealing.
Genetic algorithms have been proven as an intelligent optimization algorithm that can find the optimal solution to a problem in the sense of probability in a random manner [17]. However, standard genetic algorithms have some weaknesses, such as premature convergence and poor local search ability. On the other hand, some other heuristic algorithms, such as particle swarm optimization, simulated annealing, and clonal selection algorithm usually have powerful local search ability.
4.1. Basic Idea
In order to obtain the higher classification accuracy rate and higher efficiency of standard genetic algorithms, some hybrid GA for feature selection have been developed by combining the powerful global search ability of GA with some efficient local search heuristic algorithms. In this paper, a novel immune clonal genetic algorithm based on immune clonal algorithm, called ICGFSA, is designed to solve the feature selection problem. Immune clone algorithm is a simulation of the immune system which has the ability to identify the bacteria and designed diversity, and its search target has certain dispersion and independence. ICA can effectively maintain the diversity between populations of antibodies but also accelerate the global convergence speed [18]. The ICGFSA algorithm has more exploration and exploitation abilities due to the clonal selection theory that an antibody has the possibility to clone some similar antibodies in the solution space with each antibody in the search space specifying a subset of the possible features. The experimental results show the superiority of the ICGFSA in terms of the prediction accuracy with smaller subset of features. The overall scheme of the proposed algorithm framework is outlined in Figure 1.
4.2. Encoding
In the ICGFSA algorithm, each antibody in the population represents a candidate solution to the feature selection problem. The algorithm uses the binary coding method that “1” means “selected” and “0” means “unselected” [19]. Therefore, the chromosomes represents by a string of binary digits of zeros and ones and each gene in chromosome corresponds to a feature.
4.3. Affinity Function
We design an affinity function that combines classification accuracy rate with score, which is the evaluation criterion for the feature selection. The affinity function is defined as follows: In which, is equal to the instance of feature when feature is selected, otherwise is equal to 0, .
4.4. Basic Operation
In this section focuses on the three main operations of ICGFSA, including clonal, mutation, and selection. Mutation operation will take the binary mutation operation in standard genetic algorithm [20].
Clonal is essentially the larger antibody affinity for a certain scale replication. Clone size is calculated as follows: In which, and are the number of elements in the set and , respectively. represents the number of antibodies in the population.
The basic idea of selection operation is as follows. Firstly, select the highest affinity antibodies and generate a number of clones for them. Secondly, antibodies that have been selected directly are retained to the next generation [21].
5. Experimental Results and Discussion
5.1. Parameter Setting
In this section, in order to investigate the effectiveness and superiority of the ICGFSA algorithm for classification problems, the same conditions were used to compare with other feature selection methods such as GA and SVM; that is, the parameters of ICGFSA and GA are set as follows: population size is 50, maximum generations is 500, crossover probability is 0.7, and mutation probability is 0.2. For each dataset we have performed 50 simulations, since the test results depend on the population randomly generated by the ICGFSA algorithm.
5.2. Benchmark Datasets
To evaluate the performance of ICGFSA algorithms, the following benchmark datasets are selected for simulation experiments: Liver, WDBC, Soybean, Glass, and Wine. These datasets were obtained from the UCI machine learning repository [22] and most of them are frequently used in a comprehensive testing. They suit for feature selection methods under different conditions. Furthermore, to evaluate the algorithms for real Internet data, we also use malicious PDF file datasets from Virus Total [23]. Table 1 is given some general information about these datasets, such as instances, features, and classes.

5.3. Experimental Results
Figure 2 is the number of selected features with different generations in benchmark datasets using ICGFSA, GA, and SVM, respectively. As seen from Figure 2, it can be observed that the number of selected features is decreased with the number of generations increasing, and ICGFSA can converge to the optimal subsets of required number features since it is the stochastic search algorithms. In the Liver dataset, the number of features selected keeps decreasing, while the number of iterations keeps increasing, until ICGFSA obtained nearly 90% classification accuracy, which indicates that a good feature selection algorithm not only decreases the number of features, but also selects features relevant for improving classification accuracy. It can be observed from Figure 3(b) that when the number of iterations increases beyond certain value (say 300), the performance will no longer be improved. In the Wine dataset, there are several critical points (153, 198, 297, etc.) where the trend has been shifted or changed sharply. In the Soybean and Glass datasets, three algorithms have the best performances and significant improvements in the number of selection features.
(a) Liver dataset
(b) WDBC dataset
(c) Soybean dataset
(d) Glass dataset
(e) Wine dataset
(f) PDF dataset
(a) Liver dataset
(b) WDBC dataset
(c) Soybean dataset
(d) Glass dataset
(e) Wine dataset
(f) PDF dataset
We carried out extensive experiments to verify the ICGFSA algorithm. The running times that find the best subset of required numbers of features and number of selected features in benchmark datasets using ICGFSA, GA, and SVM are recorded in Table 2. It can be observed from Table 2 that ICGFSA algorithm can achieve significant feature reduction that selects only a small portion from the original features which better than the other two algorithms. ICGFSA is more effective than GA and SVM and, moreover, produces improvements of conventional feature selection algorithms over SVM which is known to give the best classification accuracy. From the experimental results we can obviously see that ICGFSA has the least feature number and clonal selection operations can greatly enforce the local searching ability and make the algorithm fast enough to reach its optimum, which indicates ICGFSA has the ability to break through the local optimal solution when applied to largescale feature selection problems. It can be concluded that the ICGFSA is relatively simple and can effectively reduce the computational complexity of implementation process.

Finally, we inspect the classification accuracy for the proposed algorithm. Figure 3 shows the global best classification accuracies with different generations in benchmark datasets using ICGFSA, GA, and SVM, respectively. In the Liver dataset, the global best classification accuracy of ICGFSA is 88.69%. However, the global best classification accuracy of GA and SVM are only 85.12% and 87.54%, respectively. In the WDBC dataset, the global best classification accuracy of ICGFSA is 84.89%. However, the global best classification accuracy of GA and SVM is only 79.36% and 84.72%, respectively. In the Soybean dataset, the global best classification accuracy of ICGFSA and SVM is 84.96% and 84.94%, respectively. However, the global best classification accuracy of GA is only 77.68%. In the Glass dataset, the global best classification accuracy of ICGFSA is 87.96%. However, the global best classification accuracy of GA and SVM is only 84.17% and 86.35%, respectively. In the Wine dataset, the ICGFSA obtained 94.8% classification accuracy before reaching the maximum number of iterations. In the PDF dataset, the global best classification accuracy of ICGFSA and SVM is 94.16% and 93.97%, respectively. However, the global best classification accuracy of GA is only 92.14%. ICGFSA method is consistently more effective than GA and SVN methods on six datasets.
The numerical results and statistical analysis show that the proposed ICGFSA algorithm performs significantly better than the other two algorithms in terms of running time and higher classification accuracy. ICGFSA can reduce the feature vocabulary with best performance in accuracy. It can be concluded that an effective feature selection algorithm is helpful in reducing the computational complexity of analyzing dataset. As long as the chosen features contain enough feature classification information, higher classification accuracy can be achieved.
6. Conclusions
Machine learning is a science of the artificial intelligence. The field’s main objectives of study are computer algorithms that improve their performance through experience. In this paper, the main work in machine learning field is on methods for handling datasets containing large amounts of irrelevant attributes. For the high dimensionality of feature space and the large amounts of irrelevant feature, we propose a new feature selection method base on genetic algorithm and immune clonal algorithm. In the future, ICGFSA algorithm will be applied to more datasets for testing performance.
Acknowledgments
This research work was supported by the Hubei Key Laboratory of Intelligent Wireless Communications (Grant no. IWC2012007) and the Special Fund for Basic Scientific Research of Central Colleges, SouthCentral University for Nationalities (Grant no. CZY11005).
References
 T. Peters, D. W. Bulger, T.H. Loi, J. Y. H. Yang, and D. Ma, “Twostep crossentropy feature selection for microarrayspower through complementarity,” IEEE/ACM Transactions on Computational Biology and Bioinformatics, vol. 8, no. 4, pp. 1148–1151, 2011. View at: Publisher Site  Google Scholar
 W.C. Yeh, “A twostage discrete particle swarm optimization for the problem of multiple multilevel redundancy allocation in series systems,” Expert Systems with Applications, vol. 36, no. 5, pp. 9192–9200, 2009. View at: Publisher Site  Google Scholar
 L.Y. Chuang, H.W. Chang, C.J. Tu, and C.H. Yang, “Improved binary PSO for feature selection using gene expression data,” Computational Biology and Chemistry, vol. 32, no. 1, pp. 29–37, 2008. View at: Publisher Site  Google Scholar  Zentralblatt MATH
 B. Hammer and K. Gersmann, “A note on the universal approximation capability of support vector machines,” Neural Processing Letters, vol. 17, no. 1, pp. 43–53, 2003. View at: Publisher Site  Google Scholar
 L. Yu and H. Liu, “Efficient feature selection via analysis of relevance and redundancy,” Journal of Machine Learning Research, vol. 5, pp. 1205–1224, 2004. View at: Publisher Site  Google Scholar  Zentralblatt MATH  MathSciNet
 G. Qu, S. Hariri, and M. Yousif, “A new dependency and correlation analysis for features,” IEEE Transactions on Knowledge and Data Engineering, vol. 17, no. 9, pp. 1199–1206, 2005. View at: Publisher Site  Google Scholar
 G.B. Huang, Q.Y. Zhu, and C.K. Siew, “Extreme learning machine: theory and applications,” Neurocomputing, vol. 70, no. 1–3, pp. 489–501, 2006. View at: Publisher Site  Google Scholar
 J. G. Dy, C. E. Brodley, A. Kak, L. S. Broderick, and A. M. Aisen, “Unsupervised feature selection applied to contentbased retrieval of lung images,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 25, no. 3, pp. 373–378, 2003. View at: Publisher Site  Google Scholar
 C.L. Huang and C.J. Wang, “A GAbased feature selection and parameters optimizationfor support vector machines,” Expert Systems with Applications, vol. 31, no. 2, pp. 231–240, 2006. View at: Publisher Site  Google Scholar
 J. Huang, Y. Cai, and X. Xu, “A hybrid genetic algorithm for feature selection wrapper based on mutual information,” Pattern Recognition Letters, vol. 28, no. 13, pp. 1825–1844, 2007. View at: Publisher Site  Google Scholar
 A. L. I. Oliveira, P. L. Braga, R. M. F. Lima, and M. L. Cornélio, “GAbased method for feature selection and parameters optimization for machine learning regression applied to software effort estimation,” Information and Software Technology, vol. 52, no. 11, pp. 1155–1166, 2010. View at: Publisher Site  Google Scholar
 Y. Liu, G. Wang, H. Chen, H. Dong, X. Zhu, and S. Wang, “An improved particle swarm optimization for feature selection,” Journal of Bionic Engineering, vol. 8, no. 2, pp. 191–200, 2011. View at: Publisher Site  Google Scholar
 C. Bae, W.C. Yeh, Y. Y. Chung, and S.L. Liu, “Feature selection with Intelligent Dynamic Swarm and rough set,” Expert Systems with Applications, vol. 37, no. 10, pp. 7026–7032, 2010. View at: Publisher Site  Google Scholar
 C. Deisy, S. Baskar, N. Ramraj, J. S. Koori, and P. Jeevanandam, “A novel information theoreticinteract algorithm (ITIN) for feature selection using three machine learning algorithms,” Expert Systems with Applications, vol. 37, no. 12, pp. 7589–7597, 2010. View at: Publisher Site  Google Scholar
 L.Y. Chuang, S.W. Tsai, and C.H. Yang, “Improved binary particle swarm optimization using catfish effect for feature selection,” Expert Systems with Applications, vol. 38, no. 10, pp. 12699–12707, 2011. View at: Publisher Site  Google Scholar
 C. Lee and G. G. Lee, “Information gain and divergencebased feature selection for machine learningbased text categorization,” Information Processing and Management, vol. 42, no. 1, pp. 155–165, 2006. View at: Publisher Site  Google Scholar
 J. Huang, Y. Cai, and X. Xu, “A hybrid genetic algorithm for feature selection wrapper based on mutual information,” Pattern Recognition Letters, vol. 28, no. 13, pp. 1825–1844, 2007. View at: Publisher Site  Google Scholar
 L. N. De Castro and F. J. Von Zuben, “Learning and optimization using the clonal selection principle,” IEEE Transactions on Evolutionary Computation, vol. 6, no. 3, pp. 239–251, 2002. View at: Publisher Site  Google Scholar
 H. Han, B. Gu, T. Wang, and Z. R. Li, “Important sensors for chiller fault detection and diagnosis (FDD) from the perspective of feature selection and machine learning,” International Journal of Refrigeration, vol. 34, no. 2, pp. 586–599, 2011. View at: Publisher Site  Google Scholar
 P. Kumsawat, K. Attakitmongcol, and A. Srikaew, “A new approach for optimization in image watermarking by using genetic algorithms,” IEEE Transactions on Signal Processing, vol. 53, no. 12, pp. 4707–4719, 2005. View at: Publisher Site  Google Scholar  MathSciNet
 R. Meiri and J. Zahavi, “Using simulated annealing to optimize the feature selection problem in marketing applications,” European Journal of Operational Research, vol. 171, no. 3, pp. 842–858, 2006. View at: Publisher Site  Google Scholar  Zentralblatt MATH
 C. L. Blake and C. J. Merz, “UCI repository of machine learning databases,” Department of Information and Computer Science, University of California, Irvine, Calif, USA, 1998, http://www.ics.uci.edu/mlearn/MLRepository.html. View at: Google Scholar
 VirusTotal: http://www.virustotal.com.
Copyright
Copyright © 2013 Shaohua Wu et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.