Research Article  Open Access
Zilin Zeng, Hongjun Zhang, Rui Zhang, Youliang Zhang, "A Hybrid Feature Selection Method Based on Rough Conditional Mutual Information and Naive Bayesian Classifier", International Scholarly Research Notices, vol. 2014, Article ID 382738, 11 pages, 2014. https://doi.org/10.1155/2014/382738
A Hybrid Feature Selection Method Based on Rough Conditional Mutual Information and Naive Bayesian Classifier
Abstract
We introduced a novel hybrid feature selection method based on rough conditional mutual information and Naive Bayesian classifier. Conditional mutual information is an important metric in feature selection, but it is hard to compute. We introduce a new measure called rough conditional mutual information which is based on rough sets; it is shown that the new measure can substitute Shannon’s conditional mutual information. Thus rough conditional mutual information can also be used to filter the irrelevant and redundant features. Subsequently, to reduce the feature and improve classification accuracy, a wrapper approach based on naive Bayesian classifier is used to search the optimal feature subset in the space of a candidate feature subset which is selected by filter model. Finally, the proposed algorithms are tested on several UCI datasets compared with other classical feature selection methods. The results show that our approach obtains not only high classification accuracy, but also the least number of selected features.
1. Introduction
With increase of data dimensionality in many domains such as bioinformatics, text categorization, and image recognition, feature selection has become one of the most important data mining preprocessing methods. The aim of feature selection is to find a minimal feature subset of the original datasets that is the most characterizing. Since feature selection can bring lots of advantages, such as avoiding overfitting, facilitating data visualization, reducing storage requirements, and reducing training times, it has attracted considerable attention in various areas [1].
In the past two decades, different techniques are proposed to address these challenging tasks. Dash and Liu [2] point out that there are four basic steps in a typical feature selection method, that is, subset generation, subset evaluation, stopping criterion, and validation. Most studies focus on the two major steps of feature selection: subset generation and subset evaluation. According to subset evaluation function, feature selection methods can be divided into two categories: filter method and wrapper method [3]. Filter methods are independent of predictor, whereas wrapper methods utilize their predictive power as the evaluation function. The merits of filter methods are high computation efficiency and its generality. However, the result of filter method is not always satisfactory. This is because the filter model separates feature selection from the classifier learning and selects the feature subsets that are independent from the learning algorithm. On the other hand, wrapper methods guarantee good results, but they are very slow when applied to large datasets.
In this paper, we propose a new algorithm which combined rough conditional entropy and naive Bayesian classifier to select features. First, in order to decrease the computational cost of wrapper search, a candidate feature set is selected by using rough conditional mutual information. Second, the candidate feature subset is then further refined by a wrapper procedure. We take advantages of both the filter and the wrapper. The main goal of our research is expected to obtain a few features while the classification accuracy is still very high. This approach provides the possibility of efficiently applying filterwrapper model on some datasets from UCI [4], obtaining better results than other classical feature selection approaches.
In the remainder of the paper, related work is first discussed in the next section. Section 3 presents the preliminaries on Shannon’s entropy and rough sets. Section 4 introduces the definitions of rough uncertainty measure and discusses their properties and interpretation. The proposed hybrid feature selection method is delineated in Section 5. The experimental results are presented in Section 6. Finally, a brief conclusion is given in Section 7.
2. Related Work
In filter based feature selection techniques, a number of relevance measures were applied to measure the performance of features for predicting decisions. These relevance measures can be divided into four categories: distance, dependency, consistency, and information. The most prominent distancebased method is relief [5]. This method uses Euclidean distance to select the relevance features. Since relief works only for binary classes, Kononenko generalized it to multiple classes called reliefF [6, 7]. However, relief and reliefF are unable to detect redundant features. Dependence measures or correlation measures quantify the ability to predict the value of one variable from the value of another variable. Hall’s correlationbased feature selection (CFS) algorithm [8] is typical representative of this category. Consistency measures try to preserve the discriminative power of data in the original feature space. Rough set theory is a popular technique of this sort [9]. Among these measures, mutual information (MI) is the most widely used one in computing relevance. MI is a wellknown concept from information theory and has been used to capture the relevance and redundancy among features. In this paper, we focus on the informationbased measure in the filter model.
The main advantages of MI are its robustness to noise and transformation. In contrast to other measures, MI is not limited to linear dependencies but includes any nonlinear ones. Since Battiti proposed mutual information feature selector (MIFS) [10], more and more researchers began to study informationbased feature selection. MIFS selects the feature that maximizes the information of the class, corrected by subtracting a quantity proportion to the average MI with the previously selected features. Battiti demonstrated that MI can be very useful in feature selection problems and the MIFS can be used in any classifying systems for its simplicity whatever the learning algorithm may be. Kwak and Choi [11] analyzed the limitations of MIFS and proposed method called MIFSU, which, in general, makes a better estimation of the MI between input attributes and output classes than MIFS. They showed that MIFS does not work in nonlinear problems and proposed MIFSU to improve MIFS for solving nonlinear problems. Another variant of MIFS is minredundancy maxrelevance (mRMR) criterion [12]. The method presented the theoretical analysis of the relationships of maxdependency, maxrelevance, and minredundancy. They proved that mRMR is equivalent to maxdependency for the firstorder incremental search.
The limitations of MIFS, MIFSU, and mRMR algorithms are as follows. Firstly, they are all incremental search schemes that select one feature at a time. At each pass, these methods select one feature with maximum criterion, without considering the interaction between groups of features. In many classification problems, groups of several features occurring simultaneously are relevant but not for the case of individual feature alone, for example, the XOR problem. Secondly, the coefficient is a configurable parameter, which must be set experimentally. Thirdly, they are not accurate enough to quantify the dependency among features with respect to a given decision.
Assume as an input feature set and as a target; our task is to select features from a pool such that their joint mutual information is maximized. However, the estimation of mutual information from the available data is a great challenge, especially multivariate mutual information. Martínez Sotoca and Pla [13] and Guo et al. [14] proposed different methods to approximate multivariate conditional mutual information, respectively. Nevertheless, their proofs are all based on the same inequality; that is, . The inequality does not hold under any conditions. Only if random variables , , and satisfy Markovity, then the inequality holds. Many researchers try various methods to estimate mutual information. The most common methods are histogram [15], kernel density estimation (KDE) [16], and knearest neighbor estimation (KNN) [17]. The standard histogram partitions the axes into distinct bins of width and then counts the number of observations; therefore, this estimation method is highly dependent on the choice of the width of the bins. Although the KDE is better than histogram, the bandwidth and kernel function are difficult to decide. The KNN approach uses a fixed number of nearest neighbors to estimate the MI, but it seems more suitable for continuous random variables.
This paper will compute multivariate mutual information and multivariate conditional mutual information in a new perspective. Our method is based on rough entropy uncertainty measure. Several authors [18–21] have used Shannon’s entropy and its variants to measure uncertainty in rough set theory. In this work, we will propose several rough entropybased metrics. Some important properties and relationships of these uncertainty measures will be concluded. Then we will find a candidate feature subset by using rough conditional mutual information to filter the irrelevant and redundant features in the first stage. To overcome the limitations of the filter model, in the second stage, we will use the wrapper model with the sequential backward elimination scheme to search for an optimal feature subset from the candidate feature subset.
3. Preliminaries
In this section we briefly introduce some basic concepts and notations of the information theory and rough set theory.
3.1. Entropy, Mutual Information, and Conditional Mutual Information
Shannon’s information theory, first introduced in 1948 [22], provides a way to measure the information of random variables. The entropy is a measure of uncertainty of random variables [23]. Let be a discrete random variable and let be the probability of ; the entropy of is defined by the following: Here the base of log is 2 and the unit of entropy is the bit. If and are two discrete random variables, the joint probability is , where and . The joint entropy of and is as follows:
When certain variables are known and others are not known, the remaining uncertainty is measured by the conditional entropy as follows: The information found commonly in two random variables is of importance and this is defined as the mutual information between two variables as follows: If the mutual information between two random variables is large (small), it means two variables are closely (not closely) related. If the mutual information becomes zero, the two random variables are totally unrelated or the two variables are independent. The mutual information and the entropy have the following relation: For continuous random variables, the entropy and mutual information are defined as follows: Conditional mutual information is the reduction in the uncertainty of due to knowledge of when is given. The conditional mutual information of random variables and given is defined by the following: Mutual information satisfies a chain rule; that is,
3.2. Rough Sets
Rough sets theory, introduced by Pawlak [24], is a mathematical tool to handle imprecision, uncertainty, and vagueness. It has been applied in many fields [25] such as machine learning, data mining, and pattern recognition.
The notion of an information system provides a convenient basis for the representation of objects in terms of their attributes. An information system is a pair of , where is a nonempty finite set of objects called the universe and is a nonempty finite set of attributes; that is, for , where is called the domain of . A decision table is a special case of information system , where attributes in are called condition attributes and is a designated attribute called the decision attribute.
For every set of attributes , an indiscernibility relation is defined in the following way. Two objects, and , are indiscernible by the set of attribute in , if for every . The equivalence class of is called elementary set in because it represents the smallest discernible groups of objects. For any element of , the equivalence class of in relation is represented as . For , the indiscernibility relation constitutes a partition of , which is denoted by .
Given an information system , for any subset and equivalence relation , the lower and upper approximations of are defined, respectively, as follows:
4. Rough EntropyBased Metrics
In this section, the concept of rough entropy is introduced to measure the uncertainty of knowledge in an information system and then some rough entropybased uncertainty measures are presented. Some important properties of these uncertainty measures are deduced, respectively, and the relationships among them are discussed as well.
Definition 1. Given a set of samples described by features , is a subset of attributes. Then the rough entropy of the sample is defined by and the average entropy of the set of samples is computed as where is the cardinality of .
Since for all , , , so we have . if and only if for for all , ; that is, . if and only if for all , ; that is, . Obviously, when knowledge can distinguish any two objects, the rough entropy is the largest; when knowledge can not distinguish any two objects, the rough entropy is zero.
Theorem 2. Consider , where is Shannon’s entropy.
Proof. Suppose and ,where ; then . Because for and for any , we have
Theorem 2 shows that the rough entropy equals Shannon’s entropy.
Definition 3. Suppose are two subsets of attributes; the joint rough entropy is defined as Due to , therefore, . According to Definition 3, we can observe that .
Theorem 4. Consider and .
Proof. Consider for all ;we have and , and then and . Therefore, and .
Definition 5. Suppose are two subsets of attributes; the conditional rough entropy of to is defined as
Theorem 6 (chain rule). Consider .
Proof. Consider
Definition 7. Suppose are two subsets of attributes; the rough mutual information of and is defined as
Theorem 8 (the relation between rough mutual information and rough entropy). Consider . . .
Proof. The conclusions of (1) and (3) are straightforward; here we give the proof of property (2).
(2) Consider
Definition 9. The rough conditional mutual information of and given is defined by
Theorem 10. The following equations hold: ; .
Proof. (1) Consider
(2) Consider
5. A Hybrid Feature Selection Method
In this section, we propose a novel hybrid feature selection method based on rough conditional mutual information and naive Bayesian classifier.
5.1. Feature Selection by Rough Conditional Mutual Information
Given a set of sample described by the attribute set , in terms of mutual information, the purpose of feature selection is to find a feature set with features, which jointly have the largest dependency on the target class . This criterion, called maxdependency, has the following form: According to the chain rule for information, that is to say, we can select a feature which produces the maximum conditional mutual information, formally written as where represents the selected feature set.
Figure 1 illustrates the validity of this criterion. Here, represents a feature highly correlated with , and is much less correlated with . The mutual information between vectors and is indicated by a shadowed area consisting of three different patterns of patches; that is, , where , , and are defined by different cases of overlap. In detail,(1) is the mutual information between and , that is, ;(2) is the mutual information between and , that is, ;(3) is the mutual information between and , that is, ;(4) is the conditional mutual information between and given , that is, ;(5) is the mutual information between and , that is, .
This illustration clearly shows that the features maximizing the mutual information not only depend on their predictive information individually, for example, , but also need to take account of redundancy between them. In this example, feature should be selected first since the mutual information between and is the largest, and feature should have priority for selection over in spite of the latter having larger individual mutual information with . This is because provides more complementary information to feature to predict than does (as in Figure 1); that is to say, for each round, we should select a feature which maximizes conditional mutual information. From Theorem 2, we know that rough entropy equals Shannon’s entropy; therefore, we can select a feature which produces the maximum rough conditional mutual information.
We adopt the forward feature algorithm to select features. Each single input feature is added to selected features set based on maximizing rough conditional mutual information, that is, given selected feature set , maximizing the rough mutual information of and target class , where belongs to the remain feature set. In order to apply the rough conditional mutual information measure to the filter model well, a numerical threshold value is set to . This can help the algorithm to be resistant to noise data and to overcome the overfitting problem to a certain extent [26]. The procedure can be performed until is satisfied. The filter algorithm can be described by the following procedure.(1)Initialization: set “initial set of all features,” “empty set,” and “class outputs.”(2)Computation of the rough mutual information of the features with the class outputs: for each feature , compute .(3)Selection of the first feature: find the feature that maximizes ; set and .(4)Greedy selection: repeat until the termination condition is satisfied:(a)computation of the rough mutual information for each feature ,(b)selection of the next feature: choose the feature as the one that maximizes ; set and .(5)Output the set containing the selected features: .
5.2. Selecting the Best Feature Subset on Wrapper Approach
The wrapper model uses the classification accuracy of a predetermined learning algorithm to determine the goodness of the selected subset. It searches for features that are better suited to the learning algorithm, aiming at improving the performance of the learning algorithm; therefore, the wrapper approach generally outperforms the filter approach in the aspect of the final predictive accuracy of a learning machine. However, it is more computationally expensive than the filter models. Although many wrapper methods are not exhaustive search, most of them still incur time complexity [27, 28] where is the number of features of the dataset. Hence, it is worth reducing the search space before using wrapper feature selection approach. Through the filter model, it can reduce high computational cost and avoid encountering the local maximal problem. Therefore, the final subset of the features obtained contains a few features while the predictive accuracy is still high.
In this work, we propose the reducing of the search space of the original feature set to the best candidate which can reduce the computational cost of the wrapper search effectively. Our method uses the sequential backward elimination technique to search for every possible subset of features through the candidate space.
The features are ranked according to the average accuracy of the classifier, and then features will be removed one by one from the candidate feature subset only if such exclusion improves or does not change the classifier accuracy. Different kinds of learning models can be applied to wrappers. However, different kinds of learning machines have different discrimination abilities. Naive Bayesian classifier is widely used in machine learning because it is fast and easy to be implemented. Rennie et al. [29] show that its performance is competitive with the stateoftheart models like SVM while the latter has too many parameters to decide. Therefore, we choose the naive Bayesian classifier as the core of fine tuning. The decrement selection procedure for selecting an optimal feature subset based on the wrapper approach can be seen as shown in Algorithm 1.

There are two phases in the wrapper algorithm, as shown in wrapper algorithm. In the first stage, we compute the classification accuracy of the candidate feature set which is the results of filter model (step 1), where Classperf () represents the average classification accuracy of dataset D with candidate features C. The results are obtained by 10fold crossvalidation. For each , we compute the average accuracy . Then features are ranked according to value (steps 3–6). In the second stage, we deal with the list of the ordered features once; each feature in the list determines the first till the last ranked feature (steps 8–26). In this stage, each feature in the list considers the average accuracy of the naive Bayesian classifier only if the feature is excluded. If any feature is found to lead to the most improved average accuracy and the relative accuracy [30] is more than (steps 11–14), the feature then will be removed. Otherwise, every possible feature is considered and the feature that leads to the largest average accuracy will be chosen and removed (step 15). The one that leads to the improvement or the unchanging of the average accuracy (steps 17–20) or the degrading of the relative accuracy not worse than (steps 21–24) will be removed. In general, should take value in [0, 0.1] and should take value in [0, 0.02]. In the following, if not specified, and .
This decrement selection procedure is repeated until the termination condition is satisfied. Usually, the sequential backward elimination is more computationally expensive than the incremental sequential forward search. However, it could yield a better result when considering the local maximal. In addition, the sequential forward search adding one feature at each pass does not take the interaction between the groups of the features into account [31]. In many classification problems, the class variable may be affected by grouping several features but not the individual feature alone. Therefore, the sequential forward search is unable to find the dependencies between the groups of the features while the performance can be degraded sometimes.
6. Experimental Results
This section illustrates the evaluation of our method in terms of the classification accuracy and the number of selected features in order to see how good the filter wrapper is in the situation of large and middlesized features. In addition, the performance of the rough conditional mutual information algorithm is compared with three typical feature selection methods which are based on three different evaluation criterions, respectively. These methods include correlation based feature selection (CFS), consistency based algorithm, and minredundancy maxrelevance (mRMR). The results illustrate the efficiency and effectiveness of our method.
In order to compare our hybrid method with some classical techniques, 10 databases are downloaded from UCI repository of machine learning databases. All these datasets are widely used by the data mining community for evaluating learning algorithms. The details of the 10 UCI experimental datasets are listed in Table 1. The sizes of databases vary from 101 to 2310, the numbers of original features vary from 12 to 279, and the numbers of classes vary from 2 to 19.

6.1. Unselect versus CFS, Consistency Based Algorithm, mRMR, and RCMI
In Section 5, rough conditional mutual information is used to filter the redundant and irrelevant features. In order to compute the rough mutual information, we employ Fayyad and Irani’s MDL discretization algorithm [32] to transform continuous features into discrete ones.
We use naive Bayesian and CART classifier to test the classification accuracy of selected features with different feature selection methods. The results in Tables 2 and 3 show the classification accuracies and the number of selected features obtained by the original feature (unselect), RCMI, and other feature selectors. According to Tables 2 and 3, we can find that the selected feature by RCMI has the highest average accuracy in terms of naive Bayes and CART. It can also be observed that RCMI can achieve the least average number of selected features which is the same as mRMR. This shows that RCMI is better than CFS and consistency based algorithm and is comparable to mRMR.


In addition, to illustrate the efficiency of RCMI, we experiment on Ionosphere, Sonar, and Wine datasets, respectively. A different number of the selected features obtained by RCMI and mRMR are tested on naive Bayesian classifier, as presented in Figures 2, 3, and 4. In Figures 2–4, the classification accuracies are the results of 10fold crossvalidation tested by naive Bayes. The number in axis refers to the first features with the selected order by different methods. The results in Figures 2–4 show that the average accuracy of classifier with RCMI is comparable to mRMR in the majority of cases. We can see that the maximum value of the plots for each dataset with RCMI method is higher than mRMR. For example, the highest accuracy of Ionosphere achieved by RCMI is 94.87% while the highest accuracy achieved by mRMR is 90.60%. At the same time, we can also notice that the RCMI method has the number of maximum values higher than mRMR. It shows that RCMI is an effective measure for feature selection.
However, the number of the features selected by the RCMI method is still more in some datasets. Therefore, to improve performance and reduce the number of the selected features, these problems were conducted by using the wrapper method. With removal of the redundant and irreverent features, the core of wrappers for fine tuning can perform much faster.
6.2. Filter Wrapper versus RCMI and Unselect
Similarly, we also use naive Bayesian and CART to test the classification accuracy of selected features with filter wrapper, RCMI, and unselect. The results in Tables 4 and 5 show the classification accuracies and the number of selected features.


Now we analyze the performance of these selected features. First, we can conclude that although most of features are removed from the raw data, the classification accuracies do not decrease; on the contrary, the classification accuracies increase in the majority of datasets. The average accuracies derived from RCMI and filterwrapper method are all higher than the unselect datasets with respect to naive Bayes and CART. With respect to naive Bayesian learning algorithm, the average accuracy is 91.47% for filter wrapper, while 89.86% for unselect. The average classification accuracy increased 1.8%. With respect to CART learning algorithm, the average accuracy is 89.94% for filter wrapper, while 88.08% for unselect. The average classification accuracy increased 2.1%. The average number of selected features is 5.3 for filter wrapper, while 9.3 for RCMI and 50.6 for unselect. The average number of selected features reduced 43% and 89.5%, respectively. Therefore, the average value of classification accuracy and number of features obtained from the filterwrapper method are better than those obtained from the RCMI and unselect. In other words, using the RCMI and wrapper methods as a hybrid improves the classification efficiency and accuracy compared with using the RCMI method individually.
7. Conclusion
The main goal of feature selection is to find a feature subset as small as possible, while the feature subset has highly prediction accuracy. A hybrid feature selection approach which takes advantages of filter model and wrapper model has been presented in this paper. In the filter model, measuring the relevance between features plays an important role. A number of measures were proposed. Mutual information is widely used for its robustness. However, it is difficult to compute mutual information, especially multivariate mutual information. We proposed a set of rough based metrics to measure the relevance between features and analyzed some important properties of these uncertainty measures. We have proved that the RCMI can substitute Shannon’s conditional mutual information; thus, RCMI can be used as an effective measure to filter the irrelevant and redundant features. Based on the candidate feature subset by RCMI, naive Bayesian classifier is applied to the wrapper model. The accuracy of naive Bayesian and CART classifier was used to evaluate the goodness of feature subsets. The performance of the proposed method is evaluated based on ten UCI datasets. Experimental results on ten UCI datasets show that the filterwrapper method outperformed CFS, consistency based algorithm, and mRMR at most cases. Our technique not only chooses a small subset of features from a candidate subset but also provides good performance in predictive accuracy.
Conflict of Interests
The authors declare that there is no conflict of interests regarding the publication of this paper.
Acknowledgment
This work was supported by the National Natural Science Foundation of China (70971137).
References
 I. Guyon and A. Elisseeff, “An introduction to variable and feature selection,” Journal of Machine Learning Research, vol. 3, pp. 1157–1182, 2003. View at: Google Scholar
 M. Dash and H. Liu, “Feature selection for classification,” Intelligent Data Analysis, vol. 1, no. 1–4, pp. 131–156, 1997. View at: Publisher Site  Google Scholar
 M. Dash and H. Liu, “Consistencybased search in feature selection,” Artificial Intelligence, vol. 151, no. 12, pp. 155–176, 2003. View at: Publisher Site  Google Scholar  Zentralblatt MATH  MathSciNet
 C. J. Merz and P. M. Murphy, UCI Repository of Machine Learning Databases, Department of Information and Computer Science, University of California, Irvine, Calif, USA, 1996, http://mlearn.ics.uci.edu/MLRepository.html.
 K. Kira and L. A. Rendell, “Feature selection problem: traditional methods and a new algorithm,” in Proceedings of the 9th National Conference on Artificial Intelligence (AAAI '92), pp. 129–134, July 1992. View at: Google Scholar
 M. RobnikŠikonja and I. Kononenko, “Theoretical and empirical analysis of ReliefF and RReliefF,” Machine Learning, vol. 53, no. 12, pp. 23–69, 2003. View at: Publisher Site  Google Scholar
 I. Kononenko, “Estimating attributes: analysis and extension of RELIEF,” in Proceedings of European Conference on Machine Learning (ECML '94), pp. 171–182, 1994. View at: Google Scholar
 M. A. Hall, Correlationbased feature subset selection for machine learning [Ph.D. thesis], Department of Computer Science, University of Waikato, Hamilton, New Zealand, 1999.
 J. G. Bazan, “A comparison of dynamic and nondynamic rough set methods for extracting laws from decision table,” in Rough Sets in Knowledge Discovery, L. Polkowski and A. Skowron, Eds., pp. 321–365, Physica, Heidelberg, Germany, 1998. View at: Google Scholar
 R. Battiti, “Using mutual information for selecting features in supervised neural net learning,” IEEE Transactions on Neural Networks, vol. 5, no. 4, pp. 537–550, 1994. View at: Publisher Site  Google Scholar
 N. Kwak and C. H. Choi, “Input feature selection for classification problems,” IEEE Transactions on Neural Networks, vol. 13, no. 1, pp. 143–159, 2002. View at: Publisher Site  Google Scholar
 H. Peng, F. Long, and C. Ding, “Feature selection based on mutual information: criteria of maxdependency, maxrelevance, and minredundancy,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 27, no. 8, pp. 1226–1238, 2005. View at: Publisher Site  Google Scholar
 J. Martínez Sotoca and F. Pla, “Supervised feature selection by clustering using conditional mutual informationbased distances,” Pattern Recognition, vol. 43, no. 6, pp. 2068–2081, 2010. View at: Publisher Site  Google Scholar
 B. Guo, R. I. Damper, S. R. Gunn, and J. D. B. Nelson, “A fast separabilitybased featureselection method for highdimensional remotely sensed image classification,” Pattern Recognition, vol. 41, no. 5, pp. 1670–1679, 2008. View at: Publisher Site  Google Scholar
 D. W. Scott, Multivariate Density Estimation: Theory, Practice and Visualization, John Wiley & Sons, New York, NY, USA, 1992. View at: Publisher Site  MathSciNet
 B. W. Silverman, Density Estimation for Statistics and Data Analysis, Chapman & Hall, London, UK, 1986. View at: MathSciNet
 A. Kraskov, H. Stögbauer, and P. Grassberger, “Estimating mutual information,” Physical Review E, vol. 69, no. 6, Article ID 066138, 16 pages, 2004. View at: Publisher Site  Google Scholar  MathSciNet
 T. Beaubouef, F. E. Petry, and G. Arora, “Informationtheoretic measures of uncertainty for rough sets and rough relational databases,” Information Sciences, vol. 109, no. 1–4, pp. 185–195, 1998. View at: Google Scholar
 I. Düntsch and G. Gediga, “Uncertainty measures of rough set prediction,” Artificial Intelligence, vol. 106, no. 1, pp. 109–137, 1998. View at: Publisher Site  Google Scholar  Zentralblatt MATH  MathSciNet
 G. J. Klir and M. J. Wierman, Uncertainty Based Information: Elements of Generalized Information Theory, Physica, New York, NY, USA, 1999. View at: MathSciNet
 J. Liang and Z. Shi, “The information entropy, rough entropy and knowledge granulation in rough set theory,” International Journal of Uncertainty, Fuzziness and KnowledgeBased Systems, vol. 12, no. 1, pp. 37–46, 2004. View at: Publisher Site  Google Scholar  Zentralblatt MATH  MathSciNet
 C. E. Shannon, “A mathematical theory of communication,” The Bell System Technical Journal, vol. 27, pp. 379–423, 1948. View at: Google Scholar  Zentralblatt MATH  MathSciNet
 T. M. Cover and J. A. Thomas, Elements of Information Theory, John Wiley & Sons, New York, NY, USA, 1991. View at: Publisher Site  MathSciNet
 Z. Pawlak, “Rough sets,” International Journal of Computer and Information Sciences, vol. 11, no. 5, pp. 341–356, 1982. View at: Publisher Site  Google Scholar  Zentralblatt MATH  MathSciNet
 X. Hu and N. Cercone, “Learning in relational databases: a rough set approach,” Computational Intelligence, vol. 11, no. 2, pp. 323–338, 1995. View at: Google Scholar
 J. Huang, Y. Cai, and X. Xu, “A hybrid genetic algorithm for feature selection wrapper based on mutual information,” Pattern Recognition Letters, vol. 28, no. 13, pp. 1825–1844, 2007. View at: Publisher Site  Google Scholar
 H. Liu and L. Yu, “Toward integrating feature selection algorithms for classification and clustering,” IEEE Transactions on Knowledge and Data Engineering, vol. 17, no. 4, pp. 491–502, 2005. View at: Publisher Site  Google Scholar
 L. Yu and H. Liu, “Efficient feature selection via analysis of relevance and redundancy,” Journal of Machine Learning Research, vol. 5, pp. 1205–1224, 2003/04. View at: Publisher Site  Google Scholar  MathSciNet
 J. D. M. Rennie, L. Shih, J. Teevan, and D. Karger, “Tackling the poor assumptions of naive bayes text classifiers,” in Proceedings, Twentieth International Conference on Machine Learning, pp. 616–623, Washington, DC, USA, August 2003. View at: Google Scholar
 R. Setiono and H. Liu, “Neuralnetwork feature selector,” IEEE Transactions on Neural Networks, vol. 8, no. 3, pp. 654–662, 1997. View at: Publisher Site  Google Scholar
 S. Foithong, O. Pinngern, and B. Attachoo, “Feature subset selection wrapper based on mutual information and rough sets,” Expert Systems with Applications, vol. 39, no. 1, pp. 574–584, 2012. View at: Publisher Site  Google Scholar
 U. Fayyad and K. Irani, “Multiinterval discretization of continuousvalued attributes for classification learning,” in Proceedings of the 13th International Joint Conference on Artificial Intelligence, pp. 1022–1027, Morgan Kaufmann, San Mateo, Calif, USA, 1993. View at: Google Scholar
Copyright
Copyright © 2014 Zilin Zeng et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.