BioMed Research International

BioMed Research International / 2015 / Article
Special Issue

Sequence and Structure Analysis of Biological Molecules Based on Computational Methods

View this Special Issue

Research Article | Open Access

Volume 2015 |Article ID 910515 |

Manli Zhou, Youxi Luo, Guoquan Sun, Guoqin Mai, Fengfeng Zhou, "Constraint Programming Based Biomarker Optimization", BioMed Research International, vol. 2015, Article ID 910515, 5 pages, 2015.

Constraint Programming Based Biomarker Optimization

Academic Editor: Antonello Merlino
Received28 Sep 2014
Accepted13 Dec 2014
Published05 May 2015


Efficient and intuitive characterization of biological big data is becoming a major challenge for modern bio-OMIC based scientists. Interactive visualization and exploration of big data is proven to be one of the successful solutions. Most of the existing feature selection algorithms do not allow the interactive inputs from users in the optimizing process of feature selection. This study investigates this question as fixing a few user-input features in the finally selected feature subset and formulates these user-input features as constraints for a programming model. The proposed algorithm, fsCoP (feature selection based on constrained programming), performs well similar to or much better than the existing feature selection algorithms, even with the constraints from both literature and the existing algorithms. An fsCoP biomarker may be intriguing for further wet lab validation, since it satisfies both the classification optimization function and the biomedical knowledge. fsCoP may also be used for the interactive exploration of bio-OMIC big data by interactively adding user-defined constraints for modeling.

1. Introduction

Biological big data is being accumulated at an accelerated speed, facilitated by the rapid invention and development of bio-OMIC data production technologies [1, 2]. Interactive exploration technology is widely used to mine knowledge from various big data areas [3] and may be useful to rapidly and accurately detect phenotype-associated biomarkers from the huge amount of bio-OMIC data [4]. This is usually formulated as the feature selection problem [5, 6].

Various algorithms have been proposed to choose a few from a large number of features, by optimizing a phenotypic measurement. The principle of parsimony prefers a minimum number of features for an accurate representation of the data [7]. Detailed introduction may be found for both general feature selection algorithms [8] and phenotype-associated biomarker detection algorithms [9] from the literature. Considering millions or more of bio-OMIC features for each sample, although the exhaustive search guarantees the detection of optimal feature subset, its computational requirement exceeds the capacity of any high-performance computing systems under the current parallel computing architecture. So all the existing feature selection algorithms screen for the suboptimal solutions based on some heuristic rules.

Heuristic feature selection algorithms may be grouped as two classes based on how they generate the finally chosen features. The class I wrapper or group optimization algorithms evaluate a feature subset by testing its classification performance with a learning algorithm. The features are selected by heuristic rules or randomly, and only the feature subset with the best classification performance will be kept for further investigation, for example, forward stepwise selection [10] and ant colony optimization [11]. The class II filtering or individual ranking algorithms measure each feature’s correlation with the class labels and rank the features by their measurement. A heuristic assumption is that the combination of top-ranked features should produce a good classification performance, where is an arbitrarily chosen integer. They are usually much faster than the class I algorithms but lack model robustness due to the ignorance of feature interdependence [12]. It is also difficult to determine how many features should be chosen from the ordered feature list.

This work proposes a constraint programming based interactive feature selection algorithm, fsCoP, for efficient exploration of the bio-OMIC big data. An interactive feature selection problem requires a fast and accurate detection of features and the integration of user-input features in the final result. The majority of existing feature selection algorithms do not consider how to make sure a given feature subset appears among the finally selected features. fsCoP fixes the user-input features in the result by formulating them as constraints of the programming model. Our data show that features chosen by fsCoP perform well similar to or much better than the existing feature selection algorithms in classification, even with the constraints of fixed features from both literature and other algorithms.

2. Materials and Methods

2.1. Dataset Downloading and Preprocessing

Two microarray-based gene expression profiling datasets are downloaded from the NCBI GEO database [13]. Both datasets GSE5406 [14] and GSE1869 [15] profiled ischemic cardiomyopathy samples and their controls on the Affymetrix Human Genome U133A Array (HG-U133A) platform. The transcriptomes are normalized using the RMA algorithm [16]. The gene expression profiles of ischemic cardiomyopathy samples and the nonfailing controls are kept for binary classification study in this work.

2.2. Feature Selection Based on Constraint Programming (fsCoP)

This work proposes a constraint programming based feature selection algorithm, allowing the user to determine a few features in the finally chosen feature subset. The prefixed features may be the biomarkers known to be associated with the phenotype in the literature or the features selected by other feature selection algorithms. This model is proposed to answer the biological questions like whether a few genes together with the ischemic cardiomyopathy associated ACE2 (angiotensin-converting enzyme-2) may constitute an accurate model for the disease early detection. The majority of the existing feature selection algorithms do not have the integrating component for fixing a few features in the final feature subset. Let FixedSubset be the set of features to be fixed in the final result and let be the class number. Class has samples, where . The programming model is defined as follows:The average value of the th feature is denoted as for the samples in class . Formula (2) makes that the centroid of class is the closest centroid to the samples of the class . Each prefixed feature has the weight no smaller than MinWeight. Each feature has a weight , where only features with positive weights are selected by the algorithm.

2.3. Classification Performance Measurements

A binary classification model is trained over the datasets of positive and negative samples, whose numbers are and , respectively. The classification performance is usually measured by the sensitivity and specificity , where , , , and are the numbers of true positives, false negatives, true negatives, and false positives. The overall classification performances may be measured by the overall accuracy and balanced overall accuracy . Matthew’s Correlation Coefficient is also calculated to measure how well a classification model is, and it is defined as , where is the squared root of .

Fivefold cross validation (5FCV) strategy is used to train the model and calculate how well a model performs. Fluctuation may occur for different seeds of the random number generator. So 30 runs of the 5FCV experiments are carried out with different random seeds.

2.4. Comparison with Four Feature Selection Algorithms

The proposed feature selection algorithm fsCoP is compared with two ranking algorithms, that is, -test (TRank) [17] and Wilcoxon test (WRank) [18], and two other widely used algorithms, that is, prediction analysis of microarrays (PAM) [19] and regularized random forest (RRF) [20].

The ultimate goal of the proposed model is to select a subset of features with accurate classification performance. The performance of a given feature subset is measured by five widely used classification algorithms, including support vector machine (SVM) [21], Naive Bayesian [22], decision tree (DTree) [23], Lasso [24], and -nearest neighbor [25]. The classification model with the best Matthew’s Correlation Coefficients is kept for the comparison study.

This work uses the default parameters of all the investigated algorithms implemented in the statistical software R/Rstudio version 3.1.1 released on July 10, 2014 [26, 27]. A classification model is usually obtained by trying multiple classification algorithms [28, 29]. So this work compares the feature selection algorithms based on the highest MCC values of the five aforementioned classification algorithms.

3. Results and Discussion

3.1. Constrains from the Literature

The angiotensin-converting enzyme-2 (ACE2) at the location Xp22.2 of the human genome HG19 is chosen to be fixed in the algorithm fsCoP, denoted as fsCoP(ACE2). ACE2 was observed to be differentially expressed between ischemic and nonischemic cardiomyopathy and may play a role in transducing the signal of heart failure pathophysiology [15]. The expression level of ACE2 is detected by two probe sets (219962_at and 222257_s_at) in the Affymetrix microarray platform U133A (GPL96). These two features will be fixed in the feature subset fsCoP(ACE2), and the performances of the five classification algorithms are compared using the selected features by fsCoP and fsCoP(ACE2).

Firstly, fsCoP and fsCoP(ACE2) achieve similarly good performance on the two investigated datasets, that is, GSE5406 and GSE1869. Table 1 shows that, except the decision tree algorithm on the dataset GSE1869, there are no greater than 0.021 differences in MCC between the two versions of fsCoP. The greatest difference occurs for the NBayes classification algorithm on the dataset GSE1869, where fsCoP(ACE2) (1.000 in MCC) improves fsCoP (0.979).



SVM1.000 1.000 1.000 1.000 1.000
NBayes1.000 0.998 1.000 0.999 0.999
DTree0.992 0.800 0.967 0.896 0.848
Lasso0.999 0.900 0.987 0.950 0.939
KNN1.000 0.871 0.983 0.936 0.923


SVM1.000 0.996 0.999 0.998 0.998
NBayes1.000 0.998 1.000 0.999 0.999
DTree0.993 0.796 0.967 0.894 0.847
Lasso1.000 0.907 0.988 0.953 0.944
KNN0.999 0.860 0.982 0.930 0.916



SVM1.000 0.955 0.983 0.978 0.965
NBayes1.000 0.972 0.990 0.986 0.979
DTree0.907 0.000 0.567 0.453 NaN
Lasso0.960 0.989 0.971 0.974 0.943
KNN1.000 0.994 0.998 0.997 0.996


SVM1.000 0.939 0.977 0.970 0.953
NBayes1.000 1.000 1.000 1.000 1.000
DTree0.987 0.000 0.617 0.493 NaN
Lasso0.990 0.967 0.981 0.978 0.962
KNN1.000 1.000 1.000 1.000 1.000

Secondly, if only the best classification algorithm is chosen for each subset of selected features, fsCoP(ACE2) also performs well similar with fsCoP. SVM(fsCoP) only improves NBayes(fsCoP(ACE2)) by 0.001 in MCC. The other classification performance measurements also show that this is a minor improvement, with the maximal difference being 0.002 in specificity (Sp). The comparison of the best classification models between the two datasets in Table 1 also shows that NBayes(fsCoP(ACE2)) even performs 0.002 better than KNN(fsCoP) on the dataset GSE1869.

fsCoP runs fast similar with or without fixing a few features. The running time of the algorithm fsCoP with or without fixing user-selected features is compared between fsCoP and fsCoP(ACE2). Since fsCoP runs very fast, we repeat the model testing for multiple times with different random seeds, as in Table 2. The data suggests that fsCoP(ACE2) runs slightly faster than fsCoP for most of the times, except for the case of 10 repeats.

RepeatfsCoPAvg (fsCoP)fsCoP(ACE2)Avg (fsCoP(ACE2))

511.95 2.39 11.78 2.36
1023.83 2.38 23.96 2.40
50120.01 2.40 117.79 2.36
100240.23 2.40 236.75 2.37

3.2. Comparison of fsCoP(ACE2) with the Existing Feature Selection Algorithms

A further comparison of fsCoP(ACE2) with the other existing feature selection algorithms is conducted for the best classification algorithms on each of the selected features, as shown in Figure 1. First of all, fsCoP(ACE2) performs the best (100%) in sensitivity (Sn) with the classification algorithm NBayes on both datasets, as in Figures 1(a) and 1(b). SVM(TRank) achieves the same sensitivities for both datasets, and KNN(PAM) also achieves 100% in Sn on the dataset GSE1869. NBayes(fsCoP(ACE2)) achieves 0.998 in specificity (Sp) on the dataset GSE5406, and no other feature selection algorithms reach the same specificity level. Figure 1(a) suggests that the second best feature selection algorithm may be TRank, which achieves 0.964 in MCC on the dataset GSE5406.

3.3. Constraints from the Existing Feature Selection Algorithms

Except for the features selected by TRank, fsCoP improves all the other three feature selection algorithms. fsCoP(A) is defined to be feature list selected by fsCoP, with the fixed features selected by Algorithm A. Figure 2 shows that fsCoP(TRank) achieves the same classification performance as TRank, and, for the three other feature selection algorithms, fsCoP() achieves higher averaged values and smaller standard deviations for all the five classification performance measurements. The most significant improvement of fsCoP is observed for the RRF algorithm, with 0.0916 in Sp improvement. So besides the integration of known biomarkers from the literature, fsCoP may also be used to further refine the feature subset selected by the existing feature selection algorithms. Better classification performance with smaller fluctuation may be obtained stably by fsCoP, compared with the algorithms.

After the further refining by fsCoP, features selected by all the four feature selection algorithms achieve 100% in the classification sensitivity, while maintaining at least 92% in specificity. And at least 0.95 in MCC is achieved for all the four cases.

Conflict of Interests

The authors declare that there is no conflict of interests regarding the publication of this paper.

Authors’ Contribution

Manli Zhou and Youxi Luo contributed equally to this work.


This work was supported by the Strategic Priority Research Program of the Chinese Academy of Sciences (XDB13040400), Shenzhen Peacock Plan (KQCX20130628112914301), Shenzhen Research Grants (ZDSY20120617113021359, CXB201104220026A, and JCYJ20130401170306884), the MOE Humanities Social Sciences Fund (no. 13YJC790105) and Doctoral Research Fund of HBUT (no. BSQD13050), and Key Laboratory of Human-Machine-Intelligence Synergic Systems, Chinese Academy of Sciences. Computing resources were partly provided by the Dawning supercomputing clusters at SIAT CAS. The constructive comments from the two anonymous reviewers are greatly appreciated.


  1. E. L. van Dijk, H. Auger, Y. Jaszczyszyn, and C. Thermes, “Ten years of next-generation sequencing technology,” Trends in Genetics, vol. 30, no. 9, pp. 418–426, 2014. View at: Publisher Site | Google Scholar
  2. N. Chen, W. Sun, X. Deng et al., “Quantitative proteome analysis of HCC cell lines with different metastatic potentials by SILAC,” Proteomics, vol. 8, no. 23-24, pp. 5108–5118, 2008. View at: Publisher Site | Google Scholar
  3. J. Heer and S. Kandel, “Interactive analysis of big data,” XRDS: Crossroads, vol. 19, no. 1, pp. 50–54, 2012. View at: Publisher Site | Google Scholar
  4. H. Ge, A. J. M. Walhout, and M. Vidal, “Integrating “omic” information: a bridge between genomics and systems biology,” Trends in Genetics, vol. 19, no. 10, pp. 551–560, 2003. View at: Publisher Site | Google Scholar
  5. S. Datta and V. Pihur, “Feature selection and machine learning with mass spectrometry data,” Methods in Molecular Biology, vol. 593, pp. 205–229, 2010. View at: Publisher Site | Google Scholar
  6. Y. Saeys, I. Inza, and P. Larrañaga, “A review of feature selection techniques in bioinformatics,” Bioinformatics, vol. 23, no. 19, pp. 2507–2517, 2007. View at: Publisher Site | Google Scholar
  7. D. A. Bell and H. Wang, “A formalism for relevance and its application in feature subset selection,” Machine Learning, vol. 41, no. 2, pp. 175–195, 2000. View at: Publisher Site | Google Scholar
  8. I. Iguyon and A. Elisseeff, “An introduction to variable and feature selection,” Journal of Machine Learning Research, vol. 3, pp. 1157–1182, 2003. View at: Google Scholar
  9. S. Baek, C.-A. Tsai, and J. J. Chen, “Development of biomarker classifiers from high-dimensional data,” Briefings in Bioinformatics, vol. 10, no. 5, pp. 537–546, 2009. View at: Publisher Site | Google Scholar
  10. S. Colak and C. Isik, “Feature subset selection for blood pressure classification using orthogonal forward selection,” in Proceedings of the IEEE 29th Annual Northeast Bioengineering Conference, pp. 122–123, IEEE, March 2003. View at: Google Scholar
  11. J. Yang and V. Honavar, “Feature subset selection using a genetic algorithm,” in Feature Extraction, Construction and Selection, pp. 117–136, Springer, 1998. View at: Google Scholar
  12. I. A. Gheyas and L. S. Smith, “Feature subset selection in large dimensionality domains,” Pattern Recognition, vol. 43, no. 1, pp. 5–13, 2010. View at: Publisher Site | Google Scholar
  13. T. Barrett, S. E. Wilhite, P. Ledoux et al., “NCBI GEO: archive for functional genomics data sets—update,” Nucleic Acids Research, vol. 41, no. D1, pp. D991–D995, 2013. View at: Publisher Site | Google Scholar
  14. S. Hannenhalli, M. E. Putt, J. M. Gilmore et al., “Transcriptional genomics associates FOX transcription factors with human heart failure,” Circulation, vol. 114, no. 12, pp. 1269–1276, 2006. View at: Publisher Site | Google Scholar
  15. M. M. Kittleson, K. M. Minhas, R. A. Irizarry et al., “Gene expression analysis of ischemic and nonischemic cardiomyopathy: shared and distinct genes in the development of heart failure,” Physiological Genomics, vol. 21, pp. 299–307, 2005. View at: Publisher Site | Google Scholar
  16. R. A. Irizarry, B. Hobbs, F. Collin et al., “Exploration, normalization, and summaries of high density oligonucleotide array probe level data,” Biostatistics, vol. 4, no. 2, pp. 249–264, 2003. View at: Publisher Site | Google Scholar
  17. P. Baldi and A. D. Long, “A Bayesian framework for the analysis of microarray expression data: regularized t-test and statistical inferences of gene changes,” Bioinformatics, vol. 17, no. 6, pp. 509–519, 2001. View at: Publisher Site | Google Scholar
  18. W.-M. Liu, R. Mei, X. Di et al., “Analysis of high density expression microarrays with signed-rank call algorithms,” Bioinformatics, vol. 18, no. 12, pp. 1593–1599, 2002. View at: Publisher Site | Google Scholar
  19. R. Tibshirani, T. Hastie, B. Narasimhan, and G. Chu, “Diagnosis of multiple cancer types by shrunken centroids of gene expression,” Proceedings of the National Academy of Sciences of the United States of America, vol. 99, no. 10, pp. 6567–6572, 2002. View at: Publisher Site | Google Scholar
  20. H. T. Deng and G. Runger, “Feature selection via regularized trees,” in Proceedings of the IEEE International Joint Conference on Neural Networks (IJCNN '12), pp. 1–8, Brisbane, Australia, June 2012. View at: Publisher Site | Google Scholar
  21. S. Amari and S. Wu, “Improving support vector machine classifiers by modifying kernel functions,” Neural Networks, vol. 12, no. 6, pp. 783–789, 1999. View at: Publisher Site | Google Scholar
  22. G. H. John and P. Langley, “Estimating continuous distributions in Bayesian classifiers,” in Proceedings of the Eleventh conference on Uncertainty in artificial intelligence: 1995, pp. 338–345, Morgan Kaufmann, 1995. View at: Google Scholar
  23. S. R. Safavian and D. Landgrebe, “A survey of decision tree classifier methodology,” IEEE Transactions on Systems, Man, and Cybernetics, vol. 21, no. 3, pp. 660–674, 1991. View at: Publisher Site | Google Scholar | MathSciNet
  24. D. Ghosh and A. M. Chinnaiyan, “Classification and selection of biomarkers in genomic data using LASSO,” Journal of Biomedicine and Biotechnology, vol. 2005, no. 2, pp. 147–154, 2005. View at: Publisher Site | Google Scholar
  25. M.-L. Zhang and Z.-H. Zhou, “A k-nearest neighbor based algorithm for multi-label classification,” in Proceedings of the IEEE International Conference on Granular Computing, vol. 2, pp. 718–721, Beijing, China, July 2005. View at: Publisher Site | Google Scholar
  26. R: A Language and Environment for Statistical Computing,
  27. J. S. Racine, “RSTUDIO: a platform-independent IDE for R and sweave,” Journal of Applied Econometrics, vol. 27, no. 1, pp. 167–172, 2012. View at: Publisher Site | Google Scholar
  28. F. Zhou and Y. Xu, “cBar: a computer program to distinguish plasmid-derived from chromosome-derived sequence fragments in metagenomics data,” Bioinformatics, vol. 26, no. 16, pp. 2051–2052, 2010. View at: Publisher Site | Google Scholar
  29. P. Guo, Y. Luo, G. Mai et al., “Gene expression profile based classification models of psoriasis,” Genomics, vol. 103, no. 1, pp. 48–55, 2014. View at: Publisher Site | Google Scholar

Copyright © 2015 Manli Zhou et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

More related articles

 PDF Download Citation Citation
 Download other formatsMore
 Order printed copiesOrder

Related articles

Article of the Year Award: Outstanding research contributions of 2020, as selected by our Chief Editors. Read the winning articles.