BioMed Research International

BioMed Research International / 2016 / Article

Research Article | Open Access

Volume 2016 |Article ID 5413903 |

Huan Yang, Hua Tang, Xin-Xin Chen, Chang-Jian Zhang, Pan-Pan Zhu, Hui Ding, Wei Chen, Hao Lin, "Identification of Secretory Proteins in Mycobacterium tuberculosis Using Pseudo Amino Acid Composition", BioMed Research International, vol. 2016, Article ID 5413903, 7 pages, 2016.

Identification of Secretory Proteins in Mycobacterium tuberculosis Using Pseudo Amino Acid Composition

Academic Editor: Xun Lan
Received23 Apr 2016
Accepted18 Jul 2016
Published11 Aug 2016


Tuberculosis is killing millions of lives every year and on the blacklist of the most appalling public health problems. Recent findings suggest that secretory protein of Mycobacterium tuberculosis may serve the purpose of developing specific vaccines and drugs due to their antigenicity. Responding to global infectious disease, we focused on the identification of secretory proteins in Mycobacterium tuberculosis. A novel method called MycoSec was designed by incorporating -gap dipeptide compositions into pseudo amino acid composition. Analysis of variance-based technique was applied in the process of feature selection and a total of 374 optimal features were obtained and used for constructing the final predicting model. In the jackknife test, MycoSec yielded a good performance with the area under the receiver operating characteristic curve of 0.93, demonstrating that the proposed system is powerful and robust. For user’s convenience, the web server MycoSec was established and an obliging manual on how to use it was provided for getting around any trouble unnecessary.

1. Introduction

Mycobacterium tuberculosis (M. tuberculosis or MTB), also known as acid-fast bacilli, is the causative pathogen of the contagious disease tuberculosis (TB). Over the two decades, despite booming development of molecular technologies and diagnostic platforms, TB remains a disastrous global health threat ranking alongside the human immunodeficiency virus (HIV) as the second killer worldwide [1]. There were an estimated 9.6 million new TB cases in 2014, and China is bearing 10 percentage of the global total according to the latest World Health Organization (WTO) report [2]. To reverse the severe situation, constant efforts on developing effective vaccines as well as controlling the spread of MTB are in long term needed. Since delayed or incorrect diagnosis of TB markedly aggravates the epidemic of M. tuberculosis, it is urgent to develop method for credible early-stage diagnosis of TB and comprehensive understanding of the intrinsic pathogenesis of M. tuberculosis, especially the multidrug-resistant strains of M. tuberculosis (MDR-TB).

Recent researches suggest that secretory protein antigens can be used to detect antibodies in infected specimens [3]. And it is a well-established fact that effector proteins are mostly secretory proteins that stimulate infection by manipulating the host response [4]. In line with nonsecretory proteins, M. tuberculosis secretory proteins are also assembled in ribosomes but transported to extracellular milieu through certain specific secretion system to attack host cells for survival and reproduction, thus changing the host cell microenvironment and causing grievous tubercular symptoms in infected individuals. Hence, distinguishing secretory proteins from nonsecretory proteins is a matter of grave concern for tracing the real pathogenic factors and developing vaccines or drugs against TB.

Benefited from the next-generation sequencing technology, the amount of protein sequences is exploding with an exponential growth, which is far beyond the capacity of classical biochemical analysis through advances in experimental facilities and molecular biotechnologies. Faced with such an embarrassment, appropriate and fast computational methods exploit a fresh avenue to investigate the properties of M. tuberculosis proteins. However, to the best of our knowledge, few computational works focused on secretory proteins in M. tuberculosis. Since a majority of secretory proteins have a signal peptide by which they were exported via the signal peptidase pathway, in 2009, Leversen et al. evaluated the performance of nine signal peptide prediction algorithms for identification of mycobacterial signal peptides using sequence data from proteomic methods [5]. Sequential and structural characteristics of secretory proteins also help to address the problem. In 2010, Vizcaíno et al. identified potential secretory proteins from Mycobacterium tuberculosis H37Rv by screening its genomes using machine-learning tools [6]. The prediction of subcellular locations of M. tuberculosis proteins also provides vital clue for identifying secretory proteins. Recently, Zhu et al. used a support vector machine- (SVM-) based model to predict the subcellular localization of mycobacterium proteins [7]. The success on mycobacterium proteins as mentioned above suggested that bioinformatics approaches are effective strategy for mining useful information and providing insights into both basic research and drug design.

Thus, the current study was devoted to develop a bioinformatics approach to identify secretory proteins in M. tuberculosis. The work will describe how to construct a valid benchmark dataset, how to develop an effective mathematical expression to formulate the proteins, and what kinds of machine-learning method and cross-validation tests were used in the prediction model. Finally, based on the proposed method, a web server called MycoSec was established.

2. Material and Methods

2.1. Benchmark Datasets

The original datasets used in this study were extracted from the Universal Protein Resource (UniProt) [8]. To guarantee a good quality of data, proteins in UniProt were collected confidently according to the following criteria: (I) only those from M. tuberculosis were considered; (II) only those reviewed and annotated by experts were chosen; (III) sequences with ambiguous residues, such as “,” “,” and “,” were discarded; (IV) sequences that were inferred from homologous proteins were eliminated; (V) sequences that were fragments of other proteins were excluded; (VI) sequences that have less than 16 amino acids were removed to meet the parameter () requirement (see Section 3.1); (VII) sequences with the keyword “secreted” or “secretory vesicle” in the “subcellular location” column were regarded as secretory proteins (positive samples), while sequences without these keywords were considered as nonsecretory proteins (negative samples or control samples). After going through the above processes, a total of 420 samples were reserved, including 63 positives and 357 negatives. To exclude any homologous bias which may cause overestimation problem of final prediction results, we clustered these protein sequences with the CD-HIT program [9] by setting 30%, a rigorous value of this parameter, as the cutoff threshold of sequence identity to remove the redundant part. As a result, the benchmark dataset containing 35 secretory proteins and 266 nonsecretory proteins can be expressed aswhere represented the positive subset, represented the negative subset, and the symbol “” represents “union” in the set theory. More detailed information of the datasets can be downloaded from

2.2. The Representation of Protein Samples

Translating conventional biological problems into computable mathematical models is usually adopted as an accommodation to the requirement of bioinformatics analysis. A protein sequence with total number of amino acids is usually formulated aswhere    denotes the th amino acid residue in the query protein sample , of which is the length of the protein. Almost all the existing machine-learning methods, regardless of supervised and unsupervised or semisupervised, such as SVM [1013], Artificial Neural Network (ANN) [14], -Nearest Neighbor (KNN) [15], and ensemble classifiers [1618], can only handle vectors with the same dimension rather than sequence samples [19]. Meanwhile, data discretization is also required by mainstream feature selection strategies [2022]. Accordingly, the concept of discrete vector is proposed to realize more general representations of sequence fragments.

The pseudo amino acid composition (PseAAC) is a widely used method for representing protein sequences for its annexation of long-range sequence-order information and the correlation of physicochemical properties between two residues, as well as its balance between representative capability and computational expense. Inspired by PseAAC, we made an improvement by substituting amino acid composition for -gap dipeptide composition. Thus, each protein sequence in our benchmark dataset can be denoted by a dimension vector; that is,where the first 400 elements are the -gap dipeptide composition and the next elements are the first tire to th tire correlation factors of protein sequence, which are determined on the basis of physiochemical properties. “” is a symbol of transpose operator.    can be calculated as given inwhere is the occurrence frequency of the th dipeptide in , as formulated by is how many times the th -gap dipeptide appears in . is the weight coefficient. is the number of physicochemical properties selected. is the number of total counted ranks or tires of the correlations along a protein sequence and    is the th tire correlation factor that reflects the sequence-order correlation between all the th most contiguous dipeptides along a protein sequence, as formulated by the following formula:where the correlation function is calculated by where denotes the constant value of the th kind physicochemical property for and denotes the th physiochemical property value for . Due to the different scales of values among various properties, the following normalization process is necessary to attain nondimensional data:where () is the original value of the th kind physicochemical property for , stands for the th category amino acid residue, and is defined in the same manner as . The normalized values obtained by (8) will have a zero mean value over the 20 kinds of amino acids and will remain unchanged if going through the same conversion procedure again.

It is widely accepted that physicochemical properties of proteins have impacts upon not only a series of biological processes [23], such as protein denaturation and renaturation, cell signaling transduction, and change of solution conductivity, but also maintaining tertiary structures, further underlying certain molecular functions. Among various properties, hydrophilicity and hydrophobicity were chosen here on the basis of a priori knowledge that water is the environment for the survival of all biological molecules and promotes the transportation of molecules by an intrinsic liquidity. In our model, the tire correlation of hydrophilic and hydrophobic amino acids was investigated by tuning the parameter in order to see the global placement in which amino acids were linked to each other.

2.3. Support Vector Machine

SVM is a welcome and powerful machine-learning algorithm which has been successfully used in the realm of bioinformatics for pattern recognition and classification. Established on the theories of Vapnik-Chervonenkis Dimension [24] (VC Dimension) and structural risk minimization, SVM can solve binary nonlinear classification problems among small samples and achieve good generalization ability by mapping the input data into higher dimensional feature space (Hibert space) by kernel function transformation and then determining an optimal hyperplane to separate a given set of labeled data. For multiclass classification tasks, “one-versus-one (OVO)” and “one-versus-rest (OVR)” are generally applied to extend the traditional SVM. A brief description of formulation of SVM is given in [25, 26] and, for more details, please see a monograph [27].

In this work, the toolbox LIBSVM 3.20 was employed to implement the SVM classifier and perform the prediction, which can be freely downloaded from The feature vectors of protein samples formulated by (3) were used as inputs of the SVM. Radial basis function (RBF) was adopted here on account of better validity, less deviation, and faster speed in nonlinear training process compared with other kernel functions. For pursuit of the optimal model, the penalty constant and the kernel width parameter were tuned in an optimization procedure using a grid search method, of which the search spaces for and are and with steps of and 2, respectively.

2.4. Performance Evaluation

Here we attempted to evaluate the performance of a statistical predictor from two parts: (a) evaluating its generalization ability with an impartial cross-validation method; (b) employing appropriate metrics to measure its success rate.

Independent dataset test, subsampling (or k-fold cross-validation) test, and jackknife test are the three widely used validation methods to evaluate the anticipated success rate of a predictor [28]. Among the three methods, jackknife test is deemed the least arbitrary due to its basic idea of leaving out each sample as the testing set iteratively and then finding the mean value of these calculations. However, compared to the other two methods, jackknife test is so time-consuming that it lags the overall computational efficiency. Thus, to balance between correctness and efficiency, a trade-off was made by applying fivefold cross-validation test in the stage of parameter optimization and then switching to jackknife test for evaluation of final model once the optimal model was found.

Here, a set of straightforward methods was provided to assess the prediction quality by using the following three metrics: sensitivity (Sn) which is also known as recall, specificity (Sp), and average accuracy (AA), which are defined as follows: In (9), TP, TN, FP, and FN are the number of true positives, true negatives, false positives, and false negatives. Sn indicates the ability of correctly judging positive samples, Sp suggests the ability of correctly recognizing negative samples, and Ac averages Sn and Sp.

The Receptor Operating Characteristic curves [29] (ROC curves) were also plotted to describe the performance of models by plotting the Sn (true positive rate (TPR)) against the 1-Sp (false positive rate (FPR)) across the entire range of SVM decision values, under which the area (AUC) can serve as an objective indicator for quality assessments even when the classes are of very different sizes. The value 0.5 of AUC is equivalent to random prediction while 1 represents a perfect one. Accordingly, the larger the AUC is, the more credibility the predictions have.

2.5. Feature Selection

Feature selection is very important in view of dimension reduction for pinpointing distinguishing features and then improving generalization ability [30]. In the present study, we performed feature selection by using the Analysis of Variance (ANOVA), which is known as a robust and simple method to test the difference in means between groups, even when the number of observations is uneven in each group. Besides, it is easily generalized to more than two groups without increasing the Type 1 error. According to the basic idea of ANOVA, features can be ranked by their F-values calculated as follows:where and denote the sample variance between groups (also called Mean Square Between, MSB) and sample variance within groups (also called Mean Square Within, MSW) separately; the detailed formulae were given inwhere is the value of the th feature of the th sample in the th group defined in (4) and is the sample size of each group (here and ). For only two groups were counted in the current study, the degree of freedom (DOF) between and within groups was 1 and , respectively. Still, it is apparent that a larger value of reflects a better discriminative capability of a feature.

Based on the features thus ranked, we used the incremental feature selection [31] (IFS) to determine the optimal number of features as described below. The first feature subset was initialized by a feature with the highest F-value, and the next subset was composed when the feature with the second highest F-value was added. We repeated this process by adding features sequentially from higher to lower F-values until all candidate features were added. Thus, the feature sets thus formed would be composed of ranked features. The th feature set can be formulated as

For each of such feature sets, an SVM prediction model was constructed and examined by the jackknife test on the benchmark data set. Afterwards, we obtained an IFS curve in a 2D Cartesian coordinate system with index ζ as the abscissa (or X-coordinate) and the AUC as the ordinate (or Y-coordinate). The optimal feature set is expressed aswith which the IFS curve reaches its peak. In other words, in the 2D coordinate system, the AUC value reaches its maximum when .

3. Results and Discussions

3.1. Parameter Optimization

As we can see from (3) to (8), the results of the proposed method depended on three parameters, that is, , and . represents the tiers counted for the global or long-range sequence-order effect and larger may contain more global sequence-order information; portrayed the local or short sequential tendencies by measuring the gap between two amino acid residues; and is the weight factor imposed between local and global effects which is usually within the limits of 0 to 1. To search for the optimal values of the three parameters which can achieve the highest accuracy, we performed a series of experiments according to the following standard:Hence, a total of individual combinations (or points in the 3D parameter space) had to queue up to be screened for finding the optimal one, which was actually a routine but tedious process to optimize the model via a 3D grid search [32]. In view of computational efficiency, fivefold cross-validation was firstly employed to cope with the prioritization of parameters, and once the optimal values were determined, the rigorous jackknife test was performed to evaluate the success rates of the feature set according to the four metrics defined in Performance Evaluation section. As a result, the largest AUC of 0.845 was obtained when , , and .

The dimension of the optimal feature set was , which means that 400 9-gap dipeptide compositions and 12 additional hydrophilicity/hydrophobicity components would be incorporated in a feature vector fed into the predictor. Nevertheless, to evade from noise and overfitting problems, ANOVA was employed to further optimize the feature set. By calculating the F-value of each candidate feature and ranking them in descending order, a series of feature sets in various sizes were obtained based on IFS strategy as illustrated in Feature Selection section. The IFS curve reached its peak when the feature number was 374, indicating that the feature subset was the least redundant and the most discriminative one. Consequently, the AUC was improved from 0.85 to 0.93 and the computational time and expenses were acutely reduced.

3.2. Comparison with Other Classifiers

Subsequent to the determination of optimum parameters and feature subset, we further compared the performance between various classifiers, namely, SVM, Bayes Net, Radial Basis Function Network (RBF Network), and Random Forest, by applying the same sequence and feature data under the same parameters. The software WEKA (version 3.8) ( was used to implement the last three classifiers. Likewise, calculations were made in line with the four metrics illustrated in Performance Evaluation Section. The results are summarized in Table 1. From Table 1, although relative similar Sns were produced by SVM (94.29%), Bayes Net (82.86%), and RBF Network (85.71%), the differences in other three metrics among the four classifiers were conspicuous enough to pronounce judgment. Nevertheless, Sn and Sp may lose objectivity because of an imbalance in class sizes, and even when all samples were misjudged as negative, the Sp and overall accuracy (OA) would still hang in high levels. Therefore, we attached more importance to Ac instead of Sn or Sp. As we can see from Table 1, SVM also achieved the maximum Ac with 87.18%, followed by Bayes Net (69.25%) and Random Forest (65.15%).

AlgorithmSn (%)Sp (%)AA (%)AUC

Random Forest45.7184.5965.150.69
Bayes Net82.8655.6469.250.66
RBF Network85.7136.0960.900.59

In fact, all the metrics mentioned above are static assessment indicators, except AUC, which emerges as a dynamic measure depending on ROC and remains objective under class imbalance conditions. Thus, AUC was chosen as the foremost evaluation criteria to compare performance disparities among the four algorithms. To visually view the dynamic changing process of Sn (also known as true positive rates or TPR) of a certain algorithm versus its Sp (also known as true negative rates or TNR), four curves corresponding to four classifiers were plotted in Figure 1 on the basis of their confusion matrixes varying with a range of thresholds. As shown in Figure 1, the SVM gave an AUC of 0.93 in the discrimination between secretory proteins and nonsecretory proteins, which is higher than that of the other three classifiers. Furthermore, the tendency of SVM curve was steeper and ran much closer to the vertical axis than that of the others, which signified a better performance in ROC curve. Thus, compared with the other three state-of-the-art classifiers, SVM is much more powerful and robust for classifying secretory and nonsecretory proteins. Two main reasons may lead to the result. SVM is insensitive to the number and probability distribution of training samples, along with the dimension of input space. Furthermore, SVM can gain the global optimal solution of goal function on the basis of convex optimization theory, while other classifiers based on greedy learning strategies were usually trapped into locally optimal solutions.

3.3. Web Server Construction

For the convenience of a vast majority of experimental scientists, a user-friendly web server, namely, MycoSec, was established, which is freely accessible at Below is a step-by-step guide on how to use the web server.

First, reasonable and clear as you can see in Figure 2, the homepage provides several buttons and a textbox lying to the center of the whole. A brief introduction of the predictor and the caveat when using it would be present on clicking on the Read Me button.

Second, either type/paste the query protein sequences into the input box or select a FASTA file from a certain directory, ensuring that each sequence is longer than 11 amino acid residues and without any ambiguous characters. Example sequences in FASTA format can be seen by clicking on the Example button right above the input box.

Third, click on the Submit button to see the predicted results present in a new webpage. For a query sequence, the first column of the results shows its actual label, the second gives the probability that it belongs to secretory protein class, and the third is the probability that it pertains to nonsecretory protein class.

Fourth, the benchmark data used in our analysis can be downloaded by clicking on the Data button for training and testing our system or your own model.

Finally, click on the Citation button to find the relevant papers that document the detailed development and algorithm of MycoSec.

4. Conclusions

Identification of secretory proteins in M. tuberculosis is of paramount significance for targeting of antigens and vaccine development, which may contribute to an early diagnose or cure of the terrible disease tuberculosis. In this paper, we proposed a SVM-based algorithm to identify secretory proteins in M. tuberculosis. In this method, an improved pseudo amino acid composition which integrates -gap dipeptides along with physicochemical properties was proposed and used to encode the residue sequences. By parameters search and feature optimization, we obtained the optimal model with an AUC of 0.93 validated by jackknife test. Based on this model, an online predictor MycoSec was established for the convenience of researchers in relevant fields. We hope the predictor will be a smart tool in M. tuberculosis research. In the future, we will combine other features such as evolution information to improve classification performance.

Competing Interests

The authors declare no competing financial interests.


This work was supported by the National Nature Scientific Foundation of China (no. 61301260), the Applied Basic Research Program of Sichuan Province (nos. 2015JY0100 and LZ-LY-45), the Scientific Research Foundation of the Education Department of Sichuan Province (11ZB122), the Nature Scientific Foundation of Hebei Province (no. C2013209105), the Fundamental Research Funds for the Central Universities of China (no. ZYGX2015J144), the Outstanding Youth Foundation of North China University of Science and Technology (no. JP201502), and Program for the Top Young Innovative Talents of Higher Learning Institutions of Hebei Province (no. BJ2014028).


  1. J. Zhang, H. Gou, X. Hu et al., “Status of drug-resistant tuberculosis in China: a systematic review and meta-analysis,” American Journal of Infection Control, vol. 44, no. 6, pp. 671–676, 2016. View at: Publisher Site | Google Scholar
  2. World Health Organization, Global Tuberculosis Report 2015, WHO, 2015.
  3. D. Tiwari, S. Haque, R. P. Tiwari, A. Jawed, T. Govender, and H. G. Kruger, “Fast and efficient detection of tuberculosis antigens using liposome encapsulated secretory proteins of Mycobacterium tuberculosis,” Journal of Microbiology, Immunology and Infection, 2015. View at: Publisher Site | Google Scholar
  4. H. Sonah, R. K. Deshmukh, and R. R. Bélanger, “Computational prediction of effector proteins in fungi: opportunities and challenges,” Frontiers in Plant Science, vol. 7, article 126, 2016. View at: Publisher Site | Google Scholar
  5. N. A. Leversen, G. A. de Souza, H. Målen, S. Prasad, I. Jonassen, and H. G. Wiker, “Evaluation of signal peptide prediction algorithms for identification of mycobacterial signal peptides using sequence data from proteomic methods,” Microbiology, vol. 155, no. 7, pp. 2375–2383, 2009. View at: Publisher Site | Google Scholar
  6. C. Vizcaíno, D. Restrepo-Montoya, D. Rodríguez et al., “Computational prediction and experimental assessment of secreted/surface proteins from Mycobacterium tuberculosis H37Rv,” PLoS Computational Biology, vol. 6, no. 6, Article ID e1000824, 2010. View at: Publisher Site | Google Scholar
  7. P.-P. Zhu, W.-C. Li, Z.-J. Zhong et al., “Predicting the subcellular localization of mycobacterial proteins by incorporating the optimal tripeptides into the general form of pseudo amino acid composition,” Molecular BioSystems, vol. 11, no. 2, pp. 558–563, 2015. View at: Publisher Site | Google Scholar
  8. A. Bairoch, R. Apweiler, C. H. Wu et al., “The universal protein resource (UniProt),” Nucleic Acids Research, vol. 33, pp. D154–D159, 2005. View at: Publisher Site | Google Scholar
  9. W. Li and A. Godzik, “Cd-hit: a fast program for clustering and comparing large sets of protein or nucleotide sequences,” Bioinformatics, vol. 22, no. 13, pp. 1658–1659, 2006. View at: Publisher Site | Google Scholar
  10. C. Cortes and V. Vapnik, “Support-vector networks,” Machine Learning, vol. 20, no. 3, pp. 273–297, 1995. View at: Publisher Site | Google Scholar | Zentralblatt MATH
  11. D. Li, Y. Ju, and Q. Zou, “Protein folds prediction with hierarchical structured SVM,” Current Proteomics, vol. 13, pp. 79–85, 2016. View at: Google Scholar
  12. R. Wang, Y. Xu, and B. Liu, “Recombination spot identification based on gapped k-mers,” Scientific Reports, vol. 6, Article ID 23934, 2016. View at: Google Scholar
  13. J. Chen, X. Wang, and B. Liu, “IMiRNA-SSF: improving the identification of MicroRNA precursors by combining negative sets with different distributions,” Scientific Reports, vol. 6, article 19062, 2016. View at: Publisher Site | Google Scholar
  14. S.-C. Wang, “Artificial neural network,” in Interdisciplinary Computing in Java Programming, pp. 81–100, Springer, New York, NY, USA, 2003. View at: Google Scholar
  15. L. E. Peterson, “K-nearest neighbor,” Scholarpedia, vol. 4, no. 2, article 1883, 2009. View at: Publisher Site | Google Scholar
  16. Q. Zou, J. Guo, Y. Ju, M. Wu, X. Zeng, and Z. Hong, “Improving tRNAscan-SE annotation results via ensemble classifiers,” Molecular Informatics, vol. 34, no. 11-12, pp. 761–770, 2015. View at: Publisher Site | Google Scholar
  17. C. Lin, W. Chen, C. Qiu, Y. Wu, S. Krishnan, and Q. Zou, “LibD3C: ensemble classifiers with a clustering and dynamic selection strategy,” Neurocomputing, vol. 123, pp. 424–435, 2014. View at: Publisher Site | Google Scholar
  18. B. Liu, S. Wang, Q. Dong, S. Li, and X. Liu, “Identification of DNA-binding proteins by combining auto-cross covariance transformation and ensemble learning,” IEEE Transactions on NanoBioscience, 2016. View at: Publisher Site | Google Scholar
  19. B. Liu, F. Liu, X. Wang, J. Chen, L. Fang, and K. Chou, “Pse-in-One: a web server for generating various modes of pseudo components of DNA, RNA, and protein sequences,” Nucleic Acids Research, vol. 43, pp. W65–W71, 2015. View at: Publisher Site | Google Scholar
  20. I. Jolliffe, Principal Component Analysis, Wiley Online Library, 2002.
  21. P. Rivett, “Multidimension scaling for multiobjective policies,” Omega, vol. 5, no. 4, pp. 367–379, 1977. View at: Publisher Site | Google Scholar
  22. L. Wei, M. Liao, X. Gao, and Q. Zou, “Enhanced protein fold prediction method through a novel feature extraction technique,” IEEE Transactions on Nanobioscience, vol. 14, no. 6, pp. 649–659, 2015. View at: Publisher Site | Google Scholar
  23. J. F. Zayas, “Introduction,” in Functionality of Proteins in Food, pp. 1–5, Springer, New York, NY, USA, 1997. View at: Publisher Site | Google Scholar
  24. V. Vapnik, E. Levin, and Y. Le Cun, “Measuring the VC-dimension of a learning machine,” Neural Computation, vol. 6, no. 5, pp. 851–876, 1994. View at: Publisher Site | Google Scholar
  25. K.-C. Chou and Y.-D. Cai, “Using functional domain composition and support vector machines for prediction of protein subcellular location,” The Journal of Biological Chemistry, vol. 277, no. 48, pp. 45765–45769, 2002. View at: Publisher Site | Google Scholar
  26. Y.-D. Cai, G.-P. Zhou, and K.-C. Chou, “Support vector machines for predicting membrane protein types by using functional domain composition,” Biophysical Journal, vol. 84, no. 5, pp. 3257–3263, 2003. View at: Publisher Site | Google Scholar
  27. T. S. Furey, N. Cristianini, N. Duffy, D. W. Bednarski, M. Schummer, and D. Haussler, “Support vector machine classification and validation of cancer tissue samples using microarray expression data,” Bioinformatics, vol. 16, no. 10, pp. 906–914, 2000. View at: Publisher Site | Google Scholar
  28. K.-C. Chou and C.-T. Zhang, “Prediction of protein structural classes,” Critical Reviews in Biochemistry and Molecular Biology, vol. 30, no. 4, pp. 275–349, 1995. View at: Publisher Site | Google Scholar
  29. A. P. Bradley, “The use of the area under the ROC curve in the evaluation of machine learning algorithms,” Pattern Recognition, vol. 30, no. 7, pp. 1145–1159, 1997. View at: Publisher Site | Google Scholar
  30. Q. Zou, J. Zeng, L. Cao, and R. Ji, “A novel features ranking metric with application to scalable visual and bioinformatics data classification,” Neurocomputing, vol. 173, pp. 346–354, 2016. View at: Publisher Site | Google Scholar
  31. H. Liu and R. Setiono, “Incremental feature selection,” Applied Intelligence, vol. 9, no. 3, pp. 217–230, 1998. View at: Publisher Site | Google Scholar
  32. H. Lin, E.-Z. Deng, H. Ding, W. Chen, and K.-C. Chou, “iPro54-PseKNC: a sequence-based predictor for identifying sigma-54 promoters in prokaryote with pseudo k-tuple nucleotide composition,” Nucleic Acids Research, vol. 42, no. 21, pp. 12961–12972, 2014. View at: Publisher Site | Google Scholar

Copyright © 2016 Huan Yang et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Related articles

No related content is available yet for this article.
 PDF Download Citation Citation
 Download other formatsMore
 Order printed copiesOrder

Related articles

No related content is available yet for this article.

Article of the Year Award: Outstanding research contributions of 2021, as selected by our Chief Editors. Read the winning articles.