Research Article  Open Access
Predicting βTurns in Protein Using Kernel Logistic Regression
Abstract
A βturn is a secondary protein structure type that plays a significant role in protein configuration and function. On average 25% of amino acids in protein structures are located in βturns. It is very important to develope an accurate and efficient method for βturns prediction. Most of the current successful βturns prediction methods use support vector machines (SVMs) or neural networks (NNs). The kernel logistic regression (KLR) is a powerful classification technique that has been applied successfully in many classification problems. However, it is often not found in βturns classification, mainly because it is computationally expensive. In this paper, we used KLR to obtain sparse βturns prediction in short evolution time. Secondary structure information and positionspecific scoring matrices (PSSMs) are utilized as input features. We achieved of 80.7% and MCC of 50% on BT426 dataset. These results show that KLR method with the right algorithm can yield performance equivalent to or even better than NNs and SVMs in βturns prediction. In addition, KLR yields probabilistic outcome and has a welldefined extension to multiclass case.
1. Introduction
The number of known protein sequence is increasing rapidly as a result of genome and other sequencing projects. Consequently, this increase widens sequencestructure gap rapidly [1, 2]. Thus computational tools for predicting protein structure and function are highly needed to narrow the widening gap [3]. There are four distinct levels of protein structures. These levels are primary structure which refers to amino acid linear sequence of the polypeptide, secondary structure, which is defined by the patterns of hydrogen bonds between backbone amide and carboxyl groups, tertiary structure, which is the threedimensional structure of a single protein molecule, and quaternary structure, which is a larger assembly of several protein molecules or polypeptide chains.
The basic elements of the secondary structure of proteins are helices, sheets, coils, and turns. A turn is a structural motif where the atoms of two residues are separated by few (usually from 1 to 5) peptide bonds, and the distance between them is less than , while the corresponding residues do not form a regular secondary structure element such as an helix or sheet. Different turns are classified according to the separation between the two end residues. The end residues are separated by four peptide bonds in turns, three peptide bonds in turns, two peptide bonds in turns, one bond in turns, and five bonds in turns. turns are the most common found type of turns that constitute approximately 25% of the residues in protein. They play a significant role in protein configuration and function, and its formation is a vital stage during the protein folding. They were found to be more helpful in the context of molecular recognition and in modeling interactions between peptide substrates receptors, because they tend to be more solvent exposed than buried [4]. In the recent years it has been found that turns are important in the design of various peptidomimetics for many diseases [5]. Therefore, development of effective and efficient prediction methods for turns identification in protein is useful in fold recognition and drug design [6].
turns are further classified into different types according to the dihedral angles () of the central two residues. The classification scheme proposed by Hutchinson and Thornton [7] recognizes nine distinct types of turn: I, I′, II, II′, VIa1, VIa2, VIb, VIII, and IV. In this classification, the most frequently occurring type is type IV, which constitutes approximately (35%) of the turns. Types VIa1,VIa2, and VIb are rare types.
Most of the successful turns prediction methods are based on either support vector machines (SVMs) or neural networks (NNs). Zheng and Kurgan [6] applied SVMbased ensemble to predict turns. They used positionspecific scoring matrices (PSSMs) and secondary structure information as features in their prediction model. Kountouris and Hirst [8] developed a method based on SVM; their method uses PSSMs, predicted secondary structures, and predicted dihedral angles as input features to the SVM. Shepherd et al. [9] used a neural network to predict both the location and types of turn in protein; they incorporated secondary structure information on the features used as input to the NN. Kaur and Raghava [10] used two feedforward backpropagation networks with a single hidden layer, where the first sequencestructure network is trained with the PSSMs. The initial prediction from the first network and the predicted secondary structure using PSIPRED [11, 12] are used as input to the second structurestructure network to refine the prediction obtained from the first network. Petersen et al. [13] presented a neural network method called NetTurnP, for predicting turns and turn types. Their method consists of two artificial neural network layers; they used PSSMs, secondary structure, and surface accessibility as input to their model.
There is another method that can perform well as SVMs and NNs, which is the Kernel Logistic Regression (KLR). KLR is a kernel version of logistic regression (LR). It is often not found in predicting protein secondary structures and turns due to its computational demand. However, unlike SVMs and NNs, KLR yields a posteriori probabilities based on a maximum likelihood argument that is, besides predicting class labels, KLR provides interpretation about this labeling. When it comes to turn types prediction, KLR has an additional advantage that its extension to multiclass classification is well described. Karsmakers [14, 15] proposed a fast and accurate approximate implementation of KLR for Automatic Speech Recognition (ASR). He described a different practical technique suited for large datasets, based on fixedsize least squares support vector machines (FSLSSVMs), of which he named fixedsize kernel logistic regression (FSKLR). Karsmakers used trust region Newton’s method for largescale LR [16] as a basis to solve the approximate problem and Nystrom method to approximate the features’ matrix. In this paper, we show that FSKLR can be used in predicting turns in an efficient and effective way, and it yields results that are comparable to the stateoftheart methods.
2. Methods
2.1. Data Sets
The uniform dataset of 426 nonhomologous proteins (BT426) [17], the dataset of 547 protein sequence (BT547), and the dataset of 823 protein sequence (BT823) are used to evaluate the performance of our KLR method. Several researchers used BT426 as a golden set of sequences upon which performance values are reported and compared. This dataset consists of protein chains whose structure has been determined by Xray crystallography at a resolution of or better. Each chain contains at least one turns region. In total 23,580 amino acids, corresponding to 24.9% of all amino acids, have been assigned to be located in turns. None of the sequences in the dataset shares more than 25% sequence identity. BT426 has been used by various recent turns prediction methods; therefore, we can use it to make direct comparisons with these methods. The other two datasets: BT547 and BT823 are constructed for training and testing COUDES [18].
2.2. Features Vector
The features that are used in this study include PSSMs and secondary structure information.
2.2.1. PSSMs
Several studies show that PSSMs contributed significantly to the accuracy of turns prediction [6, 13]. The PSSMs are in the form of , whererepresents the sequence length. The PSSMs were generated using the iterative PSIBLAST program [19] against National Center for Biotechnology Information (NCBI) nonredundant (nr) sequence database using the default parameters. The PSSMs values are scaled to values between 0 and 1. A window size of seven residues is used for the PSSMs. This is in accordance with Shepherd et al. [9] who found that the optimal prediction for turns is achieved using window size of seven or nine. The total number of the features that are based on PSSMs is ().
2.2.2. Secondary Structure Information
For the secondary structure information features, four secondary structure prediction methods are utilized for all protein chains. These four prediction methods are PSIPRED [12, 20], JNET [21], TRANSEC [22], and PROTEUS [22]. The secondary structures were predicted as three structure states: helix (H), strand (E), and coils (C). These three structure states are encoded as 1 0 0 for helix, 0 1 0 for strand, and 0 0 1 for coils. The secondary structure information features are organized as follows: a binary value denoting the prediction of a given secondary structure method from the aforementioned used prediction methods for the central residue; that is, if PSIPRED predicted the central amino acid to be helix, JNET predicted it to be coil, TRANSEC predicted it to be helix, and PROTEUS predicted it to be helix, then this binary value will be 1 0 0 0 0 1 1 0 0 1 0 0, so the total number of features using this organization is 12; the confidence value obtained from the central residue using the four prediction methods. The confidence score is divided by 10 to normalize it to a unit interval, and the total number of features using this organization is 4. A binary value denoting a specific configuration of the secondary structure is predicted using the four prediction methods for the central and the two adjacent residues. Here we have four patterns 1, 2, 3, and 4. If the predicted secondary structure using specific method is coils 0 0 1, the secondary structure for the pattern 1 will be CCC and for pattern 2, 3, and 4 will be CCX, XCC, and XCX, respectively, where X = E, H. The total number of features based on this organization is (4 patterns 3 secondary structures 4 prediction methods = 48 features). The ratio between the number of residues in a given secondary structures and the window size for the four prediction methods, the number of features based on this organization will be (3 secondary structure 4 prediction methods = 12). The total number of features based on secondary structure information is 76. The motivation to use this organization comes from [6].
The predicted secondary structure information is added to the PSSMs features. The total number of the features that are based on PSSMs and secondary structure information is 216. Similarly as in [6], feature’s selection methods based on information gain and CHI squared are employed to reduce the number of features to 90 features. Figure 1 shows the overall architecture of our KLR method.
2.3. Prediction Method
The fixedsize kernel logistic regression (FSKLR) was applied to predict turns. KLR is the kernel version of LR, which is a wellknown statistical model for classification. Unlike LR, KLR enables the classification of linearly nonseparable problems by transferring the input features to a higherdimensional space, via the kernel trick. The kernel is a transformation function that must satisfy Mercer’s necessary and sufficient conditions, which state that a kernel function must be expressed as an inner product and must be positive semidefinite. Similar to LR, KLR can be fitted using the maximum likelihood estimate (MLE).
Iteratively reweighted least square (IRLS) algorithm is one of the most popular techniques used to find the MLE of the LR models. IRLS is a nonlinear optimization algorithm that uses a series of weighted least squares (WLS) subproblems to search for the MLE. It is a special case of Fisher’s scoring method, a quasiNewton algorithm that replaces the objective function’s Hessian with the Fisher information. For LR, IRLS is a special form of Newton’s method in which each iteration finds the WLS estimates for a given set of weights, which are used to construct a new set of weights. KLR also can be fitted effectively using IRLS [15].
Unlike SVMs, KLR does not use risk minimization principle, but it is based on conditional maximum likelihood inference, which results in estimates of a posteriori class probabilities via logit stochastic models: where , is the vector of the KLR parameters, and is the intercept. The penalized negative log likelihood (PNLL) is normally used to infer the parameters of the KLR model. In the primal weight space, the objective function for the PNLL is as follows: where is the regularization parameter. The solution can be expressed in terms of and computed using IRLS iteration as
In the dual representation, the function values in the KLR logit models can be computed as follows: where .
The IRLS method is suitable for small size problems, but for largescale problems this method becomes computationally expensive. Based on fixedsize least squares support vector machines (FSLSSVMs) Karsmakers [14, 15] implemented a fixedsize variant of the standard KLR formulation (FSKLR) which does easily scale to very large datasets. In his method, he adopted Nystrom approximation method.
In Nystrom approximation, the kernel matrix will be decomposed into eigenvalues/eigenvectors matrices in the form: where = diag() and are the eigenvalues of the matrix K, is the matrix of the eigenvectors that correspond to the eigenvalues, and is the number of the data points. We can select the first eigenvectors and eigenvalues from the matrices and , respectively, where , to approximate the kernel matrix. This approximation is motivated by its widely usage, for example, principal component analysis. Using this approximation reduces the computational cost drastically. However, computing the eigendecomposition is also computationally expensive. To reduce the computational cost of computing the eigendecomposition we selected a small sample of size from the features’ matrix to construct the following eigenproblem:
We can extend the eigenvalues/eigenvectors of the to all the points using the following Nystrom approximation: where and are the ith eigenvalue/eigenvector of the eigenproblem and is the appropriate submatrix of .
The selected sample of size from the features’ matrix can be called prototype vectors (PVs). These PVs can be selected using kcenter clustering. The use of kcenter clustering is justified in [23], which observed that the Nystrom lowrank approximation depends crucially on the quantization error induced by encoding the sample set with landmark points. This suggests that one can simply use the clusters obtained with a kcenter (such as kmeans) algorithm, which finds a local minimum of the quantization error. The PVs selection methods using kcenter clustering suffer from the fact that they will select outliers as prototypes. In cases where the number of PVs is relatively small, the fraction of prototypes chosen to represent the nonoutlier and outlier data is unbalanced, and, therefore, the classification performance will not be optimal. When the number of PVs is increased, the performance will also increase to that of KLR. Hence taking into account outliers removal can result in a sparser model. The sparse kernel logistic regression problem is solved in the primal space using Newton’s trust region algorithm, which is given in [16]. This algorithm yielded the best performance compared to the stateoftheart alternatives. Convergence speed and cost per iteration will be balanced in that lowcost approximate because Newton’s steps will be taken in the beginning of the algorithm and full Newton directions at the end for fast convergence. In this paper, the following radial basis function (RBF) is used as a kernel function: where is the kernel parameter.
2.4. Model Selection
Model selection is the process of determining the optimal regularization parameter and the kernel parameter . It is a very important step in fitting kernel models to maximize generalization performance. The crossvalidationbased method is used to determine the optimal parameters for turns prediction.
2.5. Training and Testing
In order to evaluate a prediction method it is necessary to have different datasets for training and testing. The jackknife test is the most objective and rigorous crossvalidation method compared with independent dataset test and subdataset test [24]. In a full jackknife test of proteins, one protein is removed from the set; the training is done on the remaining proteins, and the test is done on the removed protein. This process is repeated times by removing one protein in turn. Since this training technique is very time consuming most of the recent βturns prediction methods use sevenfold crossvalidation to assess their performances. We also used sevenfold crossvalidation to assess the accuracy of FSKLR. In sevenfold crossvalidation, the datasets will be divided into seven subsets, each containing equal number of proteins. Each set is an unbalanced set that retains the naturally occurring proportion of βturns. Six of the seven subsets were merged together to form a training set that was used to train the FSKLR methods, and the seventh was used for validation. This process was repeated seven times in order to have a different set for validation each time. The final prediction results are taken as the average of the results from the seven testing sets.
2.6. Performance Measures
The quality of prediction is evaluated using five measures: MCC, , , , and Specificity. These measures are consistent with the test procedures and measures applied to evaluate competing methods. Let (true positives) TP be the number of correctly classified turns residues, (true negatives) TN be the number of correctly classified nonturns residues, (false positives) FP be the number of nonturns incorrectly classified as turns residues, and (false negatives) FN be the number of turns incorrectly classified as nonturns residues. The Matthews correlation coefficient (MCC) can be calculated as [25]
The result of MCC is in the range of −1 and 1, where a value of 1 indicates a perfect positive correlation, a value of −1 indicates a perfect negative correlation, and a value of 0 indicates no correlation.
(prediction accuracy), which is defined as the percentage of correctly classified residues, is calculated as follows:
Probability of correct prediction or is the percentage of correctly predicted turns among the predicted turns. It is also called predicted positive value (PPV), and it is given as follows:
Sensitivity or coverage (also known as ) is the percentage of correctly predicted turns among the observed turns, or it is the fraction of the total positive samples that are correctly predicted, and it is given as follows:
Specificity is the fraction of total negative samples that are correctly predicted
3. Results
The selected number of Pvs (m) where from the feature matrix affects the accuracy and the MCC of the prediction. A relatively small or big will yield low performance. To select the optimal number of vectors a crossvalidation is used starting with relatively small and adding more vectors to until a point where adding more vectors does not improve the classification performance reached. Table 1 shows the prediction accuracy and MCC using different values of . In Figure 2, we see that the highest MCC is achieved for equal 120, while Figure 3 shows that the highest accuracy is achieved using equal to 115 or 120. The dataset used for the two figures is BT426.

After a short analysis of various values of threshold, we set its value to 0.45 to obtain the results in Table 1. The has improved slightly when the threshold value is set to 0.50, while the MCC dropped to less than 0.46. Similarly, the MCC has increased when the threshold value is set to 0.40, but at the cost of , which has dropped to less than 79%. The number of selected vectors in this research is set to 120 for BT426 dataset. Using this value for we obtained a of 80.54%, MCC of 0.48, of 59%, of 62%, and Specificity of 86%. The MCC is a robust and reliable performance measure that accounts for both overpredictions and underpredictions. A high MCC value indicates a high prediction performance.
To increase the performance of our KLR model further we used state changing rules. In these rules we put in our consideration that turns occur in a group of at least four adjacent residues. After analyzing the results obtained by the KLR prediction, the state changing rules, which will make the prediction to be more turn like, are derived as follows.(1)Change isolated nonturn predictions to turn (i.e., tnt ttt).(2)Change isolated turn prediction to nonturn prediction (i.e., ntn nnn).(3)Change the residues that are neighboring two isolated turn predictions to turn (i.e., nttn tttt).(4)If there is isolated triplet of turns predictions, then change the adjacent nonturn prediction with the highest KLR probability output to turn (i.e., ntttn ttttn or ntttt).
The above rules should be executed in orders. After applying these rules, we obtained a better performance, where the MCC has increased from 0.48 to 0.50.
Table 2 shows the comparison between our KLR method and other best existing turns prediction methods. Our KLR method achieves prediction accuracy = 80.7%, = 58.98%, = 65.25%, sensitivity = 85.34%, and MCC = 0.50. We note that the of our method is 0.2% lower than the of BTNpred and ESSpred, but because βturns account for approximately 25% of the globular protein residues, is a poor measure by itself, as it is possible to achieve of 75% by predicting all residues to be nonβturns. Instead, our method shows high MCC 0.50 compared to BTNpred 0.47 and ESSpred 0.44. The NetturnP and our method have the highest MCC 0.50 among the other turns prediction methods. Other than BTNpred and ESSpred our KLR shows the highest . When combining and MCC our method has the highest performance among the other prediction methods. Considering the baseline accuracy which equals 75%, our method provides 5.7/25 = 0.23% error rate reduction, while BTNpred and ESSpred provide 5.9/25 = 0.24% error rate reduction, and the second best method (SVM) provides 4.8/25 = 0.19% error rate reduction. The of our method is higher by 9.65% than the of BTNpred, by 3.25% than the of the BTSVM, and by 16.05% than the of ESSpred. Higher values mean that a large percentage of the observed urns are correctly predicted. At the same time, the of our method shows that 58% of the actual turns are correctly predicted. We note that the of our method is 3.72% lower than the of BTNpred and 4.62% lower than the of the ESSpred. The increase in the values is a tradeoff for the decrease in the values. In spite of this tradeoff, our method shows high overall accuracy.

Besides BT426 dataset that is used for training and testing our method, we used two additional datasets, that is, BT547 and BT823 datasets, to validate the performance of our method. Results obtained based on sevenfold crossvalidation on these datasets are given in Table 3. The results show that for the BT547 dataset our method obtains = 80.46%, = 59.04%, = 65.36%, and MCC = 0.50. The MCC of our method is the best among other competing methods that are evaluated on BT547 dataset. For the BT823 dataset our method obtains = 80.66%, = 58.42%, = 64.64%, and MCC = 0.49. Also our method has the highest MCC on BT547 and BT823 datasets. The results also show that our method shows stable performances on all the three datasets used.
All the computations for KLR were carried out using Matlab version 2010b on a computer with 3 GB RAM and 1.86 GHz Genuine Intel dual core processor. We compared the average elapsed time of our method with the BTNpred and ESSpred. The results of the comparison are shown in Table 4. In this comparison, we used onefold out of the sevenfolds in all the datasets as a test set and the remaining folds as training set. Since both BTNpred and ESSpred used SVM, we used LIBSVM [30] on their features. Note that both ESSpred and BTNpred used PSSMs and secondary structure information as features. In addition to PSSMs and secondary structure information, ESSpred added amino acid (AA) composition generated with classical local coding scheme. We also compared the average execution time of the KLR method with LIBSVM on our feature vector using BT426 dataset. Figure 4 shows the average execution time of the KLR method and LIBSVM in function of the number of the training instances.

Compared to ESSpred and BTNpred as shown in Table 4, our method is faster by more than a factor of 14. Although the training data in BT823 is more than the training data in BT547, its computation time using KLR is less than the computation time of BT547, that is because the number of selected vectors for BT823 is 90, which is by far less than the number of selected vectors for BT547, which is 140. This indicates that, for a very large dataset, a very small number of selected vectors can be sufficient to approximate its Kernel matrix, which reflects the capability of the FSKLR model to handle largescale datasets.
The ROC curve, which is a plot of the sensitivity against the falsepositive rate for the evaluation of the KLR, is shown in Figure 5. From the ROC curve, we calculated the area under the curve (AUC), which is a thresholdindependent measure. An AUC value above 0.7 is an indication of a useful prediction, and a good prediction method achieves a value above 0.85 [31]. NetTurnP, DEBT, ESSpred, and SVM achieved AUC of 0.864, 0.84, 0.84, and 0.87, respectively; our method achieves AUC of 0.861.
4. Discussion
Prediction of turns has attracted researchers interest because it plays the following important roles.
turns have been proposed to be important in folding because they are capable of initiating productive structure formation without a large loss in chain entropy since the interactions involved in turn formation are largely local [32]. They can play two different roles in the folding reaction of a protein. They can be either foldingactive elements and function as initiation sites or foldingpassive elements that form only after other regions develop. These different roles are likely to arise from the relative importance of the various interactions in forming the native states of different proteins [33].
Turns can influence the stability of a protein’s native state both by their intrinsic preference to sample favorable space and by their sidechain packing interactions and local environment [34]. Since turns usually occur on the exposed surface of a protein, they are well suited to participate in ligand binding, molecular recognition, proteinprotein, or proteinnucleic acid interactions, thus modulating protein functions and intermolecular interactions. Additionally, they are frequent sites of posttranslational modifications such as phosphorylation and glycosylation, which are used to tune interactions [4].
turns are also involved in the biological activity of peptides as the bioactive structures that interact with other molecules such as receptors, enzymes, or antibodies. Recent years have seen interest in mimicking turns for the synthesis of medicines. Thus, turn is an important component of protein structure whose prediction can provide enormous information to the researchers working in the field of drug design. So the prediction of turns would not only aid in overall tertiary structure prediction but also assists in fold recognition studies.
Throughout the previous research on turns prediction, predictors based on machinelearning method emphasize selecting proper features to improve prediction performance. Secondary structures and PSSMs are widely used in the predictions and have been proven to be the most helpful features. Using these features the proposed KLR method achieves comparable results to the SVMs methods. To design a method that can be applied in βturn prediction, there are four main concerns. These concerns are the size of the dataset, the need for dealing with input examples of variable length, the need to have probabilistic outcomes, and the need to perform multiclass classification. When the dataset is very large such as the βturns data, people neglect the last two concerns and concentrate on selecting a classifier that deals with large datasets effectively. Since SVMs methods are designed in a way that can handle largescale datasets, they become the choice for most of the βturns classification methods. However, SVMs do not address the last two concerns directly. KLR is not used in largescale datasets such as βturns data classification although it provides elegant solution to the last two concerns, simply because it is inapplicable in such datasets. The last two concerns are very important for βturns classification, since there is a need for multiclass classification for the βturns type. FSKLR extends the applicability of KLR for largescale datasets. This way it can address all of the aforementioned concerns.
5. Conclusion
In this paper, we presented sparse KLR method for turns prediction. Our method is based on FSKLR in which trust region Newton’s method for largescale LR is used as a basis to solve the approximate problem, while Nystrom method is used to approximate the features’ matrix. Our method uses secondary structure information and PSSMs as input features. Empirical evaluations using three nonredundant datasets show that our predictions provide favorable , , and MCC when compared with the stateoftheart methods that used secondary structure information and PSSMs as features. Using our method we achieved and MCC of 80.7% and 0.50, respectively, on BT426 dataset. In addition, KLR yields probabilistic outputs and its extension to the multiclass case is well defined, which will be appropriate for turns types prediction. The computational complexity of our method is and its computation time is by far less than that of SVMs methods.
Conflict of Interests
The authors declare that there is no conflict of interests.
Acknowledgments
This work is supported in part by the National Natural Science Foundation of China under Grant nos. 61232001, 61073036, and 61128006, the Ph.D. Programs Foundation of Ministry of Education of China no. 20090162120073, and the Freedom Explore Program of Central South University no. 201012200124.
References
 A. Bairoch and R. Apweiler, “The SWISSPROT protein sequence database and its supplement TrEMBL in 2000,” Nucleic Acids Research, vol. 28, no. 1, pp. 45–48, 2000. View at: Google Scholar
 H. M. Berman, J. Westbrook, Z. Feng et al., “The protein data bank,” Nucleic Acids Research, vol. 28, no. 1, pp. 235–242, 2000. View at: Google Scholar
 J. Guo, H. Chen, Z. Sun, and Y. Lin, “A novel method for protein secondary structure prediction using duallayer SVM and profiles,” Proteins, vol. 54, no. 4, pp. 738–743, 2004. View at: Publisher Site  Google Scholar
 G. D. Rose, L. M. Gierasch, and J. A. Smith, “Turns in peptides and proteins,” Advances in Protein Chemistry, vol. 37, pp. 1–109, 1985. View at: Google Scholar
 K. S. Kee and S. D. S. Jois, “Design of βturn based therapeutic agents,” Current Pharmaceutical Design, vol. 9, no. 15, pp. 1209–1224, 2003. View at: Publisher Site  Google Scholar
 C. Zheng and L. Kurgan, “Prediction of βturns at over 80% accuracy based on an ensemble of predicted secondary structures and multiple alignments,” BMC Bioinformatics, vol. 9, article 430, 2008. View at: Publisher Site  Google Scholar
 E. G. Hutchinson and J. M. Thornton, “A revised set of potentials for βturn formation in proteins,” Protein Science, vol. 3, no. 12, pp. 2207–2216, 1994. View at: Google Scholar
 P. Kountouris and J. D. Hirst, “Predicting βturns and their types using predicted backbone dihedral angles and secondary structures,” BMC Bioinformatics, vol. 11, article 407, 2010. View at: Publisher Site  Google Scholar
 A. J. Shepherd, D. Gorse, and J. M. Thornton, “Prediction of the location and type of βturns in proteins using neural networks,” Protein Science, vol. 8, no. 5, pp. 1045–1055, 1999. View at: Google Scholar
 H. Kaur and G. P. S. Raghava, “Prediction of βturns in proteins from multiple alignment using neural network,” Protein Science, vol. 12, no. 3, pp. 627–634, 2003. View at: Publisher Site  Google Scholar
 D. T. Jones, “Protein secondary structure prediction based on positionspecific scoring matrices,” Journal of Molecular Biology, vol. 292, no. 2, pp. 195–202, 1999. View at: Publisher Site  Google Scholar
 L. J. McGuffin, K. Bryson, and D. T. Jones, “The PSIPRED protein structure prediction server,” Bioinformatics, vol. 16, no. 4, pp. 404–405, 2000. View at: Google Scholar
 B. Petersen, C. Lundegaard, and T. N. Petersen, “NetTurnP—neural network prediction of βturns by use of evolutionary information and predicted protein sequence features,” PLoS ONE, vol. 5, no. 11, Article ID e15079, 2010. View at: Publisher Site  Google Scholar
 P. Karsmakers, Sparse kernelbased models for speech recognition [Ph.D. thesis], Katholieke Universiteit Leuven, Arenberg, Belgium, 2010.
 P. Karsmakers, K. Pelckmans, and J. A. K. Suykens, “Multiclass kernel logistic regression: a fixedsize implementation,” in Proceedings of the International Joint Conference on Neural Networks (IJCNN '07), pp. 1756–1761, Orlando, Fla, USA, August 2007. View at: Publisher Site  Google Scholar
 C.J. Lin, R. C. Weng, and S. S. Keerthi, “Trust region Newton method for largescale logistic regression,” Journal of Machine Learning Research, vol. 9, pp. 627–650, 2008. View at: Google Scholar
 K. Guruprasad and S. Rajkumar, “β and γturns in proteins revisited: a new set of amino acid turntype dependent positional preferences and potentials,” Journal of Biosciences, vol. 25, no. 2, pp. 143–156, 2000. View at: Google Scholar
 P. F. J. Fuchs and A. J. P. Alix, “High accuracy prediction of βturns and their types using propensities and multiple alignments,” Proteins, vol. 59, no. 4, pp. 828–839, 2005. View at: Publisher Site  Google Scholar
 S. F. Altschul, T. L. Madden, A. A. Scha ffer et al., “GappedBLAST and PSIBLAST: a new generation of protein database search programs,” Nucleic Acids Research, vol. 25, pp. 3389–3402, 1997. View at: Google Scholar
 K. Bryson, L. J. McGuffin, R. L. Marsden, J. J. Ward, J. S. Sodhi, and D. T. Jones, “Protein structure prediction servers at University College London,” Nucleic Acids Research, vol. 33, no. 2, pp. W36–W38, 2005. View at: Publisher Site  Google Scholar
 J. A. Cuff and G. J. Barton, “Application of multiple sequence alignment profiles to improve protein secondary structure prediction,” Proteins, vol. 40, no. 3, pp. 502–511, 2000. View at: Google Scholar
 S. Montgomerie, S. Sundararaj, W. J. Gallin, and D. S. Wishart, “Improving the accuracy of protein secondary structure prediction using structural alignment,” BMC Bioinformatics, vol. 7, article 301, 2006. View at: Publisher Site  Google Scholar
 K. Zhang, I. W. Tsang, and J. T. Kwok, “Improved Nyström lowrank approximation and error analysis,” in Proceedings of the 25th International Conference on Machine Learning, pp. 1232–1239, Helsinki, Finland, July 2008. View at: Google Scholar
 K. C. Chou and C. T. Zhang, “Prediction of protein structural classes,” Critical Reviews in Biochemistry and Molecular Biology, vol. 30, pp. 275–349, 1995. View at: Google Scholar
 P. Baldi, S. Brunak, Y. Chauvin, C. A. F. Andersen, and H. Nielsen, “Assessing the accuracy of prediction algorithms for classification: an overview,” Bioinformatics, vol. 16, no. 5, pp. 412–424, 2000. View at: Google Scholar
 X. Hu and Q. Li, “Using support vector machine to predict β and γturns in proteins,” Journal of Computational Chemistry, vol. 29, no. 12, pp. 1867–1875, 2008. View at: Publisher Site  Google Scholar
 T. H. Pham, K. Satou, and T. B. Ho, “Prediction and analysis of βturns in proteins by support vector machine,” Genome Informatics, vol. 14, pp. 196–205, 2003. View at: Google Scholar
 L. Liu, Y. Fang, M. Li, and C. Wang, “Prediction of βturn in protein using ESSpred and support vector machine,” Protein Journal, vol. 28, no. 34, pp. 175–181, 2009. View at: Publisher Site  Google Scholar
 C.T. Zhang and K. C. Chou, “Prediction of βturns in proteins by 14 and 23 correlation model,” Biopolymers, vol. 41, no. 6, pp. 673–702, 1997. View at: Google Scholar
 C.C. Chang and C.J. Lin, “LIBSVM: A library for support vector machines,” http://www.csie.ntu.edu.tw/~cjlin/libsvm/. View at: Google Scholar
 O. Lund, M. Nielsen, C. Lundegaard, C. KeSmir, and S. Brunak, Immunological Bioinformatics, The MIT Press, London, UK, 2005.
 H. J. Dyson and P. E. Wright, “Peptide conformation and protein folding,” Current Opinion in Structural Biology, vol. 3, pp. 60–65, 1993. View at: Google Scholar
 A. Marcelino and L. Gierasch, “Roles of βturns in protein folding: from peptide models to protein engineering,” Biopolymers, vol. 89, no. 5, pp. 380–391, 2008. View at: Google Scholar
 P. F. Predki, V. Agrawal, A. T. Brunger, and L. Regan, “How amino acid substitutions in a surface turn modulate protein stability,” Nature Structural Biology, vol. 3, pp. 54–58, 1996. View at: Google Scholar
Copyright
Copyright © 2013 Murtada Khalafallah Elbashir et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.