Research Article | Open Access
Yin-Fu Huang, Shu-Ying Chen, "Extracting Physicochemical Features to Predict Protein Secondary Structure", The Scientific World Journal, vol. 2013, Article ID 347106, 8 pages, 2013. https://doi.org/10.1155/2013/347106
Extracting Physicochemical Features to Predict Protein Secondary Structure
We propose a protein secondary structure prediction method based on position-specific scoring matrix (PSSM) profiles and four physicochemical features including conformation parameters, net charges, hydrophobic, and side chain mass. First, the SVM with the optimal window size and the optimal parameters of the kernel function is found. Then, we train the SVM using the PSSM profiles generated from PSI-BLAST and the physicochemical features extracted from the CB513 data set. Finally, we use the filter to refine the predicted results from the trained SVM. For all the performance measures of our method, reaches 79.52, SOV94 reaches 86.10, and SOV99 reaches 74.60; all the measures are higher than those of the SVMpsi method and the SVMfreq method. This validates that considering these physicochemical features in predicting protein secondary structure would exhibit better performances.
Many issues on molecular biology have been addressed in the past decades, including genetics, structural biology, and drug design. A protein primary sequence is composed of amino acids; as we know, totally 20 different kinds of amino acids can be found in protein sequences. In this paper, we would investigate protein secondary structures based on protein sequences.
The secondary structure of a protein sequence comes from different folding of amino acids, due to the differences of their side chain sizes, shapes, reactivity, and the ability to form hydrogen bonds. Furthermore, owing to the differences of the side chain sizes, the number of electric charges, coupled with the affinity for water, the tertiary structures of protein sequences are not all the same. Thus, the exploration of molecular structures on protein sequences is divided into secondary, tertiary, and even quaternary structures. Given a protein primary sequence, its corresponding secondary structure can be revealed as follows: Primary sequence: MFKVYGYDSNIHKCVYCDNAKRLLTVKKQPFEFINIMPEKGV Secondary structure: CEEEEECCCCCCCCHHHHHHHHHHHHCCCCEEEEECCCCTTC.
A protein sequence affects the structure and function; in other words, a protein sequence determines its structure, and the structure determines functions. If amino acids in a protein sequence are arranged in a different order in the skeleton branch of the side chain R group, the nature of the protein would reveal specific functions. Even for different species of proteins, if they have a similar structure, their functions would be also similar. Therefore, predicting the protein structure is crucial to the function analysis. Besides, the secondary structure refers to the relative position of the space between the atoms of a certain backbone. Traditional protein structure determination was done by protein X-ray crystallography or nuclear magnetic resonance (NMR). However, all experimental analysis costs much time. In order to shorten the time to help biologists, protein structure prediction by computers facilitates reaching this goal.
The prediction of protein secondary structure has been studied for decades. Early, the statistical analysis of secondary structure was done for a single amino acid. The most representative is the Chou and Fasman method , and the accuracy is only 50%. Next, the statistical analysis for amino acid segments was done further. A segment length is usually with 9~21 amino acids. Based on an amino acid segment, predicting the structure of central residues enables promoting the accuracy. The most representative is the GOR method , and the accuracy increases more than 10% (about 63%). At present, the prediction methods on protein secondary structure have evolved into using the PSI-BLAST program  to find the protein homology information, based on PSSM (position-specific scoring matrices) profiles. The accuracy of using PSSM to predict secondary structure has reached between 70 and 80% [4–7]. However, we believe that there still exists a great improvement in predicting protein secondary structure.
The rest of this paper is organized as follows. In Section 2, basic concepts used in the proposed methods are introduced first. In Section 3, we propose the methods and relevant features to predict the secondary structure of a protein sequence. Then, we make use of window sizes and tune parameters in the experiments in Section 4, in order to obtain better experimental results. Finally, we make a conclusion in Section 5.
2. Basic Concepts
2.1. Protein Secondary Structure
Protein secondary structure derived from the experimentally determined 3D structure has been defined using DSSP (Dictionary of Secondary Structures of Proteins) , STRIDE (STRuctural IDEntification) , and DEFINE (DEFINE_structure) . DSSP is selected here so that our method can be compared with most existing methods, based on the same protein secondary structure definition. Eight secondary structure classes were defined there, that is, H(α-helix), G(310-helix), I(π-helix), E(β-strand), B(isolated β-bridge), T(turn), S(bend), and -(rest). The eight structure classes are usually reduced to three classes of helix (H), sheet (E), and coil (C). Five reductions could be performed as follows:(1)H, G and I to H; E to E; the rest to C(2)H, G to H; E, B to E; the rest to C(3)H, G to H; E to E; the rest to C(4)H to H; E, B to E; the rest to C(5)H to H; E to E; the rest to C.
The first reduction was used in the PHD (Profile network from HeiDelberg) method  which is the early secondary structure prediction method using multiple sequence alignments of proteins homologous with a query protein sequence. We also use the first reduction in order to provide a fair comparison with other prediction methods.
2.2. SVM (Support Vector Machine)
SVM was first investigated by Boser et al. in 1992 . It solves linearly inseparable problems by nonlinearly mapping the vector in a low dimensional space to a higher dimensional feature space and constructs an optimal hyper-plane in the higher dimensional space. Therefore, SVM has high performances in data classification. A classification task usually involves with training and testing data which consist of some data instances. Each instance in the training set contains one “target value” (i.e., class label) and several “attributes” (i.e., features). The goal of SVM is to produce a model which can predict the target value of data instances in the testing set by using the attributes.
3. Methods and Features
3.1. System Architecture
The system architecture of predicting protein secondary structure is divided into three steps, as illustrated in Figure 1. The first step is to determine/extract the relevant features in/from protein sequences. Then, in the second step, we feed the features into SVM, respectively, in the training and test phases. Finally, we use a filter method to refine the predicted results from the trained SVM. During the SVM training phase, we not only train the SVM using the training data, but also, in advance, find the optimal sliding window size and the cost and gamma parameters of SVM kernel function, using the entire data set. The details about each module in the system architecture are depicted in the following subsections.
3.2. Feature Extraction
Five relevant kinds of features are extracted from protein sequences to predict protein secondary structure, that is, conformation parameters, position specific scoring matrix (PSSM) profiles, net charge, hydrophobic, and side chain mass. The process of feature extraction is shown in Figure 2.
3.2.1. Extracting Sequences
First, we extract amino acid and secondary structure sequences from the PDB website (http://www.rcsb.org/pdb/home/home.do), using the PDB codes of CB513 . Then, we can further extract five different features from amino acid sequences as follows.
3.2.2. Conformation Parameters
Conformation parameters are the proportions that residues (or amino acids) tend to secondary structure. In general, protein secondary structure is divided into three types: α-helix (H), β-sheet (E), and coil (C), so that there are three values for each amino acid. In the feature extraction, all the conformation parameters are calculated from a data set. The conformation parameters for each amino acid are defined as follows: In this formula, indicates the 20 amino acids, and indicates the 3 types of secondary structure: H, E, and C. Here, is the amount of the th amino acid in a data set whereas is the amount of the th amino acids with the th secondary structure. The conformation parameters for each amino acid in a data set are shown in Table 1. The reason of using conformation parameters as features is that the folding of each residue has some correlation with forming a specific structure.
3.2.3. PSSM Profiles
PSSM profiles are generated by PSI-BLAST (Position Specific Iterative-Basic Local Alignment Search Tool) program. Since PSSM profiles are involved with biological evolution, we consider them as features in our work. A PSSM profile has L × 20 elements, where L is the length of a query sequence. These profiles are then used as the input features to feed an SVM, employing a sliding window method.
PSI-BLAST is based on BLAST which has been published by Altschul et al. in 1997 . Since PSI-BLAST program is more sensitive than other methods, we can find a lot of low similarity sequences and similarity structure function of protein sequences. First, a database containing all known sequences (or nonredundant database) is selected. Then, low complexity regions are removed from the nr database. Finally, PSI-BLAST program is used to query each sequence in CB513 and generates PSSM profiles after three iterations. Here, multiple sequence alignment (MSA) and BLOSUM62 matrix  are used in this process.
The reason of using the sliding window method is to get more surrounding information of residues. We consider a sliding window of size 7~19 at which a predicted residue is centered to extract input features. The optimal window size yielding favorable predictive performances would be obtained experimentally. For the th residue centered at the sliding window of size 7, we can get 7 × 20 features where is in the range 3, 3 and is the PSSM column from 1 to 20.
3.2.4. Net Charges
There are five amino acids with charges, that is, R, D, E, H, and K. Since residues with similar electric charges repel each other and interrupt the hydrogen bond of main chain, they are adverse to α-helix formation. Besides, the continuous residues of β-sheet cannot be with similar charges. This information facilitates predicting the secondary structure. The net charge of amino acids can be taken from Amino Acid index database (or AAindex) [14–18], as shown in Table 2. A plus sign represents a positive charge and a minus sign represents a negative charge.
For protein folding, polar residues prefer to stay outside of protein to prevent non-polar (hydrophobic) residues from exposing to polar solvent, like water. Therefore, hydrophobic residues appearing periodically can be used for predicting protein secondary structure. In general, the residues in α-helix structure are made up of one segment of hydrophobic and one segment of hydrophilic. However, β-sheet structure is usually influenced by the environment, so this phenomenon is not obvious. In other words, hydrophobic affects the stability of secondary structure. The hydrophobic values of amino acids can also be obtained from Amino Acid index database (or AAindex) [14–18], as shown in Table 3. The more positive values are, the more hydrophobic is.
3.2.6. Side Chain Mass
Although the basic structure as shown in Figure 3 is the same for 20 amino acids, the size of the side chain R group still influences structure folding. Here, we explain the influences as follows. First, the side chain R group is distributed in the outside of the main chain of α-helix structure, but the continuous large R groups can make α-helix structure unstable, thereby disabling amino acids from forming α-helix structure. Next, the R group with ring structure like proline (P) is not easy to form α-helix structure. Proline is composed of 5 atoms in a ring, which is not easy to reverse and is also not easy to generate a hydrogen bond. Finally, we observe that the R group of β-sheet structure is smaller than those of other structures, in general. Therefore, we include the side chain mass as a feature, as shown in Table 4.
3.3. SVM (Finding the Optimal Window Size and Parameters)
The SVM used in the experiments is a classifier for predicting the secondary structure H, E, and C. Threefold cross-validation is employed on the CB513 data set to find the optimal window size in the range [7, 19] and the optimal parameters of the kernel function, such as cost C and gamma γ. Here, the kernel function used in the SVM is RBF (i.e., Radial Basis Function). To solve the multiclass problem confronted in the work, we employ the “one-against-one” approach. For 3 classes, we need 3 binary classifiers and set the labels of the secondary structure (H, E, C) to (−1, +1, +2). Then, we use the max-wins voting strategy to determine the class; in other words, each binary classifier casts a vote, and the winning class is with the highest number of votes. In the experiments, the LIBSVM tool kit proposed by Chang and Lin  would be used to implement the program. After the optimal window size and parameters are found, we would use the SVM for training and test.
A single residue in its natural state cannot be alone folded into α-helix or β-sheet. Thus, setting thresholds on the length of consensus secondary structure can be used to filter out incorrect predicted results. For example, at least three contiguous residues are for α-helix and at least two contiguous residues are for β-sheet. For the current scanning window (, , + 1) in the predicted secondary structure, two possible structures could happen at position :Case H: if str() and str() are H, then str() is not changed; otherwise, extend the examined segment to (, , , , , , ) and replace str() with the majority structure in the examined segment.Case E: if str() or str() is E, then str() is not changed; otherwise, extend the examined segment to (, , , , , , ) and replace str() with the majority structure in the examined segment.For the example as shown in Figure 4, after the filtering, for 9INSb is improved from 76.7 to 80 and SOV99 is improved from 77.8 to 93.3 where and SOV99 will be described in Section 4.2.
4.1. Data Set
In the previous work, some typical data sets were frequently used in protein secondary structure prediction, such as RS126 , CB513 , CASP , and EVA . Here, we consider the selected data set should be with low similarity; that is, the protein sequences within the data set are not similar to each other. Thus, the protein secondary structure prediction we develop would enable predicting an unknown protein sequence more accurately.
In our work, the data set we choose is nonhomologous CB513 data set constructed by Cuff and Barton and contains 513 protein chains. Almost all the sequences in the RS126 data set are also included in the CB513 data set. The CB513 data set contains 16 chains of ≦30 residues. Although very short chains would slightly decrease the accuracy for the hard definition of secondary structures, we still include them in the set for comprehensive study. We retrieve the CB513 data set from the website: http://paraschopra.com/projects/evoca_prot/index.php, which contains 84,093 residues where 34.59% of the residues is for helix, 21.35% for sheet, and 44.06% for coil, as shown in Table 5.
4.2. Performance Measures
Two kinds of performance measures are frequently used in protein secondary structure prediction; that is, or accuracy (three-state overall per-residue accuracy) and SOV99  (or SOV94 ) (Segment Overlap measure). is a residue-based measure of three-structure overall percentage of correctly classified residues, which can be represented as where is the total number of predicted residues, Hpre is the correctly classified secondary structure for helix, Epre for sheet, and Cpre for coil.
SOV99 is a segment-based measure of three structures, whose value is within the range 0, 100, as shown in Formula (3). SOV99 differs from in the prediction unit such that SOV99 would penalize wrong predictions; for example, a single helix predicted as a multiply-split helix is unrealistic prediction where and denote segments of secondary structure (H, E, or C), and are both in structure, is a normalization value, (, ) is the length of actual overlap of and , , ) is the length of total extent for and , and (, ) can be represented as The definition of and the normalization value N are different for SOV99 and SOV94.
4.3. Optimal Parameters and Window Sizes
As introduced in Section 2.2, we adopt the well-known LIBSVM developed by Chang and Lin  as an SVM classifier. The kernel function used here is RBF (Radial Basis Function) since it is more accurate and effective than the other kernel ones. The parameters and are determined by the optimum performance of 6 × 6 combinations between and for each window size. Moreover, the feature vector is normalized in the range 0, 1 and the number of features in a larger window size would become more. The optimal parameters and classification accuracy are evaluated in threefold cross-validation, as shown in Table 6.
According to the experimental results, we found the optimal parameters and window size are = 21, = 2−4, and WS = 13. Then, we use these parameters and window size to conduct the further experiments.
4.4. Experimental Results
In this section, we compare the experimental results without filtering and with filtering. For the classification results, a confusion matrix is employed to present the correct and false predictions based on the precision and recall, as shown in Tables 7 and 8. The precision and recall are expressed as follows:
Obviously, the classification accuracy with filtering (i.e., 79.52%) is higher than that without filtering (i.e., 77.40%). The precision for and the recall for especially are improved from 83.64 to 92.00 (with filtering) and from 82.43 to 88.31 (with filtering), respectively. Therefore, the filter rules are required to improve the accuracy in predicting protein secondary structure.
4.5. Comparing with Other Methods
Here, we compare our methods with other four methods; that is, PHD, SVMfreq, PMSVM, and SVMpsi as shown in Table 9. Both the PHD and SVMfreq methods are based on the frequency profiles with multiple sequence alignment; however, the classifier used in the PHD method is a neural network (or NN) whereas the classifier used in the SVMfreq method is a support vector machine (or SVM). Similarly, both the PMSVM and SVMpsi methods are based on the PSSM profiles generated from PSI-BLAST. Although they use the same-type classifier (or SVM), the former adopts one-versus-one classifier (i.e., , , ) and the latter adopts the one-versus-rest classifier (i.e., /~, /~, /~).
As shown in Table 9, we found that all the performance measures of our method (i.e., the version with filtering), including , SOV94, and SOV99, are higher than those of the other four methods, regardless using the CB513 or RS126 data sets. for the version with filtering (or without filtering) is improved by 2.92 (or 0.8), SOV94 for the version with filtering (or without filtering) is improved by 6 (or 10.1), and SOV99 for the version with filtering is improved by 1.1, compared with the results of the SVMpsi method for CB513 (i.e., the next best one).
However, our method (i.e., the version with filtering) has lower R(H) than the SVMpsi method (i.e., 76.91 versus 78.1). One of the possible reasons is that the threshold on the length of consensus secondary structure (i.e., at least three contiguous residues for H) is set in the filter. Although the recall for H is decreased, the predicted structures are more structurally meaningful. Besides, we found that two SOV measures in the SVMpsi and our methods vary greatly. Although SOV94 is decreased (i.e., from 90.20 to 86.10) after applying the filter in our method, the latest definition (i.e., SOV99) is still the highest.
In this paper, we propose a protein secondary structure prediction method using PSSM profiles and four physicochemical features, including conformation parameters, net charges, hydrophobic, and side chain mass. In the experiments, the SVM with the optimal window size and the optimal parameters of the kernel function is found first. Then, we train the SVM using the PSSM profiles and physicochemical features extracted from the CB513 data set. Finally, we use the filter to refine the predicted results from the trained SVM. For the experimental results, , SOV94, SOV99, and recall of our method are higher than those of the SVMpsi method based on the PSI-BLAST profiles as well as the SVMfreq method based on the frequency profiles with multiple sequence alignment for the CB513 data set. In summary, considering these physicochemical features in predicting protein secondary structure would exhibit better performances.
This work was supported by National Science Council of Taiwan under Grant NSC100-2218-E-224-011-MY3.
- P. Y. Chou and G. D. Fasman, “Empirical predictions of protein conformation,” Annual Review of Biochemistry, vol. 47, pp. 251–276, 1978.
- J. Garnier, D. J. Osguthorpe, and B. Robson, “Analysis of the accuracy and implications of simple methods for predicting the secondary structure of globular proteins,” Journal of Molecular Biology, vol. 120, no. 1, pp. 97–120, 1978.
- S. F. Altschul, T. L. Madden, A. A. Schäffer et al., “Gapped BLAST and PSI-BLAST: a new generation of protein database search programs,” Nucleic Acids Research, vol. 25, no. 17, pp. 3389–3402, 1997.
- J. Guo, H. Chen, Z. Sun, and Y. Lin, “A novel method for protein secondary structure prediction using dual-layer SVM and profiles,” Proteins, vol. 54, no. 4, pp. 738–743, 2004.
- S. Hua and Z. Sun, “A novel method of protein secondary structure prediction with high segment overlap measure: support vector machine approach,” Journal of Molecular Biology, vol. 308, no. 2, pp. 397–407, 2001.
- H. Kim and H. Park, “Protein secondary structure prediction based on an improved support vector machines approach,” Protein Engineering, vol. 16, no. 8, pp. 553–560, 2003.
- B. Rost and C. Sander, “Prediction of protein secondary structure at better than 70% accuracy,” Journal of Molecular Biology, vol. 232, no. 2, pp. 584–599, 1993.
- W. Kabsch and C. Sander, “Dictionary of protein secondary structure: pattern recognition of hydrogen-bonded and geometrical features,” Biopolymers, vol. 22, no. 12, pp. 2577–2637, 1983.
- D. Frishman and P. Argos, “Knowledge-based protein secondary structure assignment,” Proteins, vol. 23, no. 4, pp. 566–579, 1995.
- F. M. Richards and C. E. Kundrot, “Identification of structural motifs from protein coordinate data: secondary structure and first-level supersecondary structure,” Proteins, vol. 3, no. 2, pp. 71–84, 1988.
- B. E. Boser, I. M. Guyon, and V. N. Vapnik, “Training algorithm for optimal margin classifiers,” in Proceedings of the 5th Annual ACM Workshop on Computational Learning Theory, pp. 144–152, July 1992.
- J. A. Cuff and G. J. Barton, “Evaluation and improvement of multiple sequence methods for protein secondary structure prediction,” Proteins, vol. 34, no. 4, pp. 508–519, 1999.
- S. Henikoff and J. G. Henikoff, “Amino acid substitution matrices from protein blocks,” Proceedings of the National Academy of Sciences of the United States of America, vol. 89, no. 22, pp. 10915–10919, 1992.
- S. Kawashima, H. Ogata, and M. Kanehisa, “AAindex: amino acid index database,” Nucleic Acids Research, vol. 27, no. 1, pp. 368–369, 1999.
- S. Kawashima and M. Kanehisa, “AAindex: amino acid index database,” Nucleic Acids Research, vol. 28, no. 1, p. 374, 2000.
- S. Kawashima, P. Pokarowski, M. Pokarowska, A. Kolinski, T. Katayama, and M. Kanehisa, “AAindex: Amino acid index database, progress report 2008,” Nucleic Acids Research, vol. 36, no. 1, pp. D202–D205, 2008.
- K. Nakai, A. Kidera, and M. Kanehisa, “Cluster analysis of amino acid indices for prediction of protein structure and function,” Protein Engineering, Design and Selection, vol. 2, no. 2, pp. 93–100, 1988.
- K. Tomii and M. Kanehisa, “Analysis of amino acid indices and mutation matrices for sequence comparison and structure prediction of proteins,” Protein Engineering, vol. 9, no. 1, pp. 27–36, 1996.
- C.-C. Chang and C.-J. Lin, “LIBSVM: a library for support vector machines,” http://www.csie.ntu.edu.tw/~cjlin/libsvm, 2001.
- J. Moult, J. T. Pedersen, R. Judson, and K. Fidelis, “A large-scale experiment to assess protein structure prediction methods,” Proteins, vol. 23, no. 3, pp. ii–iv, 1995.
- V. A. Eyrich, M. A. Martí-Renom, D. Przybylski et al., “EVA: continuous automatic evaluation of protein structure prediction servers,” Bioinformatics, vol. 17, no. 12, pp. 1242–1243, 2002.
- A. Zemla, C. Venclovas, K. Fidelis, and B. Rost, “A modified definition of SOV, a segment based measure for protein secondary structure prediction assessment,” Proteins, vol. 34, no. 2, pp. 220–223, 1999.
- B. Rost, C. Sander, and R. Schneider, “Redefining the goals of protein secondary structure prediction,” Journal of Molecular Biology, vol. 235, no. 1, pp. 13–26, 1994.
Copyright © 2013 Yin-Fu Huang and Shu-Ying Chen. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.