Research Article  Open Access
Lei Yang, Yukun Han, Huixue Zhang, Wenlong Li, Yu Dai, "Prediction of ProteinProtein Interactions with Local WeightSharing Mechanism in Deep Learning", BioMed Research International, vol. 2020, Article ID 5072520, 11 pages, 2020. https://doi.org/10.1155/2020/5072520
Prediction of ProteinProtein Interactions with Local WeightSharing Mechanism in Deep Learning
Abstract
Proteinprotein interactions (PPIs) are important for almost all cellular processes, including metabolic cycles, DNA transcription and replication, and signaling cascades. The experimental methods for identifying PPIs are always timeconsuming and expensive. Therefore, it is important to develop computational approaches for predicting PPIs. In this paper, an improved model is proposed to use a machine learning method in the study of proteinprotein interactions. With the consideration of the factors affecting the prediction of the PPIs, a method of feature extraction and fusion is proposed to improve the variety of the features to be considered in the prediction. Besides, with the consideration of the effect affected by the different input order of the two proteins, we propose a “Ytype” BiRNN model and train the network by using a method which both needs backward and forward training. In order to insure the training time caused on the extra training either a backward one or a forward one, this paper proposes a weightsharing policy to minimize the parameters in the training. The experimental results show that the proposed method can achieve an accuracy of 99.57%, recall of 99.36%, sensitivity of 99.76%, precision of 99.74%, MCC of 99.14%, and AUC of 99.56% under the benchmark dataset.
1. Introduction
Protein plays an important role in the regulation of cell life activities, such as transcriptional regulation and signaling. At present, proteomics research with proteinprotein interactions (PPIs) as the main research content is widely used in the field of medical drug target discovery. It is of great significance to promote the development of the biomedical industry.
Currently, PPI prediction research mainly uses the following two techniques: (1) experimental methods, including yeast twohybrid [1, 2], protein chips [3, 4], coimmunoprecipitation [5], and Surface Plasmon Resonance (SPR) [6, 7]; (2) protein informationbased computational methods, including protein primary sequence information [8] and spatial structure information of protein secondary structure [9–11]. Compared with the experimental method, the computational method has the advantages of fast verification speed and strong repeatability. Currently, with the development of the machine learning, some methods based on machine learning have been proposed. Currently, when using the machine learning to predict the PPIs, the methods always combine the two sequences of the protein together as the input [12, 13]. However, this may double the length of the feature vector and is prone to an overfitting problem. Besides, either a forward or a backward training method may sometimes overlook the long and shortterm effect which may result in the low prediction accuracy. In addition, the current methods always just consider 7 amino acids which may result in the low diversity of the features and the low accurate prediction performance.
In order to improve the prediction accuracy, the protein feature extraction and fusion method based on the combination of discrete wavelet transform and continuous wavelet transform is used to describe the protein amino acid sequence according to eight physicochemical properties of amino acid to improve the variety of the features to be considered in the prediction. Besides, with the consideration of the effect affected by the different input order of the two proteins, we propose a “Ytype” BiRNN model and use both the backward and forward training methods to train the model. In addition, in order to insure the training time caused on the extra training either a backward one or a forward one, this paper proposes a weightsharing policy to minimize the parameters in the training. The experimental results verify that the method can effectively improve the accuracy and precision of the prediction results.
The following of the paper is organized as the follows: Section 2 introduces the related works of PPIs, Section 3 introduces the data and methods used in experiments, Section 4 introduces the experimental results and comparison with other methods in performance, and Section 5 introduces the conclusion of this paper and future work.
2. Related Work
Predicting PPIs using traditional experiments is often expensive and timeconsuming, so many computational methods are used to infer PPIs from different sources of information, including phylogenetic profiles, tertiary structures, protein domains, and secondary structures [14, 15]. However, these approaches cannot be employed when prior knowledge about a protein of interest is not available.
With the rapid growth of protein sequence data, protein sequencebased prediction methods are becoming the most widely used PPI prediction tool. Therefore, many protein sequencebased methods have been developed to predict PPIs. Bock and Gough [16] used support vector machines (SVM) in conjunction with physicochemical descriptors to predict PPIs, proving that only protein sequence information is valid for predicting PPIs. Shen et al. [17] proposed a method for PPI prediction using only the amino acid sequence information of proteins. The method combines the kernel function and the combined ternary feature to describe the amino acid and obtains an accuracy of 83.90%. In 2010, Yu et al. [18] also used the primary structure of protein alone to deal with proteinprotein interactions in unbalanced data. In 2013, You et al. [14] used amino acid sequence information and proposed ensemble extreme learning machines and principal component analysis methods to predict PPIs.
However, there are widespread problems of falsepositive rate and falsenegative rate in the above various experimental methods. In order to solve this problem, the researchers are more committed to the following two aspects to improve the accuracy of the prediction results. On the one hand, the robustness of feature extraction is improved to make the classification model effectively distinguish between positive and negative samples, thereby reducing the falsepositive rate and falsenegative rate. Thus, the dipeptide composition method [19, 20], the composition transformation distribution description method [21], the quasisequenceorder descriptor (QSOD) [22], the wavelet transform [23], and other protein feature extraction methods are used in combination, such as the methods proposed by Du et al. [12], An et al. [24], and Huang et al. [25]. On the other hand, researchers hope to increase the complexity of the training model to increase the prediction results. For example, You et al. [26] use a random forest method. Xu et al. [27] used ensemble learning methods to predict PPIs. Li et al. [28] proposed a deep learning method DNNPPI for CNN and LSTM. Although the above method has achieved certain effects, the training model takes too long, the model is too complicated, and the problem of overfitting the training model is also generated. Ref. [29] is to use the feature extraction and coding. The main work of us is to solve the problem of feature extraction and ensemble to get more appropriate features in order to ensure the accuracy. Besides, our work is to solve the problem of the effect caused by different input orders of the proteins. The extraction and coding method mentioned in the first reference can be used as our future work to improve the efficiency. The work of Ref. [30] takes the sequence of the two proteins as the input of the neural networks and uses a sLSTM to train the model. The work is similar to Li’s work which is also referred in our manuscript. Such works’ problem is that they do not consider the problem caused by the order of the two protein sequences. Ref. [31] adopts a Res2Vec to express the biological sequence and uses DNN to train the model, which also does not consider the relationship between the effect of the prediction and the order of the two protein sequences. But in the future work, we can use the proposed Res2Vec method to express the sequence also in order to avoid the overfitting problem. Ref. [32] is to use the RCNN to automatically select the features while the method does not consider the features in our work. Among the existing models, Wang et al. [13] proposed a method with a better comprehensive performance. They used discrete wavelet transform and continuous wavelet combination to extract protein features and obtained the accuracy of 97.38% under the yeast dataset. However, there are two shortcomings in this method. On the one hand, the method only considers the hydrophobicity of amino acids and regards hydrophobicity as the only property affecting protein interaction, ignoring other properties of amino acids. On the other hand, when using continuous wavelet transform to extract protein features, the timefrequency matrix is treated as an image, and the image is forcibly compressed to pixels and then subjected to singular value extraction, thus forcing the compression of the image, which loses some of the information of the timefrequency matrix.
In order to improve the prediction accuracy, this paper proposes a feature extraction and fusion method. Besides, with the consideration of the effect affected by the different input order of the two proteins, a “Ytype” BiRNN model is proposed which uses both the backward and forward training methods to train the model. In addition, in order to insure the training time, this paper uses a weightsharing policy to minimize the parameters in the training.
3. Materials and Methods
3.1. Datasets
3.1.1. Benchmark Dataset
We got the original baseline data from the PPI dataset of PAN: http://http://www.csbio.sjtu.edu.cn/bioinf/LR_PPI/Data.htm[48]. The dataset contains 36,630 pairs of positive samples and 36,480 pairs of negative samples. Positive samples were obtained from the Human Protein Reference Database (HPRD) (2007 version). Negative samples (noninteraction pairs) are produced by paired proteins found at different subcellular locations. After eliminating the protein pairs of more than 1200 residue sequences, the benchmark dataset contains 29071 positive samples and 31496 negative samples. We randomly selected 6000 samples as the retention test set for model validation and the remaining samples as the training set. See Table 1 for details.

3.1.2. Species Dataset
This paper selects five PPI datasets in the open source DIP database, the Human dataset, the Saccharomyces cerevisiae (S. Cerevisiae) dataset, the Escherichia coli (E. coli) dataset, the Helicobacter pylori (H. pylori) dataset, the Caenorhabditis elegans (C. elegans) dataset, and the Mus musculus (M. musculus) dataset.
Human dataset: in this work, the proposed method was verified with a high confidence PPI dataset. We collected this dataset from publicly available database of interacting proteins (DIP https://dip.doembi.ucla.edu/), version 20170205. There are 37027 interactive pairs (positive sets) and 37027 noninteractive pairs (negative sets). The interacting pairs in this dataset were measured by the yeast twohybrid assay in the DIP database. For the negative dataset, we followed the previous work [13, 27] and assumed that the proteins in different subcellular compartments do not interact with each other [33]. Specifically, the negative dataset was randomly generated from the SwissProt database by excluding protein sequences which met the following conditions: (i) protein sequences without a certain subcellular location, (ii) protein sequences annotated with more than one subcellular location or “fragment” term, and (iii) protein sequences of less than 64 amino acids.
The interacting proteins of S. cerevisiae, E. coli, H. pylori, C. elegans, and M. musculus interacting proteins are derived from the DIP database. Among them, except for the dataset of M. musculus, the negative set generation method is the same as the Human dataset. The M. musculus dataset contains only positive samples for testing. The specific numbers are shown in Table 2.

3.2. Feature Extraction
Traditional feature extraction methods include dipeptide composition methods, composition conversion distribution description methods, and quasisequenceorder descriptors (QSOD), among which the most commonly used amino acid physicochemical. The properties are hydrophobic , hydrophilic , side chain residue size , polarity , polarizability , solvent accessible surface area SASA, and amino acid side chain net charge number NCI. Many previous studies have proved that it can effectively express protein characteristics. Xu et al. [27] selected the isoelectric point (PI) properties of amino acids for analysis, and the experimental results show that this physicochemical property has an important influence on the polypeptide chain. At the same time, Yu et al. [34] found that the amino acid’s isoelectric point (PI) and ionization equilibrium constant (pKa) properties of the two amino acids can effectively express protein characteristics. Therefore, based on the physicochemical properties of the seven amino acids used in the conventional method, the isoelectric point (PI) and ionization equilibrium constant (pKa) of the amino acid were introduced for characterization. Considering the correlation between PI value and pKa value, this paper combines PI value and pKa value to obtain value that can comprehensively measure amino acid and further extract protein sequence information according to value. The specific calculation formula is as formula (1) and (2):
In summary, the physical and chemical properties of the 20 amino acids are shown in Table 3.

Next, standardize the above 8 physical and chemical properties according to formula (3): wherein represents the ^{th} physicochemical property of the ^{th} amino acid, represents the mean of the ^{th} physicochemical property of 20 amino acids, and represents the standard deviation of the ^{th} physicochemical property of 20 amino acids. Each protein amino acid sequence passes with hydrophobic , hydrophilic , side chain residue size , polarity , polarizability , solvent accessible surface area SASA, amino acid side chain net charge number NCI, and . The values of these eight attributes are converted into a sequence of numbers.
Wang et al. [15] proposed using wavelet to describe protein features and experimentally proved the feasibility of using wavelet transform for protein feature extraction. Discrete wavelet transform (DWT) is an implementation of the wavelet transform that uses discrete wavelet scale sets and translations and decomposes the input signal into mutually orthogonal wavelet sets. In this paper, the Dmeyer wavelet function is used in the discrete wavelet transform part, and Dmeyer is the discrete Meyer wavelet, which is used for the calculation of fast discrete wavelet transform.
This paper converts each amino acid sequence into a digital sequence by eight amino acid properties. By applying the DWT on any of these 8 digital sequences of a protein, each sequenceorder vector is considered as a discrete time series and will put into one halfband highpass filter and one halfband lowpass filter; then, the output sequence of the lowpass filter is then iterated four times to finally obtain 5 subsequences. In each subsequence, three kinds of data are extracted to reflect the internal information of the subsequence, which are (1) the average of the wavelet coefficients in each subsequence, (2) the standard deviation of the wavelet coefficients in each subsequence, and (3) the first four values with the largest absolute value in each subsequence and their relative positions; the relative position is calculated as where is the length of the current subsequence and is the position in the sequence in which the current value is located.
In addition, this paper uses a 25scale mexh continuous wavelet transform to transform each amino acid sequence. The Mexican hat “mexh” wavelet function is the second derivative of the Gauss function, with good regularity, large vanishing moment, and decomposition signal energy concentration. It has localized properties in the time domain and frequency domain. A matrix of ( represents the length of the amino acid sequence) can be extracted by the mexh continuous wavelet transform, and the vector features of the 25 dimensions are extracted according to the decomposition of the singular value matrix. Eventually, each protein sequence is converted into a feature vector of 600 dimensions.
3.3. Deep Neural Network with Local Weight Sharing
Currently, some works have been proposed to predicting PPIs based on neural networks, such as Du et al. [12] and Li et al. [28]. When using a neural network model to predict protein interactions, you can reduce the impact of protein input order on the prediction result by entering two protein features and training separately. This method of inputting proteins can coordinate protein features and protein interactions. The relationship between the feature characteristics improves the overall prediction result. However, neural network models often require too many training parameters, resulting in redundant training time. In this paper, considering the influence of protein input order on prediction results, a deep neural network with local weight sharing is proposed. The network model adopts a “Ytype” neural network model, including a weightsharing BiRNN layer, a buffer layer, and a dense layer, wherein the weightsharing BiRNN layer can reduce the influence of the protein on the predicted input order and accelerate the training. The weightsharing BiRNN to the same parameter that needs to be learned on the BiRNN layer on both sides, that is, the parameter values that need to be learned in the corresponding positions on both sides of the layer, are the same, thereby improving the accuracy of the model prediction result and the model training speed.
The neural network model of local weight sharing is shown as Figure 1. The input layer is divided into two parts containing 1200 neurons, half of which is for the 600 features of protein A and the other half of which is for the 600 features of protein B. The layer of BIRNN is also divided into 2 parts, respectively, for proteins a and b. Each part has 2 layers. Each of the layer contains 512 neurons. Then, the total number of neurons in the BIRNN layer is . The next layer is buffer layer whose function is to connect the parts of a and b into the dense layer for training. In the buffer layer, there will be totally 256 neurons for the proteins a and b which means there will be 256 features to be used. The last layer is the dense layer which contains 3 layers with 32, 8, and 2 neurons in each layer, respectively. The layer with 2 neurons is the last layer in the dense layer which is used to output the classification result. Here, the output (1,0) means there exists interaction, and (0,1) otherwise.
This paper uses the ReLU activation function. Since the ReLU activation function will make the output of a part of neurons zero, the network has a certain sparsity, which reduces the interdependence of parameters to a certain extent, and thus effectively avoids overfitting. The ReLU activation function is used in the model, and the ReLU activation function expression is as
RNN can be regarded as a neural network for information transmission in time series. The depth of the model corresponds to the length of the sequence. The gradient disappearance problem that often occurs in the neural network appears correspondingly in the time dimension of the RNN. In order to solve the gradient dispersion problem of RNN in the time dimension, the researchers proposed the long and short memory unit LSTM, which proved that LSTM is very effective in solving long sequence dependence problems. However, since the use of RNN will cause the gradient to disappear, in order to solve this problem, this paper uses Long ShortTerm Memory (LSTM) [35, 36]; the neurons are as Figure 2.
LSTM is a variant model of RNN. The parameters of an LSTM neuron include input gate, forgetting gate, output gate, and unit and unit input activation vectors, which are represented by , , , and , respectively. The specific expression is as shown in formula (6)–(11), where is the output of the previous cell and is the input of the current cell. represents the sigmoid function.
3.4. Forward and Backward Model Training
In the construction of protein sequences, the traditional method uses the way of binding and sequentially to construct PPI characteristics ( represents the number of features described by the protein), as
In view of the problem that the classification model is easy to overfitting and the order of protein feature binding affects the results, this paper proposes a combination of forward and backward protein feature sequences. The improvement based on the training method is based on the use of the above formula to construct the PPI characteristics. On the basis of the training model in the sequential combination, the PPI feature A in the training set is reinversely combined into B and added to the training set, so that the number of training sets of the training sample reconstruction is doubled and the number of test sets remains the same, as shown in Figure 3.
Every protein interaction data record can be expressed as , where represents the features of protein A, represents the features of protein B, and (l_{1},l_{2}) represents the label. The data of the labels are (1,0) and (0,1). (1,0) indicates that there exists an interaction relationship, and (0,1) means not. In the environment of python3.5, TensorFlow1.12.0 can be used to build a neural network according to the method in the paper. The batch_size is set to 128, learning_rate is set to 0.05, and trains are set for 200 rounds.
4. Experimental Results and Comparison
4.1. Evaluation Measures
In this experiment, we used a fivefold crossvalidation process to prevent the calculation method from overfitting on the dataset or misjudging the result due to unbalanced data extraction. The evaluation criteria for the algorithm use several widely used parameters: accuracy, recall, specificity, precision, and MCC (Mathew’s correlation coefficient). Some of the above parameters are defined as formula (13)–(17): where the truepositive (TP) value denotes the number of true samples which are predicted correctly; the falsenegative (FN) value is the number of true samples predicted to be noninteracting pairs incorrectly; the falsepositive (FP) value is the number of true noninteracting pairs predicted to be PPIs falsely; and the truenegative (TN) value is the number of true noninteracting pairs predicted correctly.
4.2. Training and Validation on the Benchmark Dataset
We randomly divide the training set into five sets by fivefold crossvalidation, and crosstrain. Finally, the best performance model is tested under the holdout test set. The specific data are shown in Table 4. We can see that the accuracy of the fivefold crossvalidation is almost the same. For testing the holdout set, the model for test set 2 is used which achieved an accuracy of 99.57%, recall of 99.36%, sensitivity of 99.76%, precision of 99.74%, MCC of 99.14%, and AUC of 99.56%.

To further confirm the effectiveness of the model, we compare the experimental results by constantly switching the ratio of training set to test set, as shown in Table 5. We can see that the model can maintain a robust effect in different proportions.

4.3. Multispecies Training Evaluation
To further explore the validity of the model, we used the same method to conduct experiments on different species datasets. Similarly, the training sets of each species were randomly divided into five sets for crosstraining, and the best model was selected for testing. Table 6 shows the results. In each training set, the performance remains robustness, and the data fluctuation does not exceed 1%. Finally, we can see that the accuracy in E. coli, C. elegans, S. cerevisiae, and human is 95.04%, 98.14%, 99.96%, and 99.94%, respectively.

4.4. Compare the Performance of Feature Extraction and Fusion
In order to verify the feature extraction and fusion, under the same dataset, compare the performance of the feature extraction and fusion in our work and the works (signified as AC) which uses the amino acid composition based on position information to extract the features, as well as the works (signified as WT) which uses the feature extraction method based on multiattribute wavelet transform. For just comparing the performance of the feature extraction and fusion, we use the SVM classifier to train the model. The result can be seen in Figure 4. From the result, we can see that our work has a higher accuracy than others.
4.5. Comparison with Other Methods
In this section, we compare the performance of the proposed method with other different methods, by using the S. cerevisiae dataset, as shown in Table 7. We can see that in the counterpart methods, the accuracy is between 83.35% and 98.78%, most of them have exceeded 90%, and the highest accuracy rate is 98.78%. Compared with our method, except that precision in Ref. [28], the other attributes are higher than other methods, the minimum difference of accuracy is 1.16%, the minimum difference of sensitivity is 1.72%, the minimum difference of precision is 0%, and the minimum difference of MCC is 1.29%.

4.6. Comparison Time Performance with Other Deep Learning Methods
Based on the same machine configuration, this paper also compares the training time required for different training models. Table 8 lists the comparison of the training time required between different algorithms. It can be seen that the training time of the partialinput twoway cyclic neural network shared by local weights is almost the same as that of the fivelayer fully connected neural network under the sameorder neurons. At the same time, compared with the DeepPPI training model proposed by Du et al. [12], the training speed is increased by 70 seconds, which is 251 seconds higher than that of the deep neural network without local weight sharing.

4.7. Training Other Models by Using the Features of Forward and Backward Reconstruction
The experimental results are shown in Table 9. We can see that the training methods based on the characteristics of forward and backward training can improve the performance of Xu’s method which can improve the accuracy by 1.26%, recall by 1.98%, sensitivity by 0.99%, precision by 2.13%, and MCC by 2.91%.

4.8. CrossSpecies CrossValidation Evaluation
According to Table 10, we can see that when the training set is human, the test results are better, most of which are more than 90 percent. However, other species can only play a certain role in the M. musculus dataset for testing. For data analysis, since human datasets are almost the sum of other kinds of datasets, robust models can be effectively trained. The best is 98.39%.

5. Conclusion and Future Work
The protein feature extraction and fusion method based on the combination of discrete wavelet transform and continuous wavelet transform is used to describe the protein amino acid sequence to improve the variety of the features to be considered in the prediction. When constructing protein interaction features, the common method is to directly combine the two protein features in sequence. However, this method of constructing interaction features doubles the length of the feature vector, which is likely to cause overfitting under limited protein interaction datasets. At the same time, because two interacting protein features are equally important, constructing mutual features of proteins is likely to result in different predictions due to different binding sequences. In this paper, based on the use of sequential binding protein features, the forward and backward model training methods are proposed to address this issue. Besides, a weightsharing policy is proposed to minimize the parameters in the training in order to insure the training time. This paper defines protein interactions as a binary classification problem. In the future, we can consider whether the problem can be mapped to a multiclassification problem where the prediction result can be a certain type of interaction. Besides, currently, expressing the sequence by using an appropriate way can be used to solve the overfitting problem. Then, in the future, the coding method for expressing the features will be studied.
Data Availability
The data supporting the conclusions are presented in the main article. The code can be found in https://github.com/joddiedai/ppi.
Conflicts of Interest
The authors declare no competing interests.
Authors’ Contributions
Lei Yang reviewed the paper. Yukun Han wrote the paper. Huixue Zhang analyzed the experiment result. Yu Dai designed the main idea and supervised the study. Wenlong Li did the experiment.
Acknowledgments
This work was supported in part by the National Key Research and Development Plan of China (Grant No. 2017YFB1401803).
References
 P. Uetz, L. Giot, G. Cagney et al., “A comprehensive analysis of proteinprotein interactions in Saccharomyces cerevisiae,” Nature, vol. 403, no. 6770, pp. 623–627, 2000. View at: Publisher Site  Google Scholar
 T. Ito, T. Chiba, R. Ozawa, M. Yoshida, M. Hattori, and Y. Sakaki, “A comprehensive twohybrid analysis to explore the yeast protein interactome,” Proceedings of the National Academy of Sciences, vol. 98, no. 8, pp. 4569–4574, 2001. View at: Publisher Site  Google Scholar
 H. Zhu, M. Bilgin, R. Bangham et al., “Global analysis of protein activities using proteome chips,” Science, vol. 293, no. 5537, pp. 2101–2105, 2001. View at: Publisher Site  Google Scholar
 A. H. Y. Tong, B. Drees, G. Nardelli et al., “A combined experimental and computational strategy to define protein interaction networks for peptide recognition modules,” Science, vol. 295, no. 5553, pp. 321–324, 2001. View at: Publisher Site  Google Scholar
 E. Harlow, P. Whyte, B. R. Franza Jr., and C. Schley, “Association of adenovirus earlyregion 1A proteins with cellular polypeptides,” Molecular and Cellular Biology, vol. 6, no. 5, pp. 1579–1589, 1986. View at: Publisher Site  Google Scholar
 C. Williams and T. A. Addona, “The integration of SPR biosensors with mass spectrometry: possible applications for proteome analysis,” Trends in Biotechnology, vol. 18, no. 2, pp. 45–48, 2000. View at: Publisher Site  Google Scholar
 G. Multhaup, D. Strausak, K. D. Bissig, and M. Solioz, “Interaction of the CopZ copper chaperone with the CopA copper ATPase of Enterococcus hirae assessed by surface plasmon resonance,” Biochemical and Biophysical Research Communications, vol. 288, no. 1, pp. 172–177, 2001. View at: Publisher Site  Google Scholar
 R. Yang, C. Zhang, R. Gao, and L. Zhang, “An ensemble method with hybrid features to identify extracellular matrix proteins,” PLoS One, vol. 10, no. 2, article e0117804, 2015. View at: Publisher Site  Google Scholar
 J. Wang, Y. Li, X. Liu, Q. Dai, Y. Yao, and P. He, “Highaccuracy prediction of protein structural classes using PseAA structural properties and secondary structural patterns,” Biochimie, vol. 101, no. 1, pp. 104–112, 2014. View at: Publisher Site  Google Scholar
 S. Zhang, Y. Liang, and X. Yuan, “Improving the prediction accuracy of protein structural class: approached with alternating word frequency and normalized LempelZiv complexity,” Journal of Theoretical Biology, vol. 341, no. 1, pp. 71–77, 2014. View at: Publisher Site  Google Scholar
 P. N. Palma, L. Krippahl, J. E. Wampler, and J. J. G. Moura, “BiGGER: a new (soft) docking algorithm for predicting protein interactions,” Proteins, vol. 39, no. 4, pp. 372–384, 2000. View at: Publisher Site  Google Scholar
 X. Du, S. Sun, C. Hu, Y. Yao, Y. Yan, and Y. Zhang, “DeepPPI: boosting prediction of protein–protein interactions with deep neural networks,” Journal of Chemical Information and Modeling, vol. 57, no. 6, pp. 1499–1510, 2017. View at: Publisher Site  Google Scholar
 T. Wang, L. Li, Y. A. Huang, H. Zhang, Y. Ma, and X. Zhou, “Prediction of proteinprotein interactions from amino acid sequences based on continuous and discrete wavelet transform features,” Molecules, vol. 23, no. 4, p. 823, 2018. View at: Publisher Site  Google Scholar
 Z. H. You, Y. K. Lei, L. Zhu, J. Xia, and B. Wang, “Prediction of proteinprotein interactions from amino acid sequences with ensemble extreme learning machines and principal component analysis,” BMC Bioinformatics, vol. 14, no. S8, 2013. View at: Publisher Site  Google Scholar
 H. Wang, H. Huang, C. Ding, and F. Nie, “Predicting protein–protein interactions from multimodal biological data sources via nonnegative matrix trifactorization,” Journal of Computational Biology, vol. 20, no. 4, pp. 344–358, 2013. View at: Publisher Site  Google Scholar
 J. R. Bock and D. A. Gough, “Wholeproteome interaction mining,” Bioinformatics, vol. 19, no. 1, pp. 125–134, 2003. View at: Publisher Site  Google Scholar
 J. Shen, J. Zhang, X. Luo et al., “Predicting proteinprotein interactions based only on sequences information,” Proceedings of the National Academy of Sciences, vol. 104, no. 11, pp. 4337–4341, 2007. View at: Publisher Site  Google Scholar
 C. Y. Yu, L. C. Chou, and D. Chang, “Predicting proteinprotein interactions in unbalanced data using the primary structure of proteins,” BMC Bioinformatics, vol. 11, no. 1, p. 167, 2010, Article No.167. View at: Publisher Site  Google Scholar
 J. Grassmann, M. Reczko, S. Suhai, and L. Edler, “Protein fold class prediction: new methods of statistical classification,” in International Conference on Intelligent Systems for Molecular Biology, pp. 106–112, Heidelberg, Germany, 1999 Aug 6. View at: Google Scholar
 M. Reczko and H. Bohr, “The DEF data base of sequence based protein fold class predictions,” Nucleic Acids Research, vol. 22, no. 17, pp. 3616–3619, 1994. View at: Google Scholar
 I. Dubchak, I. Muchnik, C. Mayor, I. Dralyuk, and S. H. Kim, “Recognition of a protein fold in the context of the SCOP classification,” Proteins: Structure, Function, and Genetics, vol. 35, no. 4, pp. 401–407, 1999. View at: Publisher Site  Google Scholar
 K. C. Chou, “Prediction of protein subcellular locations by incorporating quasisequenceorder effect,” Biochemical and Biophysical Research Communications, vol. 278, no. 2, pp. 477–483, 2000. View at: Publisher Site  Google Scholar
 F.M. Li and Q.Z. Li, “Predicting Protein Subcellular Location Using Chous Pseudo Amino Acid Composition and Improved Hybrid Approach,” Protein & Peptide Letters, vol. 15, no. 6, pp. 612–616, 2008. View at: Publisher Site  Google Scholar
 J. Y. An, F. R. Meng, Z. H. You, Y. H. Fang, Y. J. Zhao, and M. Zhang, “Using the relevance vector machine model combined with local phase quantization to predict proteinprotein interactions from protein sequences,” BioMed Research International, vol. 2016, Article ID 4783801, 9 pages, 2016. View at: Publisher Site  Google Scholar
 Y. A. Huang, Z. H. You, X. Gao, L. Wong, and L. Wang, “Using weighted sparse representation model combined with discrete cosine transformation to predict proteinprotein interactions from protein sequence,” BioMed Research International, vol. 2015, Article ID 902198, 10 pages, 2015. View at: Publisher Site  Google Scholar
 Z. H. You, K. C. C. Chan, and P. Hu, “Predicting proteinprotein interactions from primary protein sequences using a novel multiscale local feature representation scheme and the random forest,” PLoS One, vol. 10, no. 5, article e0125811, 2015. View at: Publisher Site  Google Scholar
 P. Xu, J. Luo, and T. Dou, “Predict proteinprotein interactions from protein primary sequences: using wavelet transform combined with stacking algorithm,” PeerJ Preprints, vol. 5, article e2964v1, 2017. View at: Google Scholar
 H. Li, X. J. Gong, H. Yu, and C. Zhou, “Deep neural network based predictions of protein interactions using primary sequences,” Molecules, vol. 23, no. 8, p. 1923, 2018. View at: Publisher Site  Google Scholar
 X. Du, J. Cheng, T. Zheng, Z. Duan, and F. Qian, “A novel feature extraction scheme with ensemble coding for protein–protein interaction prediction,” International Journal of Molecular Sciences, vol. 15, no. 7, pp. 12731–12749, 2014. View at: Publisher Site  Google Scholar
 B. Zhang, J. Li, L. Quan, Y. Chen, and Q. Lü, “Sequencebased prediction of proteinprotein interaction sites by simplified long shortterm memory network,” Neurocomputing, vol. 357, pp. 86–100, 2019. View at: Publisher Site  Google Scholar
 Y. Yao, X. Du, Y. Diao, and H. Zhu, “An integration of deep learning with feature embedding for proteinprotein interaction prediction,” PeerJ, vol. 7, article e7126, 2019. View at: Publisher Site  Google Scholar
 M. Chen, C. J. . T. Ju, G. Zhou et al., “Multifaceted protein–protein interaction prediction based on Siamese residual RCNN,” Bioinformatics, vol. 35, no. 14, pp. i305–i314, 2019. View at: Publisher Site  Google Scholar
 C. J. Shin, S. Wong, M. J. Davis, and M. A. Ragan, “Proteinprotein interaction as a predictor of subcellular location,” BMC Systems Biology, vol. 3, no. 1, p. 28, 2009, Article No. 28. View at: Publisher Site  Google Scholar
 L. Yu, Y. Zhang, I. Gutman, Y. Shi, and M. Dehmer, “Protein sequence comparison based on physicochemical properties and the positionfeature energy matrix,” Scientific Reports, vol. 7, no. 1, 2017. View at: Publisher Site  Google Scholar
 F. A. Gers, J. Schmidhuber, and F. Cummins, “Learning to forget: continual prediction with LSTM,” Neural Computation, vol. 12, no. 10, pp. 2451–2471, 2000. View at: Publisher Site  Google Scholar
 Y. Z. Zhou, Y. Gao, and Y. Y. Zheng, “Prediction of proteinprotein interactions using local description of amino acid sequence,” Advances in Computer Science and Education Applications, vol. 202, pp. 254–262, 2011. View at: Publisher Site  Google Scholar
Copyright
Copyright © 2020 Lei Yang et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.