BioMed Research International

BioMed Research International / 2021 / Article
Special Issue

Scalable Machine Learning Algorithms in Computational Biology and Biomedicine 2021

View this Special Issue

Research Article | Open Access

Volume 2021 |Article ID 9923112 | https://doi.org/10.1155/2021/9923112

Guohua Huang, Qingfeng Shen, Guiyang Zhang, Pan Wang, Zu-Guo Yu, "LSTMCNNsucc: A Bidirectional LSTM and CNN-Based Deep Learning Method for Predicting Lysine Succinylation Sites", BioMed Research International, vol. 2021, Article ID 9923112, 10 pages, 2021. https://doi.org/10.1155/2021/9923112

LSTMCNNsucc: A Bidirectional LSTM and CNN-Based Deep Learning Method for Predicting Lysine Succinylation Sites

Academic Editor: Quan Zou
Received15 Mar 2021
Revised25 Apr 2021
Accepted03 May 2021
Published29 May 2021

Abstract

Lysine succinylation is a typical protein post-translational modification and plays a crucial role of regulation in the cellular process. Identifying succinylation sites is fundamental to explore its functions. Although many computational methods were developed to deal with this challenge, few considered semantic relationship between residues. We combined long short-term memory (LSTM) and convolutional neural network (CNN) into a deep learning method for predicting succinylation site. The proposed method obtained a Matthews correlation coefficient of 0.2508 on the independent test, outperforming state of the art methods. We also performed the enrichment analysis of succinylation proteins. The results showed that functions of succinylation were conserved across species but differed to a certain extent with species. On basis of the proposed method, we developed a user-friendly web server for predicting succinylation sites.

1. Introduction

Protein post-translational modification (PTM) refers to the chemical interaction occurring prior to protein biosynthesis and after mRNAs are translated into polypeptide chains. PTM has different categories and is very prevalent in the cells. More than 450 categories of PTMs were discovered to date, such as phosphorylation, methylation, and acetylation [13]. PTM increases diversity of protein structures and functions, viewed as one of most regulating mechanisms in the cellular process. Lysine succinylation is a type of protein TPMs, in which a succinyl group (-CO-CH2-CH2-CO2H) is attached to lysine residue of proteins [4]. Succinylation is reversible, dynamic, and evolutionarily conserved, widely existing in the prokaryote and the eukaryotes cells [5, 6]. The succinylation of proteins induces shift in the charge and the structural alteration and thus would yield effects on functions of proteins [6]. Growing evidences also showed aberrant succinylations were involved in the pathogenesis of some diseases including cancers [7], metabolism disease [8, 9], and nervous system diseases [10]. Thus, identifying succinylation sites and understanding its mechanism are crucial to develop drugs for related diseases.

Identifying succinylation sites has two main routes: experimental and computational methods. The experimental methods were represented by mass spectrometry, which contributed to the validation of succinylation and collection of first-hand data. On the other hand, the experimental methods are labor-intensive and time-consuming without assist of the computational methods. The computational methods are based on data yielded by the experimental methods and build machine learning-based models to predict new succinylations. Therefore, identifying succinylation is a cyclic iterative process from experiment to computation and again from computation to experiment. We focused on the computational methods to predict succinylation. In the past decades, more than ten computational methods have been developed for identifying succinylation [1129]. Most of these computational methods extracted features directly from protein sequences, which were subsequently used for training model. For example, Zhao et al. [11] used the auto-correlation functions, the group weight-based encoding, the normalized van der Waals volume, and the position weight amino acid composition. Kao et al. [25] exploited the amino acid composition and informative -spaced amino acid pairs. Xu et al. [12] and Jia et al. [13, 19] employed pseudo amino acid composition. Dehzangi et al. [23] exploited the structure information. Hasan et al. [28] compared 12 types of feature as well as two learning methods: random forest and support vector machine for succinylation prediction. Different features have different performance with species. So does the learning methods. The best performance was no more than 0.83 AUC (area under receiver operating characteristic curve) for independent test. Like sentences of language, the protein sequences should have semantic. However, all the methods above failed to seize semantic relationship hidden among residues. Thapa et al. [29] presented a convolutional neural network- (CNN-) based deep learning method DeepSuccinylSite for predicting succinylation. Different from traditional methods, the DeepSuccinylSite exploited word embedding which translated word into vector, which was an extensively used method in the field of natural language process. The CNN is a widely used method to extract local features especially in the field of image processing. Inspired by the DeepSuccinylSite and loss of semantic relationship between residues, we fused long short-term memory (LSTM) and CNN into a deep learning method for succinylation prediction.

2. Data

All the succinylated proteins were downloaded from the PLMD (Protein Lysine Modifications Database) database which is dedicated to specifically collect protein lysine modification [3032]. The PLMD has evolved to version 3.0, housing 284780 modification events in 53501 proteins for 20 types of lysine modification. We extracted 6377 proteins containing 18593 succinylation sites. To remove dependency of the proposed method on the homology, we used the software CD-Hit [33, 34] to cluster 6377 protein sequences. The sequence identify cut-off was set to 0.4, and we obtained 3560 protein sequences, of which any two kept sequence similarity less than 0.4. We randomly divided these 3560 proteins into the training and the testing samples at the ratio of training to testing 4 : 1, resulting in 712 testing and 2848 training sequences. For each protein sequence, we extracted all the peptides which centered the lysine residue with 15 amino acid residues in the downstream/upstream of it. For peptides less than 15 amino acid residues, we prefixed or suffixed “” to supply it. The length of the amino acids is influential in prediction of succinylation sites. The short amino acid peptides would miss key information, while the long peptides would include noise or redundancy. Whether the short or the long peptides would cause low accuracy of prediction. Among methods to predict succinylation sites, iSuc-PseAAC [12] adopted the shorter peptides of 15 amino acid residues; SuccinSite2.0 [20] and GPSuc [22] adopted the longer 41 amino acid residues, while the most methods including SSEvol-Suc [23], Success [24], iSuc-PseOpt [13], pSuc-Lys [19], SucStruct [18], and PSSM-Suc [17] adopted peptides of 31 amino acid residues, which is of moderate length. Thus, we chose 31 amino acid residues as basic peptides. The peptides with succinylation sites were viewed positive samples and the others as negative ones. For the training set, the negative samples extremely outnumbered the positive ones. Unbalanced training set would cause preference to negative samples in the process of prediction. Therefore, we randomly sampled the same size of negative examples as the positive ones. Finally, the training set comprised 6512 positive and 6512 negative samples, while the testing set 1479 positive and 16457 negative samples. All the experimental data are freely available to scientific communities.

3. Method

As shown in Figure 1, the proposed deep learning network consisted mainly of embedding, 1D convolution, pooling, bidirectional LTSM, dropout, flatten, and fully connected layers. Peptides with 31 amino acid residues were entered to the embedding layer and were translated into vectors with shape of (31, 64). Then, two different network structures, respectively, took the embedding as input, and their outputs were concatenated as input to the fully connected layer. One structure was the convolution neural network, and another was the bidirectional LSTM neural network. The final output was a neuron representing probability of belonging to the positive sample. The parameters and the shape of output of each layers in the deep neural network are listed in Table 1. The total number of trainable parameters is 336,897.


LayersParametersOutput

Embedding1472(None, 31, 64)
Bidirectional LSTM197632(None, 31, 256)
Dropout0(None, 31, 256)
Flatten0(None, 7936)
1D convolution10272(None, 27, 32)
Pooling0(None, 32)
Dense (16)127504(None, 16)
Dense (1)17(None, 1)

3.1. Embedding Layer

Most machine learning-based methods for predicting protein post-translational modification generally required an encoding step which translated sequences into vector representation. For example, the frequently used encoding schemes included position specific scoring matrix [35], amino acid composition, composition of -space amino acid pair [14], and pseudo amino acid composition [36]. For sequences of text, these methods might lose hidden semantic. The word2vec [37, 38] is different from the above methods, embedding word into vector. The word2vec is capable of extracting semantic of word. An interesting example is that . Similar to the word2vec [37, 38], the embedding layer translated words into vector representations. In this method, the character of amino acid corresponds to word.

3.2. 1D Convolution Layer

The convolution neural network (CNN) proposed by LeCun et al. [39, 40] is a feed forward network. Compared with the conventional neural network, the CNN has two notable properties: local connectivity and parameter sharing. The local connectivity lies that two neighboring layers are not fully connected but locally connected. That is to say, the neuron in a layer is not connected to all neurons in the neighboring layers. The CNN implemented the parameter sharing via the filter (also called convolution kernel). The filter slides on the image and convoluted with all sections in image. The filter is shared by the image. In the last ten years, many deep convolution neural networks such as AlexNet [41], VGG [42], GoogleNet [43], and ResNet [44] have been proposed and applied to computer vision. The CNN achieved significant advance in terms of classification error in comparison with the previous deep neural network. The convolution is of 1-dimension, 2-dimension, or more than 2 dimensions. Here, we used 1D convolution. Suppose a discrete sequence was , and the convolution kernel was . The 1D convolution product of and was expressed by

where was the stride of convolution and was the length of the output sequence. Generally, was the most integer less than or equal to .

3.3. Pooling Layer

The pooling operation firstly appeared in the AlexNet [41] and is increasingly becoming one of components of the deep CNN architecture. The pooling operation has such categories as max pooling, min pooling, and mean pooling. The role of pooling operation included removal of redundancy information and reduction of overfitting. Here, we used the max pooling operation. Given an -channel input , the max pooling operation was defined by

3.4. Bidirectional LSTM Layer

Recurrent neural network (RNN) [45, 46] is a different framework of neural network from multiple layer perception. The RNN shares weights and is especially suitable to the field of sequence analysis such as language translation and semantic understanding. An unfolded RNN model was shown in Figure 2(a). The hidden state at the time step was not only dependent on the current input but also on the previous hidden state, which was computed by where was an activation function and was a bias. The output at the time step was computed by where was also an activation function and was a bias. For long sequences, there was a fatal question with the RNN, i.e., vanishing gradient. Among all the solutions to the vanishing gradient, the LSTM [47] is one of the better. The LSTM contains a candidate memory cell and three gates: forget gate, input gate, and output gate, as shown in Figure 2(b). The forget gate , the input gate , and the output gate at the time step were computed, respectively, by where and were weights of the LSTM from input to forget gate and from the hidden state to the forget gate, respectively. and were link weights from input to input gate and from the hidden state to the input gate, respectively. and were link weights from input to output gate and from the hidden state to the output gate, respectively. , , and were the bias of the forget and the input and the output gate, respectively. was the activation function. The candidate memory cell was calculated by where and were weights of the LSTM from input to the candidate memory and from the hidden state to the candidate memory, respectively, and was the bias. The memory cell at the time step was computed by where was defined as element-wise multiplication. The hidden state was updated by

The previous RNN was forward. The output at the time step was only dependent on the preceding inputs and the hidden state. In fact, the output might be relevant to the latter input and the hidden state. Schuster et al. [48] proposed a bidirectional RNN to model this relationship, showed in Figure 2(c). The forward hidden state at the time step was computed by while the backward hidden state was computed by

The output at the time step was computed by

3.5. Dropout Layer

The deep neural network is prone to lead to overfitting when the number of training samples was too less. To deal with this issue, Hinton et al. [49] proposed the dropout concept. Due to its effect and efficiency, the dropout is increasingly becoming the frequently used trick in the deep learning area [41, 5053]. The neurons were dropped out at a certain rate of dropout, and parameters of only preserved neurons were updated in the training stage, while all the neurons were used in the predicting stage.

3.6. Flatten Layer and Fully Connected Layer

The role of flatten layer was only to convert the data into one-dimension and then facilitated connection of the fully connected layer. No parameters were trainable in the flatten layer. The fully connected layer was similar to hidden layer in the MLP, each neuron connected to the neurons in the preceding layer.

4. Metrics

We adopted to evaluate the predicted result these frequently used metrics in the binary classification questions such as sensitivity (SN), specificity (SP), accuracy (ACC), and Matthews correlation coefficient (MCC), which were defined by where TP and TN were defined as numbers of the true positive and the true negative samples, respectively, FP and FN, respectively, as numbers of the false positive and the false negative samples in the prediction. SN reflected the accuracy of the correctly predicted positive samples, SP accuracy of the correctly predicted negative samples, and ACC the average accuracy of the correctly predicted samples. SN, SP, and ACC ranged from 0 to 1, larger meaning better performance. MCC was Matthews correlation coefficient, representing correlation between the true class and the predicted class. MCC ranged from -1 to 1. 1 meant perfect prediction, 0 random prediction, and -1 meant that the prediction was completely opposite to the true.

5. Results

Table 2 showed the predicting performance of the trained model on the 712 testing sequences. Although more than ten approaches or tools for predicting succinylation have been proposed in the past ten years, either they did not provide online predicting server or the web server could not work. We compared the proposed method to three methods whose web predicting server still can work [28]: SuccinSite [15], iSuc-PseAAC [12], and DeepSuccinylSite [29]. 712 testing sequences were used to examine three approaches. Among 712 testing sequences, at least 225 sequences repeated in the training set of the SuccinSite, and at least 223 repeated in the training set of DeepSuccinylSite. These minus 225 sequences were used to examine the SuccinSite and these minus 223 sequences to test the DeepSuccinylSite. iSuc-PseAAC [12] obtained best SP and best ACC but worst SN and worst MCC. The SuccinSite [15] reached better SP and better ACC but worse MCC and worse SN. The iSuc-PseAAC [12] and the SuccinSite [15] were in favor of predicting the negative samples. The DeepSuccinylSite [29] was better than the LSTMCNNsucc in terms of SN, worse than the LSTMCNNsucc in terms of sp. The overall performance of the LSTMCNNsucc was slightly better than that of the DeepSuccinylSite.


MethodSNSPACCMCC

LSTMCNNsucc0.59160.79570.77890.2508
SuccinSite [15]0.39770.86350.82720.1925
iSuc-PseAAC [12]0.12580.89290.82960.0165
DeepSuccinylSite [29]0.74380.68790.69230.2438

5.1. Functional Analysis

We used the statistical over-representation test of gene list analysis in the PANTHER classification system [54, 55] to perform function enrichment analysis of the succinylated proteins. The significant biological process, the molecular function, and the cellular component terms ( value0.01) were listed in the supplementary materials 1 and 2. For five species, Escherichia coli (E. coli), Homo sapiens (H. sapiens), Mus musculus (M. musculus), Mycobacterium tuberculosis (M. tuberculosis), and Saccharomyces cerevisiae (S. cerevisiae), they shared some common functions, but they had also own specific functions. The numbers of shared terms among five species are shown in Figure 3. H. sapiens and M. musculus shared 36 significant biological process terms and 35 cellular component terms, much more than the numbers of shared terms between any other two species (Figures 3(a) and 3(b)). Five species shared eight biological process GO terms: “biosynthetic process (GO:0009058)”, “carboxylic acid metabolic process (GO:0019752)”, “organic acid metabolic process (GO:0006082)”, “organic substance biosynthetic process (GO:1901576)”, “organonitrogen compound biosynthetic process (GO:1901566)”, “organonitrogen compound metabolic process (GO:1901564)”, “oxoacid metabolic process (GO:0043436)”, and “small molecule metabolic process (GO:0044281)”; 5 cellular component GO terms: “cytoplasm (GO:0005737)”, “cytoplasmic part (GO:0044444)”, “cytosol (GO:0005829)”, “intracellular (GO:0005622)”, and “intracellular part (GO:0044424)”; and two molecular function GO terms: “catalytic activity (GO:0003824)”, and “molecular_function (GO:0003674)”. H. sapiens had much more own specific functions than other species, with 75 specific biological process GO terms, 14 GO cellular component terms, and 21 molecular function GO terms. No specific functions existed in both M. tuberculosis and S. cerevisiae whether for biological process, cellular component, or molecular functions.

We also performed enrichment analysis of Kyoto Encyclopedia of Genes and Genomes (KEGG) pathway by functional annotation in the DAVID tool [56, 57] to investigate in which pathway the succinylated proteins were involved. The statistically significant KEGG terms () are listed in Table 3. Different species were involved in some identical pathways. For example, both metabolic pathways and biosynthesis of antibiotics were enriched in the succinylated proteins for five species, implying the universal role of succinylation. On the other hand, different pathways were involved in different species. H. sapiens and M. musculus shared more pathway and had more pathways than other three species, implying species-specific role of the succinylation.


SpeciesKEGG termsBenjamini

E. coliMetabolic pathways3.30-08
Biosynthesis of amino acids1.00-06
Biosynthesis of secondary metabolites2.40-04
Biosynthesis of antibiotics7.40-04
Lysine biosynthesis3.30-03

H. sapiensBiosynthesis of antibiotics3.70-10
Metabolic pathways2.80-09
Ribosome3.40-08
Valine, leucine, and isoleucine degradation1.30-06
Carbon metabolism6.20-06
Oxidative phosphorylation1.10-05
Parkinson’s disease2.60-05
Citrate cycle (TCA cycle)1.00-04
Huntington’s disease4.10-04
Alzheimer’s disease7.80-04
Aminoacyl-tRNA biosynthesis1.00-03
Butanoate metabolism3.40-03
Proteasome8.20-03

M. musculusMetabolic pathways6.20-26
Parkinson’s disease8.50-11
Oxidative phosphorylation3.40-10
Nonalcoholic fatty liver disease (NAFLD)1.00-09
Huntington’s disease2.80-09
Alzheimer’s disease1.40-08
Ribosome3.30-07
Peroxisome1.80-06
Glycine, serine, and threonine metabolism1.50-05
Pyruvate metabolism9.00-05
Propanoate metabolism2.40-04
Valine, leucine, and isoleucine degradation1.90-03
Glyoxylate and dicarboxylate metabolism3.10-03
Biosynthesis of antibiotics5.60-03

M. tuberculosisMetabolic pathways1.00-04
Microbial metabolism in diverse environments2.50-04
Biosynthesis of antibiotics4.40-04
Biosynthesis of secondary metabolites1.00-02
Propanoate metabolism1.00-02

S. cerevisiaeMetabolic pathways5.20-05
Biosynthesis of amino acids3.30-04
2-Oxocarboxylic acid metabolism7.90-04
Biosynthesis of antibiotics3.50-03
Oxidative phosphorylation3.50-03

5.2. LSTMCNNsucc Web Server

We built a web server of the proposed LSTMCNNsucc at http://8.129.111.5/. Users either directly input protein sequences in a fasta format or upload a file of fasta format to perform prediction. When both protein sequences and files were submitted, the file was given to priority of prediction.

6. Conclusion

We presented a bidirectional LSTM and CNN-based deep learning method for predicting succinylation sites. The method absorbed semantic relationship hidden in the succinylation sequences, outperforming state-of-the-art method. The functions of succinylation proteins were conserved to a certain extent across species but also were species-specific. We also implemented the proposed method into a user-friendly web server which is available at http://8.129.111.5/.

Data Availability

The experimental succinylation and nonsuccinylation sites used to support the findings of this study have been deposited in the website http://8.129.111.5/ and are freely available to all scientific communities.

Conflicts of Interest

The authors declare that no competing interest exists.

Acknowledgments

This work is supported by the National Natural Science Foundation of China (11871061, 61672356), by the Scientific Research Fund of Hunan Provincial Education Department (18A253), by the open project of Hunan Key Laboratory for Computation and Simulation in Science and Engineering (2019LCESE03), and by Shaoyang University Graduate Research Innovation Project (CX2020SY060).

Supplementary Materials

Supplementary 1. KEGG enrichmeent analysis of KEGG pathway for 5 species.

Supplementary 2. GO enrichment analysis of GO terms for 5 species.

References

  1. E. S. Witze, W. M. Old, K. A. Resing, and N. G. Ahn, “Mapping protein post-translational modifications with mass spectrometry,” Nature Methods, vol. 4, no. 10, pp. 798–806, 2007. View at: Publisher Site | Google Scholar
  2. B. S. Sharma, “Post-translational modifications (PTMs), from a cancer perspective: an overview,” Oncogen Journal, vol. 2, no. 3, p. 12, 2019. View at: Google Scholar
  3. G. Huang and X. Li, “A review of computational identification of protein post-translational modifications,” Mini-Reviews in Organic Chemistry, vol. 12, no. 6, pp. 468–480, 2015. View at: Publisher Site | Google Scholar
  4. Z. Zhang, M. Tan, Z. Xie, L. Dai, Y. Chen, and Y. Zhao, “Identification of lysine succinylation as a new post-translational modification,” Nature Chemical Biology, vol. 7, no. 1, pp. 58–63, 2011. View at: Publisher Site | Google Scholar
  5. J. Du, Y. Zhou, X. Su et al., “Sirt5 is a NAD-dependent protein lysine demalonylase and desuccinylase,” Science, vol. 334, no. 6057, pp. 806–809, 2011. View at: Publisher Site | Google Scholar
  6. A. Sreedhar, E. K. Wiese, and T. Hitosugi, “Enzymatic and metabolic regulation of lysine succinylation,” Genes & Diseases, vol. 7, no. 2, pp. 166–171, 2020. View at: Publisher Site | Google Scholar
  7. Y. Xiangyun and N. Xiaomin, “Desuccinylation of pyruvate kinase M2 by SIRT5 contributes to antioxidant response and tumor growth,” Oncotarget, vol. 8, no. 4, pp. 6984–6993, 2017. View at: Publisher Site | Google Scholar
  8. M. Yang, Y. Wang, Y. Chen et al., “Succinylome analysis reveals the involvement of lysine succinylation in metabolism in pathogenic Mycobacterium tuberculosis,” Molecular & Cellular Proteomics, vol. 14, no. 4, pp. 796–811, 2015. View at: Publisher Site | Google Scholar
  9. Y. Yang and G. E. Gibson, “Succinylation links metabolism to protein functions,” Neurochemical Research, vol. 44, no. 10, pp. 2346–2359, 2019. View at: Publisher Site | Google Scholar
  10. G. E. Gibson, H. Xu, H. L. Chen, W. Chen, T. T. Denton, and S. Zhang, “Alpha-ketoglutarate dehydrogenase complex-dependent succinylation of proteins in neurons and neuronal cell lines,” Journal of Neurochemistry, vol. 134, no. 1, pp. 86–96, 2015. View at: Publisher Site | Google Scholar
  11. X. Zhao, Q. Ning, H. Chai, and Z. Ma, “Accurate in silico identification of protein succinylation sites using an iterative semi-supervised learning technique,” Journal of Theoretical Biology, vol. 374, pp. 60–65, 2015. View at: Publisher Site | Google Scholar
  12. Y. Xu, Y. X. Ding, J. Ding, Y. H. Lei, L. Y. Wu, and N. Y. Deng, “iSuc-PseAAC: predicting lysine succinylation in proteins by incorporating peptide position-specific propensity,” Scientific Reports, vol. 5, no. 1, article 10184, 2015. View at: Publisher Site | Google Scholar
  13. J. Jia, Z. Liu, X. Xiao, B. Liu, and K. C. Chou, “iSuc-PseOpt: identifying lysine succinylation sites in proteins by incorporating sequence-coupling effects into pseudo components and optimizing imbalanced training dataset,” Analytical Biochemistry, vol. 497, pp. 48–56, 2016. View at: Publisher Site | Google Scholar
  14. H. D. Xu, S. P. Shi, P. P. Wen, and J. D. Qiu, “SuccFind: a novel succinylation sites online prediction tool via enhanced characteristic strategy,” Bioinformatics, vol. 31, no. 23, pp. 3748–3750, 2015. View at: Publisher Site | Google Scholar
  15. M. M. Hasan, S. Yang, Y. Zhou, and M. N. Mollah, “SuccinSite: a computational tool for the prediction of protein succinylation sites by exploiting the amino acid patterns and properties,” Molecular BioSystems, vol. 12, no. 3, pp. 786–795, 2016. View at: Publisher Site | Google Scholar
  16. H. Ai, R. Wu, L. Zhang et al., “pSuc-PseRat: predicting lysine succinylation in proteins by exploiting the ratios of sequence coupling and properties,” Journal of Computational Biology, vol. 24, no. 10, pp. 1050–1059, 2017. View at: Publisher Site | Google Scholar
  17. A. Dehzangi, Y. López, S. P. Lal et al., “PSSM-Suc: accurately predicting succinylation using position specific scoring matrix into bigram for feature extraction,” Journal of Theoretical Biology, vol. 425, pp. 97–102, 2017. View at: Publisher Site | Google Scholar
  18. Y. López, A. Dehzangi, S. P. Lal et al., “SucStruct: prediction of succinylated lysine residues by using structural properties of amino acids,” Analytical Biochemistry, vol. 527, pp. 24–32, 2017. View at: Publisher Site | Google Scholar
  19. J. Jia, Z. Liu, X. Xiao, B. Liu, and K. C. Chou, “pSuc-Lys: predict lysine succinylation sites in proteins with PseAAC and ensemble random forest approach,” Journal of Theoretical Biology, vol. 394, pp. 223–230, 2016. View at: Publisher Site | Google Scholar
  20. M. M. Hasan, M. S. Khatun, M. N. H. Mollah, C. Yong, and D. Guo, “A systematic identification of species-specific protein succinylation sites using joint element features information,” International Journal of Nanomedicine, vol. Volume 12, pp. 6303–6315, 2017. View at: Publisher Site | Google Scholar
  21. Q. Ning, X. Zhao, L. Bao, Z. Ma, and X. Zhao, “Detecting succinylation sites from protein sequences using ensemble support vector machine,” BMC Bioinformatics, vol. 19, no. 1, p. 237, 2018. View at: Publisher Site | Google Scholar
  22. A. G. de Brevern, M. M. Hasan, and H. Kurata, “GPSuc: Global Prediction of Generic and Species-specific Succinylation Sites by aggregating multiple sequence features,” PLoS One, vol. 13, no. 10, 2018. View at: Google Scholar
  23. A. Dehzangi, Y. López, S. P. Lal et al., “Improving succinylation prediction accuracy by incorporating the secondary structure via helix, strand and coil, and evolutionary information from profile bigrams,” PLoS One, vol. 13, no. 2, article e0191900, 2018. View at: Publisher Site | Google Scholar
  24. Y. López, A. Sharma, A. Dehzangi et al., “Success: evolutionary and structural properties of amino acids prove effective for succinylation site prediction,” BMC Genomics, vol. 19, no. S1, p. 923, 2018. View at: Publisher Site | Google Scholar
  25. H.-J. Kao, V.-N. Nguyen, K.-Y. Huang, W.-C. Chang, and T.-Y. Lee, “SuccSite: incorporating amino acid composition and informative k-spaced amino acid pairs to identify protein succinylation sites,” Genomics, Proteomics & Bioinformatics, vol. 18, no. 2, pp. 208–219, 2020. View at: Publisher Site | Google Scholar
  26. K.-Y. Huang, J. B.-K. Hsu, and T.-Y. Lee, “Characterization and identification of lysine succinylation sites based on deep learning method,” Scientific Reports, vol. 9, no. 1, article 16175, 2019. View at: Publisher Site | Google Scholar
  27. L. Zhang, M. Liu, X. Qin, G. Liu, and H. Ding, “Succinylation Site Prediction Based on Protein Sequences Using the IFS- LightGBM (BO) Model,” Computational and Mathematical Methods in Medicine, vol. 2020, Article ID 8858489, 15 pages, 2020. View at: Publisher Site | Google Scholar
  28. M. M. Hasan, M. S. Khatun, and H. Kurata, “Large-scale assessment of bioinformatics tools for lysine succinylation sites,” Cell, vol. 8, no. 2, p. 95, 2019. View at: Publisher Site | Google Scholar
  29. N. Thapa, M. Chaudhari, S. McManus et al., “DeepSuccinylSite: a deep learning based approach for protein succinylation site prediction,” BMC Bioinformatics, vol. 21, no. S3, p. 63, 2020. View at: Publisher Site | Google Scholar
  30. Z. Liu, J. Cao, X. Gao et al., “CPLA 1.0: an integrated database of protein lysine acetylation,” Nucleic Acids Research, vol. 39, suppl_1, pp. D1029–D1034, 2011. View at: Publisher Site | Google Scholar
  31. Z. Liu, Y. Wang, T. Gao et al., “CPLM: a database of protein lysine modifications,” Nucleic Acids Research, vol. 42, no. D1, pp. D531–D536, 2014. View at: Publisher Site | Google Scholar
  32. H. Xu, J. Zhou, S. Lin, W. Deng, Y. Zhang, and Y. Xue, “PLMD: an updated data resource of protein lysine modifications,” Journal of Genetics and Genomics, vol. 44, no. 5, pp. 243–250, 2017. View at: Publisher Site | Google Scholar
  33. Y. Huang, B. Niu, Y. Gao, L. Fu, and W. Li, “CD-HIT suite: a web server for clustering and comparing biological sequences,” Bioinformatics, vol. 26, no. 5, pp. 680–682, 2010. View at: Publisher Site | Google Scholar
  34. W. Li and A. Godzik, “Cd-hit: a fast program for clustering and comparing large sets of protein or nucleotide sequences,” Bioinformatics, vol. 22, no. 13, pp. 1658-1659, 2006. View at: Publisher Site | Google Scholar
  35. G. Huang, L. Lu, K. Feng et al., “Prediction of S-nitrosylation modification sites based on kernel sparse representation classification and mRMR algorithm,” BioMed Research International, vol. 2014, Article ID 438341, 10 pages, 2014. View at: Publisher Site | Google Scholar
  36. Q. Xiang, K. Feng, B. Liao, Y. Liu, and G. Huang, “Prediction of lysine malonylation sites based on pseudo amino acid,” Combinatorial Chemistry & High Throughput Screening, vol. 20, no. 7, pp. 622–628, 2017. View at: Publisher Site | Google Scholar
  37. T. Mikolov, I. Sutskever, K. Chen, G. S. Corrado, and J. Dean, “Distributed representations of words and phrases and their compositionality,” Advances in neural information processing systems, Curran Associates Inc., pp. 3111–3119, 2013. View at: Google Scholar
  38. T. Mikolov, K. Chen, G. Corrado, and D. JJapa, “Efficient estimation of word representations in vector space,” 2013, https://arxiv.org/abs/1301.3781. View at: Google Scholar
  39. Y. LeCun, B. Boser, J. S. Denker et al., “Backpropagation applied to handwritten zip code recognition,” Neural Computation, vol. 1, no. 4, pp. 541–551, 1989. View at: Publisher Site | Google Scholar
  40. Y. Lecun, L. Bottou, Y. Bengio, and P. Haffner, “Gradient-based learning applied to document recognition,” Proceedings of the IEEE, vol. 86, no. 11, pp. 2278–2324, 1998. View at: Publisher Site | Google Scholar
  41. A. Krizhevsky, I. Sutskever, and G. E. Hinton, “Imagenet classification with deep convolutional neural networks,” Communications of the ACM, vol. 60, no. 6, pp. 84–90, 2017. View at: Publisher Site | Google Scholar
  42. K. Simonyan and A. Zisserman, “Very deep convolutional networks for large-scale image recognition,” 2014, https://arxiv.org/abs/1409.1556. View at: Google Scholar
  43. C. Szegedy, W. Liu, Y. Jia et al., “Going deeper with convolutions,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1–9, 2015. View at: Google Scholar
  44. K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” pp. 778-779. View at: Google Scholar
  45. B. A. Pearlmutter, “Learning state space trajectories in recurrent neural networks,” Neural Computation, vol. 1, no. 2, pp. 263–269, 1989. View at: Publisher Site | Google Scholar
  46. C. L. Giles, G. M. Kuhn, and R. J. Williams, “Dynamic recurrent neural networks: theory and applications,” IEEE Transactions on Neural Networks, vol. 5, no. 2, pp. 153–156, 1994. View at: Publisher Site | Google Scholar
  47. S. Hochreiter and J. Schmidhuber, “Long short-term memory,” Neural Computation, vol. 9, no. 8, pp. 1735–1780, 1997. View at: Publisher Site | Google Scholar
  48. M. Schuster and K. K. Paliwal, “Bidirectional recurrent neural networks,” IEEE Transactions on Signal Processing, vol. 45, no. 11, pp. 2673–2681, 1997. View at: Publisher Site | Google Scholar
  49. G. E. Hinton, N. Srivastava, A. Krizhevsky, I. Sutskever, and R. R. Salakhutdinov, “Improving neural networks by preventing co-adaptation of feature detectors,” 2012, https://arxiv.org/abs/1207.0580. View at: Google Scholar
  50. N. Srivastava, “Improving neural networks with dropout,” University of Toronto, vol. 182, no. 566, p. 7, 2013. View at: Google Scholar
  51. X. Bouthillier, K. Konda, P. Vincent, and R. Memisevic, “Dropout as data augmentation,” 2015, https://arxiv.org/abs/1506.08700. View at: Google Scholar
  52. P. Baldi and P. Sadowski, “The dropout learning algorithm,” Artificial Intelligence, vol. 210, pp. 78–122, 2014. View at: Publisher Site | Google Scholar
  53. N. Srivastava, G. Hinton, A. Krizhevsky, I. Sutskever, and R. Salakhutdinov, “Dropout: a simple way to prevent neural networks from overfitting,” Journal of Machine Learning Research, vol. 15, no. 1, pp. 1929–1958, 2014. View at: Google Scholar
  54. H. Mi, A. Muruganujan, D. Ebert, X. Huang, and P. D. Thomas, “PANTHER version 14: more genomes, a new PANTHER GO-slim and improvements in enrichment analysis tools,” Nucleic Acids Research, vol. 47, no. D1, pp. D419–D426, 2018. View at: Google Scholar
  55. H. Mi, A. Muruganujan, X. Huang et al., “Protocol update for large-scale genome and gene function analysis with the PANTHER classification system (v.14.0),” Nature Protocols, vol. 14, no. 3, pp. 703–721, 2019. View at: Publisher Site | Google Scholar
  56. D. W. Huang, B. T. Sherman, and R. A. Lempicki, “Bioinformatics enrichment tools: paths toward the comprehensive functional analysis of large gene lists,” Nucleic Acids Research, vol. 37, no. 1, pp. 1–13, 2009. View at: Google Scholar
  57. D. W. Huang, B. T. Sherman, and R. A. Lempicki, “Systematic and integrative analysis of large gene lists using DAVID bioinformatics resources,” Nature Protocols, vol. 4, no. 1, pp. 44–57, 2009. View at: Google Scholar

Copyright © 2021 Guohua Huang et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Related articles

No related content is available yet for this article.
 PDF Download Citation Citation
 Download other formatsMore
 Order printed copiesOrder
Views755
Downloads833
Citations

Related articles

No related content is available yet for this article.

Article of the Year Award: Outstanding research contributions of 2021, as selected by our Chief Editors. Read the winning articles.