Next Article in Journal
Holistic Type Extension for Classical Logic via Toffoli Quantum Gate
Next Article in Special Issue
Sequence Versus Composition: What Prescribes IDP Biophysical Properties?
Previous Article in Journal
Divergence-Based Risk Measures: A Discussion on Sensitivities and Extensions
Previous Article in Special Issue
A Suggestion of Converting Protein Intrinsic Disorder to Structural Entropy Using Shannon’s Information Theory
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Prediction of MoRFs in Protein Sequences with MLPs Based on Sequence Properties and Evolution Information

College of Electronic Information and Optical Engineering, Nankai University, Tianjin 300350, China
*
Author to whom correspondence should be addressed.
Entropy 2019, 21(7), 635; https://doi.org/10.3390/e21070635
Submission received: 19 April 2019 / Revised: 26 June 2019 / Accepted: 26 June 2019 / Published: 27 June 2019

Abstract

:
Molecular recognition features (MoRFs) are one important type of intrinsically disordered proteins functional regions that can undergo a disorder-to-order transition through binding to their interaction partners. Prediction of MoRFs is crucial, as the functions of MoRFs are associated with many diseases and can therefore become the potential drug targets. In this paper, a method of predicting MoRFs is developed based on the sequence properties and evolutionary information. To this end, we design two distinct multi-layer perceptron (MLP) neural networks and present a procedure to train them. We develop a preprocessing process which exploits different sizes of sliding windows to capture various properties related to MoRFs. We then use the Bayes rule together with the outputs of two trained MLP neural networks to predict MoRFs. In comparison to several state-of-the-art methods, the simulation results show that our method is competitive.

1. Introduction

Intrinsically disordered proteins (IDPs) possess flexible and instable structures which make them play a crucial role in a variety of important biological functions [1]. Being an important type of functional region in IDPs, molecular recognition features (MoRFs), generally consisting of 10–70 consecutive residues and are located in the long disordered regions, can undergo a disorder-to-order transition through binding to their interaction partners [2,3]. There are four types of MoRFs, i.e., α-MoRFs, β-MoRFs, γ-MoRFs and complex-MoRFs, which correspond to α-helix, β-strand, coil secondary structures and multiple secondary structures [4]. Many MoRFs, acting as molecular switches in molecular-interaction networks, play a role in the signaling and alternative splicing of cells [2]. It is observable that MoRFs are abundant in proteins with recognition functions [5]. Prediction of MoRFs is crucial, as the functions of MoRFs are associated with many diseases and may therefore be potential drug targets [6].
In recent years, many computational schemes for predicting MoRFs have been reported, including α-MoRF-Pred I [5], α-MoRF-Pred II [7], ANCHOR [8], MoRFpred [9], MSPSSMpred [10], DISOPRED3 [11] and MoRFCHiBi [12]. Of these reported methods, α-MoRF-Pred I and α-MoRF-Pred II are only capable of predicting α-MoRFs. ANCHOR, using estimated energy calculations [13] to capture the biophysical properties of MoRFs, cannot yield enough prediction accuracy in many cases. MoRFpred is a comprehensive method relying on a lot of features, such as evolutionary information [14] and physiochemical properties [15], solvent accessibility [16], and B-factors [17], as well as predicted disorder probabilities from several other predictors [18,19,20,21]. MoRFCHiBi predicts MoRFs by extracting features from physicochemical properties [15] and utilizing two support vector machine (SVM) models [12]. MoRFCHiBi does not depend on the results of other predictors, and obtains better prediction performance than MoRFpred. However, the prediction accuracy of MoRFCHiBi is still expected to improve. Besides, MoRFCHiBi_Web [22] and MoRFCHiBi_Light [23] are two methods based on combining MoRFCHiBi with other predictors, and obtain better performance than MoRFCHiBi. OPAL [24] is also a combined predictor, which utilizes the results of two independent predictors. The first one is MoRFCHiBi, and the second one is PROMIS [24], obtained by half-sphere exposure, solvent accessible surface area and backbone angle information. These combined predictors usually perform better than individual predictors. However, they all need to combine MoRFCHiBi with other predictors.
In this paper, we design an individual prediction method for MoRFs based on two distinct multi-layer perceptron (MLP) neural networks where one of them is MLP1 trained by the 16 sequence properties and the other is MLP2 with the evolutionary information. We present a procedure to train these two MLP neural networks. The training procedure utilizes the data from the preprocessing process developed by us, which involves different sizes of the sliding windows to capture various properties related to MoRFs. The outputs of MLP1 and MLP2 are then used to predict MoRFs based on the Bayes rule. Finally, the TEST464 and TEST_EXP53 sets are used to compare the performances of our method with ANCHOR, MoRFpred, MoRFCHiBi, MoRFCHiBi_Web, MoRFCHiBi_Light, PROMIS and OPAL. The simulation results show that the AUCs generated from our method are equal to 0.806 and 0.792 on the TEST464 and TEST_EXP53 sets, respectively.

2. Materials and Methods

In this section, we employ the sequence properties as well as the evolutionary information of the protein sequence to train two distinct MLP neural networks respectively. Utilizing the probability distributions yielded from these two distinct MLP neural networks, we then follow the Bayes rule to predict MoRFs.

2.1. Datasets

For comparison with other methods, we use the same datasets created by Disfani et al. [9], which is from Protein Data Bank (PDB) [25]. The datasets include 840 MoRFs, which contain 181 helical, 34 strand, 595 coil and 30 complex MoRF regions. In [9], the 840 MoRFs dataset are divided into the TRAINING and TEST set with which there are 421 and 419 protein sequences respectively. Thus, the TRAINING set contains 245,984 residues including 5396 MoRF residues, while the TEST set has 258,829 residues with 5153 MoRF residues. After that, Disfani et al. also used another test set, TESTNEW. TESTNEW has 45 sequences consisting of 37,533 residues with 626 MoRF residues. We combine the TEST and TESTNEW sets into single set TEST464. In addition, we use the TEST_EXP53 assembled by Malhis et al. [22] as the third test set. TEST_EXP53 has 53 sequences, including 2432 MoRF residues, which consist of 729 short MoRF residues (up to 30 residues) and 1703 long MoRF residues (more than 30 residues). We use the same TRAINING set to train our two distinct MLP neural networks and TEST set to evaluate it. The TEST464 and TEST_EXP53 sets are utilized to compare our method with other methods.

2.2. Feature Selection

We rely on the sequence properties and evolutionary information of the protein sequences to predict MoRFs. Protein sequences with MoRFs generally comprise of MoRFs, MoRFs’ flanking (Flanks) and other non-MoRF residues. We define the Flanks as other residues in the disordered regions where MoRFs are located. Our computation of the sequence properties and evolutionary information of the protein sequences does not require to do any special treatment on Flanks relies only on the protein sequences.
The sequence properties we use include 13 physicochemical properties from the Amino Acid Index [15] listed in Table A1 of Appendix A and 3 structural properties, which include the topological entropy [26], and the Remark 465 and Deleage/Roux propensities, both from the GlobPlot NAR paper [27]. Evolutionary information in this paper is obtained by the Position Specific Scoring Matrix (PSSM) through executing three iterations of PSI-BLAST against NCBI [14] non-redundant database with default parameters.
The 16 sequence properties we select are preprocessed as follows: Given a protein sequence of length L , we select a sliding window of size   N ( N < L ) and append ⎣ N / 2 ⎦ zeros to both ends of the sequence. In each region determined by the window, we compute the topological entropy through Equation 14 of [26]. Each of the rest 15 sequence properties is assigned to the average value of the amino acid scales of the residues in this region. Thus, each window can obtain a 16-dimensional vector v i ( 1 i L ) . Then, we associate it with every residue in the window. Finally, for each residue, we assign the average value of all the v i associated with it as the feature vector of this specific residue. The feature vector x j ( 1 j L ) can be computed as
x j = { 1 j + N 0 i = 1 j + N 0 v i , 1 j N 0 1 N i = j + N 0 N + 1 j + N 0 v i , N 0 < j L N 0 1 L 0 j N 0 + 1 i = j + N 0 N + 1 L 0 N + 1 v i , L N 0 < j L ,
where N 0 = N / 2 and L 0 = L + 2 N 0 . In this paper, we use three sizes of the sliding windows, i.e., N = 10, 45 and 90, to compute 16 sequence properties. The smaller size of the sliding window is used to capture properties especially related to MoRFs, as they are usually much shorter than the surrounding disordered regions. The longer sliding windows are used to extract information from the surrounding regions of MoRFs. Thus, using these three sizes of sliding windows, we can compute a 48-dimensional feature vector associated with each residue of the protein sequence.
For this given protein sequence of length L , the evolutionary information is computed through the PSSM which yields a 20 × L   matrix [28]. We first transform this 20 × L   matrix into a 20 × L 0 matrix by appending 20 × N / 2 zeros at the beginning and end of this 20 × L   matrix, respectively. Then we again choose three sizes of the sliding windows, i.e., N = 10, 45 and 90, to slice a 20 × N matrix from the 20 × L 0 transformed matrix. We can compute an average value for each row for this 20 × N   matrix, and then use Equation (1) to yield a 20 × 1 vector for each residue. Since three sizes of the sliding windows are employed, a 60-dimensional feature vector containing the evolutionary information is obtained for each residue.

2.3. MLP Prediction Models

We train two distinct multi-layer perceptron (MLP) neural networks, where one of them is trained using the sequence properties and the other is trained through the evolutionary information obtained from the above section. Both these models contain two hidden layers, with each hidden layer having 12 perceptrons and one bias. We use the ReLU functions as the activation functions in each hidden layer and the sigmoid functions in the output layers. During the training process, a dropout algorithm [29] is utilized to avoid overfitting. The forward propagation with dropout algorithm proceeds as follows:
Z [ l ] = W [ l ] · A [ l 1 ] + b [ l ] ,
A [ l ] = g [ l ] ( Z [ l ] ) R ( p d )   ,
where g [ l ] denotes the vector activation function of the l -th layer, and l = 1 , 2 , 3   in our model. R ( p d ) is a vector obeying the Bernoulli distribution with p d being the dropout parameter which represents the remaining probability of each perceptron in the hidden layers. Furthermore, A [ 0 ] is the input feature matrix and A [ 3 ] is the prediction result. Then, we employ the Adam algorithm [30] to optimize W [ l ] and b [ l ] in the back propagation.
From Section 2.1, the TRAINING set contains 245,984 residues, among which there are 5396 MoRF residues. We only randomly select 5396 non-MoRF residues from the TRAINING set to train our two MLP neural networks, which ensures our trained MLP neural networks to being capable of more effectively identifying both MoRF and non-MoRF residues. Finally, in order to increase the robust of our MLP neural networks and reduce the influence of the initial weights, we train the two MLP neural networks five times and use the average values of them as the final outputs of our MLP neural networks. Finally, we use the Bayes rule together with the outputs of our two MLP neural networks to compute the prediction of MoRFs. The detailed paradigm of our method is shown in Figure 1.

2.4. Performance Evaluation

We use the ROC curve and three evaluation metrics to evaluate performance in this paper. These are the AUC (the area under the ROC curve), TPR (the true positive rate) and FPR (the false positive rate). The computation equations of TPR and FPR are TPR = T P / N MoRF   ,   FPR = T N / N non , where T P and T N respectively represent the numbers of accurately predicted MoRFs and non-MoRFs residues. In addition, we denote the total number of MoRFs and non-MoRFs residues as NMoRF and Nnon, respectively.

3. Results and Discussion

Using the TEST set defined in Section 2.1, we run our trained multi-layer perceptron (MLP) neural networks. The outputs of two trained MLP neural networks MLP1 and MLP2 are then utilized to predict MoRFs based on the Bayes rule. Finally, the TEST464 and TEST_EXP53 sets are used to compare the performances of our method with ANCHOR, MoRFpred, MoRFCHiBi, MoRFCHiBi_Web, MoRFCHiBi_Light, PROMIS and OPAL. The simulation results show that the AUC values generated from our method are equal to 0.806 and 0.792 on the TEST464 and TEST_EXP53 sets, respectively.

3.1. Prediction Performance of Sequence Properties

For the 16 sequence properties, three windows with lengths of 10, 45 and 90 are used to perform preprocessing and calculate the feature matrix. Then we train the first MLP neural network MLP1 as shown in Figure 1. In MLP1, the perceptron number of two hidden layers is set to Nneur = [12,12], where the two numbers correspond to the perceptron numbers of the first hidden layer and the second hidden layer. The dropout parameter is pd = 0.5, and the learning rate is 0.001. Figure 2 shows the ROC curves of five independent MLPs of MLP1 on the TEST set, and the ROC curves of the average values which are described by the red curves. Figure 2a shows the overall ROC curves, and Figure 2b shows the ROC curves in the low FPR region.
From Figure 2a, the red curve is higher than the other curves. Furthermore, in Figure 2b, although the pink curve is the highest, the red one is very close to it, and the pink curve is obviously lower than the red one when FPR > 0.2 . Similar to the pink curve, other curves may be slightly higher than the red curve in some small regions, but may be lower in other regions. Therefore, the prediction performance can be improved by training five independent networks and taking their average values of outputs as the final outputs of the MLP1.

3.2. Impact of Different MLP1 Parameters

In this section, we change the perceptron number Nneur and the dropout parameter pd in MLP1 to analysis their influence. Firstly, we change the perceptron number, and set the dropout parameter and the learning rate to 0.5 and 0.001, respectively. Figure 3 shows the ROC curves of MLP1 on the TEST set when Nneur = [12,12], [25,25], [50,50].
The curves in Figure 3 are the prediction results of MLP1 calculated by the average values of five independent networks. The red curves are higher than other curves in Figure 3a,b. Thus, the perceptron number of MLP1 is set to Nneur = [12,12].
After determining the perceptron number, we change the dropout parameter pd in MLP1. Figure 4 shows the ROC curves on the TEST set when pd = 0.5, 0.7, 1. Although these three curves are approximate to each other in Figure 4a,b, the red curve in Figure 4a is slightly higher than the others, and the red and blue curves in Figure 4b are slightly higher than the pink curve. Finally, the dropout parameter of MLP1 is set to pd = 0.5.

3.3. Prediction Performance of Evolutionary Information

For the evolutionary information, after calculating the PSSM for each protein sequence, three windows of 10, 45 and 90 are also used to perform preprocessing and calculate the feature matrix. Then we train the second MLP neural network MLP2 as shown in Figure 1. We first set the perceptron number to Nneur = [25,25] in MLP2. The dropout parameter is pd = 0.7, and the learning rate is 0.0001. Similar to Figure 2, Figure 5 shows the ROC curves of five independent MLPs of MLP2 on the TEST set, and the ROC curves of the average values, which are also the red curves.
From Figure 5, the red curves are higher than other curves both in (a) and (b). Thus, for the evolutionary information, the prediction performance is also improved by training five independent networks and taking their average values of outputs as the final outputs of MLP2.

3.4. Impact of Different MLP2 Parameters

We change the perceptron number Nneur and the dropout parameter pd in MLP2 to analyze their influence. Firstly, we only change Nneur; simultaneously, the dropout parameter is set to pd = 0.7 and the learning rate is also 0.0001. Figure 6 shows the ROC curves of MLP2 on the TEST set when Nneur = [12,12], [25,25], [50,50]. In Figure 6a, the blue curve is higher than the other curves, and it is also higher when FPR > 0.06 in Figure 6b. Finally, we set Nneur = [25,25] in MLP2.
After determining the perceptron number in MLP2, we change the dropout parameter pd. Figure 7 shows the ROC curves on the TEST set when pd = 0.5, 0.7, 1. In Figure 7a, the blue and pink curves are better than the red curve. However, the blue curve is better than the other curves in Figure 7b. Therefore, the dropout parameter of MLP2 is set to pd = 0.7.

3.5. Prediction Performance of the Fusion Results

In this section, the outputs of MLP1 and MLP2 are fused using the Bayes rule to predict MoRFs. Figure 8 shows the ROC curves of MLP1, MLP2 and the fusion results on the TEST set. The ROC curve of MLP2 gets a better performance in the low FPR region, while the ROC curve of MLP1 performs better when the FPR is higher than 0.2. Thus, the ROC curves of MLP1 and MLP2 cross each other in Figure 8a. However, the curve of the fusion results is higher than the other curves both in low FPR and high FPR regions.

3.6. Comparison with other Methods

In this section, using the TEST464 and TEST_EXP53 sets, we compare our method, named MoRFMLP, with ANCHOR, MoRFpred, MoRFCHiBi, MoRFCHiBi_Web, MoRFCHiBi_Light, PROMIS and OPAL. The results of other methods are from [24] and the online predictor of MoRFCHiBi system. Table 1 shows the AUC values of these methods that run on these two test sets. Since TEST_EXP53 contains long MoRF regions, we not only compare the AUC values on the overall dataset (EXP53_all), but also compare the AUC values on the datasets that only contain long MoRF regions (EXP53_long) and short MoRF regions (EXP53_short), respectively. In these methods, MoRFCHiBi_Web, MoRFCHiBi_Light and OPAL are combined component predictors. They usually perform better than individual predictors. The bold data in Table 1 indicate the best values in individual and combined component predictors, respectively. Our method is an individual predictor, so we mainly compare with ANCHOR, MoRFpred, MoRFCHiBi and PROMIS, which are also individual predictors. From Table 1, MoRFMLP obtains a higher AUC on the TEST464 set, and PROMIS gets a higher AUC on the TEST_EXP53 set.
In addition, to further analyze the prediction performance of these methods, we also calculate the FPR values at different TPR on TEST464 and EXP53_all sets, as shown in Table 2. From Table 2, MoRFMLP gets the lowest FPR when TPR is set to 0.2, 0.3 and 0.4 in five individual predictors, which indicates that MoRFMLP can obtain higher TPR at low FPR. Therefore, as an individual predictor, MoRFMLP is competitive.

4. Conclusions

In this paper, we propose a new method, MoRFMLP, to predict MoRFs. We employ the sequence properties as well as the evolutionary information to train two distinct MLP neural networks. The sequence properties contain 13 physicochemical properties and 3 structural properties, and are extracted by preprocessing using 3 different windows. The evolutionary information is extracted from PSSM and preprocessed by the same windows as sequence properties. Then, the outputs of the two MLP neural networks are utilized to predict MoRFs based on Bayes rule. Finally, we test MoRFMLP using TEST464 and TEST_EXP53 sets. Compared to other individual predictors, the simulation results show that MoRFMLP achieves higher AUC on TEST464 set, and gets higher TPR at low FPR on TEST464 and EXP53_all sets.

Author Contributions

Conceptualization, H.H. and J.Z.; methodology, H.H. and J.Z.; software, H.H.; validation, H.H. and J.Z.; formal analysis, H.H.; investigation, H.H.; resources, J.Z.; data curation, H.H. and J.Z.; writing—original draft preparation, H.H.; writing—review and editing, J.Z. and G.S.; visualization, H.H.; supervision, J.Z. and G.S.; project administration, J.Z.; funding acquisition, G.S.

Funding

This research was funded by the National Science Foundation of China, grant number 61771262.

Acknowledgments

We would like to thank Disfani et al. and Malhis et al. for publicity providing the datasets for the MoRFs prediction. We would also like to thank the authors of MoRFCHiBi and OPAL for their detailed prediction results, which provide us with great convenience for comparison with other methods.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

The 13 physicochemical properties selected from the Amino Acid Index are given in the following list.
Table A1. The properties selected from the Amino Acid Index.
Table A1. The properties selected from the Amino Acid Index.
Index:Definition:
CIDH920101Normalized hydrophobicity scales for alpha-proteins
EISD860103Direction of hydrophobic moment
NISK86010114 A contact number
QIAN880105Weights for alpha-helix at the window position of -2
ROBB760101Information measure for alpha-helix
ROBB760108Information measure for turn
ROBB760112Information measure for coil
ROBB760113Information measure for loop
CORJ870103PRIFT index
CORJ870106ALTLS index
CORJ870107TOTFT index
CORJ870108TOTLS index
MIYS990104Optimized relative partition energies—method C

References

  1. Uversky, V.N. Functional roles of transiently and intrinsically disordered regions within proteins. FEBS J. 2015, 282, 1182–1189. [Google Scholar] [CrossRef] [PubMed]
  2. Mohan, A.; Oldfield, C.J.; Radivojac, P.; Vacic, V.; Cortese, M.S.; Dunker, A.K.; Uversky, V.N. Analysis of molecular recognition features (MoRFs). J. Mol. Biol. 2006, 362, 1043–1059. [Google Scholar] [CrossRef] [PubMed]
  3. Cumberworth, A.; Lamour, G.; Babu, M.M.; Gsponer, J. Promiscuity as a functional trait: Intrinsically disordered regions as central players of interactomes. Biochem. J. 2013, 454, 361–369. [Google Scholar] [CrossRef] [PubMed]
  4. Vacic, V.; Oldfield, C.J.; Mohan, A.; Radivojac, P.; Cortese, M.S.; Uversky, V.N.; Dunker, A.K. Characterization of molecular recognition features, MoRFs, and their binding partners. J. Proteome Res. 2007, 6, 2351–2366. [Google Scholar] [CrossRef] [PubMed]
  5. Oldfield, C.J.; Cheng, Y.G.; Cortese, M.S.; Romero, P.; Uversky, V.N.; Dunker, A.K. Coupled folding and binding with alpha-helix-forming molecular recognition elements. Biochemistry 2005, 44, 12454–12470. [Google Scholar] [CrossRef] [PubMed]
  6. Babu, M.M.; van der Lee, R.; de Groot, N.S.; Gsponer, J. Intrinsically disordered proteins: Regulation and disease. Curr. Opin. Struc. Biol. 2011, 21, 432–440. [Google Scholar] [CrossRef] [PubMed]
  7. Cheng, Y.G.; Oldfield, C.J.; Meng, J.W.; Romero, P.; Uversky, V.N.; Dunker, A.K. Mining alpha-helix-forming molecular recognition features with cross species sequence alignments. Biochemistry 2007, 46, 13468–13477. [Google Scholar] [CrossRef] [PubMed]
  8. Dosztanyi, Z.; Meszaros, B.; Simon, I. ANCHOR: Web server for predicting protein binding regions in disordered proteins. Bioinformatics 2009, 25, 2745–2746. [Google Scholar] [CrossRef]
  9. Disfani, F.M.; Hsu, W.L.; Mizianty, M.J.; Oldfield, C.J.; Xue, B.; Dunker, A.K.; Uversky, V.N.; Kurgan, L. MoRFpred, a computational tool for sequence-based prediction and characterization of short disorder-to-order transitioning binding regions in proteins. Bioinformatics 2012, 28, I75–I83. [Google Scholar] [CrossRef]
  10. Fang, C.; Noguchi, T.; Tominaga, D.; Yamana, H. MFSPSSMpred: Identifying short disorder-to-order binding regions in disordered proteins based on contextual local evolutionary conservation. BMC Bioinform. 2013, 14. [Google Scholar] [CrossRef]
  11. Jones, D.T.; Cozzetto, D. DISOPRED3: Precise disordered region predictions with annotated protein-binding activity. Bioinformatics 2015, 31, 857–863. [Google Scholar] [CrossRef] [PubMed]
  12. Malhis, N.; Gsponer, J. Computational identification of MoRFs in protein sequences. Bioinformatics 2015, 31, 1738–1744. [Google Scholar] [CrossRef] [Green Version]
  13. Meszaros, B.; Simon, I.; Dosztanyi, Z. Prediction of Protein Binding Regions in Disordered Proteins. PLoS Comput. Biol. 2009, 5. [Google Scholar] [CrossRef] [PubMed]
  14. Altschul, S.; Madden, T.; Schaffer, A.; Zhang, J.H.; Zhang, Z.; Miller, W.; Lipman, D. Gapped BLAST and PSI-BLAST: A new generation of protein database search programs. FASEB J. 1998, 12, A1326. [Google Scholar] [CrossRef]
  15. Kawashima, S.; Pokarowski, P.; Pokarowska, M.; Kolinski, A.; Katayama, T.; Kanehisa, M. AAindex: Amino acid index database, progress report 2008. Nucleic. Acids Res. 2008, 36, D202–D205. [Google Scholar] [CrossRef]
  16. Faraggi, E.; Xue, B.; Zhou, Y.Q. Improving the prediction accuracy of residue solvent accessibility and real-value backbone torsion angles of proteins by guided-learning through a two-layer neural network. Proteins 2009, 74, 847–856. [Google Scholar] [CrossRef]
  17. Schlessinger, A.; Yachdav, G.; Rost, B. PROFbval: Predict flexible and rigid residues in proteins. Bioinformatics 2006, 22, 891–893. [Google Scholar] [CrossRef] [PubMed]
  18. Dosztanyi, Z.; Csizmok, V.; Tompa, P.; Simon, I. IUPred: Web server for the prediction of intrinsically unstructured regions of proteins based on estimated energy content. Bioinformatics 2005, 21, 3433–3434. [Google Scholar] [CrossRef]
  19. Ward, J.J.; McGuffin, L.J.; Bryson, K.; Buxton, B.F.; Jones, D.T. The DISOPRED server for the prediction of protein disorder. Bioinformatics 2004, 20, 2138–2139. [Google Scholar] [CrossRef] [PubMed]
  20. McGuffin, L.J. Intrinsic disorder prediction from the analysis of multiple protein fold recognition models. Bioinformatics 2008, 24, 1798–1804. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  21. Mizianty, M.J.; Stach, W.; Chen, K.; Kedarisetti, K.D.; Disfani, F.M.; Kurgan, L. Improved sequence-based prediction of disordered regions with multilayer fusion of multiple information sources. Bioinformatics 2010, 26, i489–i496. [Google Scholar] [CrossRef] [Green Version]
  22. Malhis, N.; Wong, E.T.C.; Nassar, R.; Gsponer, J. Computational identification of MoRFs in protein sequences using hierarchical application of bayers rule. PLoS ONE 2015, 10, e0141603. [Google Scholar] [CrossRef]
  23. Malhis, N.; Jacobson, M.; Gsponer, J. MoRFchibi system: Software tools for the identification of MoRFs in protein sequences. Nucleic Acids Res. 2016, 44, 488–493. [Google Scholar] [CrossRef]
  24. Sharma, R.; Raicar, G.; Tsunoda, T.; Patil, A.; Sharma, A. OPAL: Prediction of MoRF regions in intrinsically disordered protein sequences. Bioinformatics 2018, 34, 1850–1858. [Google Scholar] [CrossRef]
  25. Berman, H.; Henrick, K.; Nakamura, H.; Markley, J.L. The worldwide Protein Data Bank (wwPDB): ensuring a single, uniform archive of PDB data. Nucleic. Acids Res. 2007, 35, D301–303. [Google Scholar] [CrossRef]
  26. He, H.; Zhao, J.X. A Low Computational Complexity Scheme for the Prediction of Intrinsically Disordered Protein Regions. Math. Probl. Eng. 2018. [Google Scholar] [CrossRef]
  27. Linding, R.; Russell, R.B.; Neduva, V.; Gibson, T.J. GlobPlot: Exploring protein sequences for globularity and disorder. Nucleic. Acids Res. 2003, 31, 3701–3708. [Google Scholar] [CrossRef] [PubMed]
  28. He, H.; Zhao, J.X.; Sun, G.L. The Prediction of Intrinsically Disordered Proteins Based on Feature Selection. Algorithms 2019, 12, 46. [Google Scholar] [CrossRef]
  29. Srivastava, N.; Hinton, G.; Krizhevsky, A.; Sutskever, I.; Salakhutdinov, R. Dropout: A Simple Way to Prevent Neural Networks overfitting. J. Mach. Learn. Res. 2014, 15, 1929–1958. [Google Scholar]
  30. Kingma, D.P.; Ba, J.L. Adam: A Method for Stochastic Optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
Figure 1. Detailed paradigm of the proposed method.
Figure 1. Detailed paradigm of the proposed method.
Entropy 21 00635 g001
Figure 2. ROC curves of MLP1 on the TEST set. (a) The overall ROC curves. (b) The ROC curves at low FPR region.
Figure 2. ROC curves of MLP1 on the TEST set. (a) The overall ROC curves. (b) The ROC curves at low FPR region.
Entropy 21 00635 g002
Figure 3. ROC curves of MLP1 with different Nneur. (a) The overall ROC curves; (b) the ROC curves in the low FPR region.
Figure 3. ROC curves of MLP1 with different Nneur. (a) The overall ROC curves; (b) the ROC curves in the low FPR region.
Entropy 21 00635 g003
Figure 4. ROC curves of MLP1 with different pd. (a) The overall ROC curves. (b) The ROC curves in the low FPR region.
Figure 4. ROC curves of MLP1 with different pd. (a) The overall ROC curves. (b) The ROC curves in the low FPR region.
Entropy 21 00635 g004
Figure 5. ROC curves of MLP2 on the TEST set. (a) The overall ROC curves; (b) the ROC curves at low FPR region.
Figure 5. ROC curves of MLP2 on the TEST set. (a) The overall ROC curves; (b) the ROC curves at low FPR region.
Entropy 21 00635 g005
Figure 6. ROC curves of MLP2 with different Nneur. (a) The overall ROC curves. (b) The ROC curves in the low FPR region.
Figure 6. ROC curves of MLP2 with different Nneur. (a) The overall ROC curves. (b) The ROC curves in the low FPR region.
Entropy 21 00635 g006
Figure 7. ROC curves of MLP2 with different pd. (a) The overall ROC curves. (b) The ROC curves in the low FPR region.
Figure 7. ROC curves of MLP2 with different pd. (a) The overall ROC curves. (b) The ROC curves in the low FPR region.
Entropy 21 00635 g007
Figure 8. ROC curves of MLP1, MLP2 and the fusion results. (a) The overall ROC curves. (b) The ROC curves at low FPR region.
Figure 8. ROC curves of MLP1, MLP2 and the fusion results. (a) The overall ROC curves. (b) The ROC curves at low FPR region.
Entropy 21 00635 g008
Table 1. AUC on TEST464 and TEST_EXP53.
Table 1. AUC on TEST464 and TEST_EXP53.
TEST464EXP53_AllEXP53_ShortEXP53_Long
MoRFMLP 0.8060.7920.8180.781
ANCHOR0.6050.6150.6830.586
MoRFpred0.6750.6200.6730.598
PROMIS0.7900.8180.8230.815
MoRFCHiBi0.7430.7120.7900.679
MoRFCHiBi_Light,0.7770.7990.8690.770
MoRFCHiBi_Web0.8050.7970.8860.758
OPAL0.8160.8360.8700.823
Table 2. FPR at different TPR on TEST464 and EXP53_all.
Table 2. FPR at different TPR on TEST464 and EXP53_all.
TPR = 0.2 TPR = 0.3 TPR = 0.4
TEST464EXP53_AllTEST464EXP53_AllTEST464EXP53_All
MoRFMLP 0.0150.0300.0290.0510.0510.079
ANCHOR0.0790.1040.1630.1730.2460.263
MoRFpred0.0330.0830.0710.1460.1430.221
PROMIS0.0310.0320.0690.0560.1030.081
MoRFCHiBi0.0310.0310.0630.0640.1040.125
MoRFCHiBi_Light,0.0200.0160.0400.0430.0730.068
MoRFCHiBi_Web0.0160.0160.0330.0330.0570.061
OPAL0.0250.0150.0520.0290.0740.056

Share and Cite

MDPI and ACS Style

He, H.; Zhao, J.; Sun, G. Prediction of MoRFs in Protein Sequences with MLPs Based on Sequence Properties and Evolution Information. Entropy 2019, 21, 635. https://doi.org/10.3390/e21070635

AMA Style

He H, Zhao J, Sun G. Prediction of MoRFs in Protein Sequences with MLPs Based on Sequence Properties and Evolution Information. Entropy. 2019; 21(7):635. https://doi.org/10.3390/e21070635

Chicago/Turabian Style

He, Hao, Jiaxiang Zhao, and Guiling Sun. 2019. "Prediction of MoRFs in Protein Sequences with MLPs Based on Sequence Properties and Evolution Information" Entropy 21, no. 7: 635. https://doi.org/10.3390/e21070635

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop