Next Article in Journal
Carbazole Alkaloids from Clausena anisum-olens: Isolation, Characterization, and Anti-HIV Evaluation
Previous Article in Journal
QSAR Model of Indeno[1,2-b]indole Derivatives and Identification of N-isopentyl-2-methyl-4,9-dioxo-4,9-Dihydronaphtho[2,3-b]furan-3-carboxamide as a Potent CK2 Inhibitor
Previous Article in Special Issue
Prediction of Protein Subcellular Localization Based on Fusion of Multi-view Features
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

PredPSD: A Gradient Tree Boosting Approach for Single-Stranded and Double-Stranded DNA Binding Protein Prediction

1
School of Computer Science and Engineering, Central South University, Changsha 410075, China
2
School of Software, Xinjiang University, Urumqi 830008, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Molecules 2020, 25(1), 98; https://doi.org/10.3390/molecules25010098
Submission received: 7 December 2019 / Revised: 20 December 2019 / Accepted: 21 December 2019 / Published: 26 December 2019
(This article belongs to the Special Issue Computational Analysis for Protein/Gene Structure and Interaction)

Abstract

:
Interactions between proteins and DNAs play essential roles in many biological processes. DNA binding proteins can be classified into two categories. Double-stranded DNA-binding proteins (DSBs) bind to double-stranded DNA and are involved in a series of cell functions such as gene expression and regulation. Single-stranded DNA-binding proteins (SSBs) are necessary for DNA replication, recombination, and repair and are responsible for binding to the single-stranded DNA. Therefore, the effective classification of DNA-binding proteins is helpful for functional annotations of proteins. In this work, we propose PredPSD, a computational method based on sequence information that accurately predicts SSBs and DSBs. It introduces three novel feature extraction algorithms. In particular, we use the autocross-covariance (ACC) transformation to transform feature matrices into fixed-length vectors. Then, we put the optimal feature subset obtained by the minimal-redundancy-maximal-relevance criterion (mRMR) feature selection algorithm into the gradient tree boosting (GTB). In 10-fold cross-validation based on a benchmark dataset, PredPSD achieves promising performances with an AUC score of 0.956 and an accuracy of 0.912, which are better than those of existing methods. Moreover, our method has significantly improved the prediction accuracy in independent testing. The experimental results show that PredPSD can significantly recognize the binding specificity and differentiate DSBs and SSBs.

1. Introduction

Protein–DNA interaction is a crucial prerequisite for cell function, such as gene replication, transcription, and protein expression translation [1,2,3,4]. DNA can be categorized into single-stranded DNA (ssDNA) and double-stranded DNA (dsDNA). Accordingly, double-stranded DNA-binding proteins (DSBs) specifically bind with dsDNA, while single-stranded DNA-binding proteins (SSBs) specifically bind with ssDNA [5,6].
Knowledge about DNA-binding residues and binding specificity are important references for rational drug design [7,8,9,10]. The availability of binding specificity encourages researchers to focus on analyzing the specific binding sites of DSBs [11,12,13,14,15], the classification prediction of DNA-binding proteins [16,17,18], the function prediction of DNA-binding proteins [19,20,21,22] and the specificity of a protein to DNA binding [23,24], etc. However, the few existing methods for large-scale identification of DSBs and SSBs need further improvement. There are three main classification methods: (1) experimental techniques [25,26], (2) structure-based methods [27], and (3) sequence-based methods [28,29,30]. However, there are few large-scale identification methods for DSBs and SSBs.
In many earlier studies, biological functions have been mainly studied by X-ray crystallography, NMR, and filter binding assays [31,32,33]. However, the use of these experimental techniques for identifying DSBs and SSBs require expensive experimental setups and massive human resource allocations and are time-consuming. Hence, the development of computational methods has been emphasized by several investigators in the field. Initially, Wang et al. [27] proposed a support vector machine (SVM) method (Wang, 2014) with structure-based features related to surface clefts and OB-folds [30,34] as the input features. The results showed that this method achieved an accuracy of 0.8251 and an MCC of 0.6632. Because the gap between available sequences and structures of DNA binding proteins in UniProtKB/Swiss-Prot (www.uniprot.org) and the PDB (www.rcsb.org/pdb/) has been growing exponentially, structure-based methods can no longer meet the needs of high-throughput research [35,36]. Subsequently, Wei Wang et al. [37] developed a machine learning method (Wang, 2017) with only single sequence information such as overall amino acid composition (OAAC) features, dipeptide compositions, and position-specific scoring matrix profiles (PSSMs). The results showed an accuracy of 88.7% and an AUC (area under the curve) of 0.919 on the benchmark datasets.
Although these computational prediction methods have been gradually developed, there are still some problems that make DSB and SSB classification prediction a very challenging task. On the one hand, the performance of commonly used feature extraction methods is still unsatisfactory, and the sequence information cannot be fully utilized to extract more effective features. On the other hand, novel feature selection algorithms and high-performance ensemble learning algorithms, such as gradient tree boosting (GTB), are rarely used in this field.
In this work, we have developed a novel approach, PredPSD, for classifying DSBs and SSBs through a more complete combination of sequence features, such as local structural entropy (LSE), NetSurfP, and DisEMBL. The results show an accuracy of 91.2% and an AUC (area under the curve) of 0.956 on benchmark datasets and indicate that the GTB algorithm and novel feature combinations are essential determinants in the classification of DSBs and SSBs. Furthermore, our algorithm achieves a significantly improved overall performance on an independent dataset. The workflow of our method is shown in Figure 1.

2. Results

2.1. Analysis of mRMR Results

We calculated a total of 1510 sequence features for each protein, including local structural entropy (LSE), NetSurfP, DisEMBL, overall amino acid composition (OAAC), dipeptide composition, PSSM profiles, and physicochemical properties. The maximum correlation between feature and category is computed. Different subsets of characteristics are obtained by setting different thresholds. The number of features changes with the threshold value, as shown in Figure 2.
Through the verification of the machine learning algorithm in terms of time complexity and accuracy, we finally set the threshold as 0.005. Therefore, we obtained an optimal set of 207 features. These selected features are shown in Supplementary Table S1. We also compared the experimental performance before and after using the feature selection algorithm. Specific experimental performance comparison results are shown in Supplementary Table S2 and S3.
To evaluate the mRMR method, we calculated the ratio of the number of selected features to the number of candidate features. A pie chart of the selected feature extraction methods and the corresponding ratios is illustrated in Figure 3, which shows that the total probability of the selected features obtained by NetSurfP, LSE, and DisEMBL is 43%. As expected, the introduction of sequence-based features has a strong correlation between the classification of two different types of proteins.
Through the analysis of 207 optimal features, we found that most of the top 10 selected features are obtained by NetSurfP and PSSM. Among the 207 features, the cumulative score of dipeptide composition features was close to that of PSSM features, ranking third. This shows the importance of global sequence information. The average score of NetSurfP was much higher than other features, suggesting that the binding specificity of single-stranded and double-stranded DNA-binding proteins is closely correlated with structural information. The average value of the maximum correlation scores of each feature is shown in Figure 4.

2.2. Feature Extraction Results Analysis

DisEMBL is a method for predicting disordered regions in protein sequences with high accuracy. It profits from predicting protein disorder according to multiple definitions, including COILS, REM465, and HOTLOOPS [38]. To observe the prediction of the above three concepts more intuitively, we used a histogram to compare the prediction results processed by ACC algorithm, as shown in Figure 5. By observing the chart, we can easily find that the average scores of the three properties on SSBs are slightly higher than those on DSBs. This specificity is what we use to distinguish the two proteins.
In addition, the feature matrix of each protein obtained by NetSurfP contains five properties in total, and 30-dimensional row vectors are obtained after five iterations of ACC transformation. Overall, it can be concluded that the scores of three of the properties of DSBs (relative surface accessibility (RSA), absolute surface accessibility, and probability for beta-strand) are significantly higher than those of SSBs. Although SSBs and DSBs have many similar properties, proteins can specifically recognize single and double-stranded DNA and bind to the right place. The essential reason may be related to their structural characteristics. Therefore, NetSurfP plays a crucial role in the prediction.

2.3. Predictive Performance of Features

To further verify the predictive performance of the seven features, we constructed a classification model for each feature by combining the GTB algorithm with 10-fold cross-validation. Figure 6 depicts the ROC curves for the different features.
Different features have different abilities to identify SSBs and DSBs correctly. Consistent with the analysis of the feature selection results, PSSM had the optimal predictive ability for positive and negative samples, followed by NetSurfP. Overall, each feature plays an indispensable role in the prediction. Because of the low dimension of the feature matrix of features AAindex, LSE, and DisEMBL, they are not individually outstanding. However, they have functional complementarity and thus can play better roles when combined. For performance evaluation, we used several widely used measurement methods: accuracy, sensitivity (SN), specificity (SP), F1-score (F1), Matthew’s correlation coefficient (MCC), and area under ROC curve (AUC). Table 1 shows that the three features with the highest accuracies are PSSM, NetSurfP, and dipeptide, reaching 0.913, 0.874, and 0.836, respectively. On the one hand, it shows that the feature extraction algorithm we selected is very effective. On the other hand, it also reflects the fit between the GTB algorithm and this study.

2.4. Comparison with Previous Work

To our knowledge, there is only one existing study on SSB and DSB prediction based on sequence information, which uses two machine learning algorithms, SVM (Wang 2017_SVM) and RF (Wang 2017_RF). Furthermore, the accuracy of the two methods trains the model on Uniprot1065 dataset with an accuracy of 0.860 and 0.887, respectively. Finally, due to the defects of high sensitivity and low specificity in the performance of the model based on random forest, the SVM model was taken as the optimal model. In this paper, based on our larger dataset, we compare PredPSD with the above two methods. Overall, our method is superior to the existing methods in terms of accuracy, specificity, and sensitivity.
The results are presented in Figure 7 and Table 2. In general, we can see that our PredPSD method shows an advantage in all six metrics (Accuracy, SN, SP, F1, MCC, and AUC) on the training set. The optimal predictive energy of our method in the training dataset after redundancy removal is as follows: accuracy of 0.912, SN of 0.784, SP of 0.975, AUC of 0.956, and MCC of 0.799. By comparing with Wang 2017_SVM, we can see that the F1 score of our PredPSD approach has been significantly improved by 14.5%.

2.5. Comparative Analysis of Independent Test Results

We used the trained model to complete the test on the independent dataset to further verify the generalization ability of the model and avoid overfitting. The processing of the independent dataset is consistent with that of the training set. The exception is that no additional feature selection is carried out for the independent dataset, and the feature matrix of the independent dataset is reconstructed directly using the feature selection results obtained from the training set. The experimental results show that PredPSD presents the best predictability based on the combination of all features: accuracy of 0.770, SN of 0.512, SP of 0.855, AUC of 0.708, and F1 of 0.525. The ROC curve is shown in Figure 8.
The information in the figure indicates that PredPSD is superior to the existing optimal method of predicting SSBs and DSBs based on sequence information in the independent dataset. Detailed results of the comparisons between our approach and existing methods are shown in Table 3. From the overall analysis of the method comparison table, it can be seen that PredPSD has the most prominent comprehensive performance among the three methods. First, compared with method Wang 2017_SVM, our accuracy improved by 6.7%, and other evaluation indexes are greatly improved. The specificity and AUC values of the two methods are similar, but the sensitivity, MCC, and F1 are all improved by more than 10%. In particular, the MCC value increased by 18.8%. Second, although the specificity of Wang 2017_RF method is less than 0.05 different from our method, the sensitivity is only 0.341, so it has low validity. Finally, by comparing the performance of the three algorithms on the training set and the independent dataset, the results show that our model has stronger generalization ability. Furthermore, we annotate all the SSBs we used for classification by using the InterProScan tool. Then, proteins belonging to OB folds, KH domains, RRMs, and whirly domains were obtained by processing annotation information. Finally, the prediction results show that PredPSD has higher recognition ability for proteins with OB folds and whirly domains. However, the prediction accuracy of proteins with KH domains and RRMs structures is low. The reason may be that KH domain and RRMs can also specifically bind ssDNA sequences specifically, but these domains generally have a smaller ligand-binding site than OB folds and thus specify for fewer positions [39].

3. Materials and Methods

3.1. Datasets

For sequence-based feature calculation, we extracted 8833 DNA-binding proteins. Which contains 2136 DSBs and 339 SSBs obtained from the literature of Wang et al. [37] And the other part is collected from UniProtKB/Swiss-Prot (www.uniprot.org). To eliminate redundancy, CD-HIT was used to remove proteins with a sequence similarity > 70% [40]. Finally, we obtained a dataset of DNA binding proteins containing 1271 SSBs and 2252 DSBs. We took SSBs as the positive sample dataset and DSBs as the negative sample dataset.
Furthermore, to evaluate the classification performance and avoid overfitting, we obtained a non-redundant independent set of 124 DSBs and 41 SSBs from the PDB (www.rcsb.org/pdb/), which has the following characteristics: (1) sequence similarity with the training set is less than 40%; (2) sequence length is greater than 40 residues; (3) structure of each DNA-binding protein is known, and the resolution is better than 3 Å.

3.2. Feature Extraction

Selecting representative features is a crucial step because they directly determine prediction performance [41,42]. Seven sequence-based feature extraction methods were used: local structural entropy (LSE), NetSurfP, DisEMBL, overall amino acid composition (OAAC), dipeptide composition, PSSM profiles, and physicochemical properties. These methods have proven to be associated with the classification of DSBs and SSBs or have been used in similar fields. A more detailed description of how to extract and encode these different sequence-based features is provided below.

3.2.1. Local Structural Entropy (LSE)

LSE describes the conformational isomeric degree sequence of small proteins [43,44], which can provide useful information for protein classification. We used a method of computing LSE directly from sequence information.

3.2.2. NetSurfP

NetSurfP [45] is a tool that has been used to predict the secondary structure and surface accessibility of proteins based on sequence information [46]. It is an architecture composed of neural network [47] training on proteins with known structures. Local structural features such as relative surface accessibility, probability for alpha-helix, and probability for beta-strand play important roles in revealing the function of proteins and can also be used for protein classification.

3.2.3. DisEMBL

DisEMBL [38] is a method for predicting disordered regions in protein sequences. Disordered regions of proteins usually contain short linear peptide motifs such as SH3 ligands and targeting signals that are important for the classification of proteins by function.

3.2.4. Overall Amino Acid Composition (OAAC)

Existing studies have shown that the overall amino acid composition of 20 standard amino acids is a sequence feature widely used in the field of protein identification. Using the OAAC method, a 20-dimensional vector can be calculated, where each value describes the frequency of the amino acid in the sequence. Previous literature has shown that the square root of probability is more conducive to research [48]. Therefore, the following formula is used to define the probability
p i = n i L   ( i = 1 , 2 , 20 )
where n i refers to the number of times the i th amino acid in the protein sequence. Given a protein sequence S with length L, its OAAC feature vector can be expressed as
S O A A C = [ p 1 , p 2 , p 3 , , p 20 ]

3.2.5. Dipeptide Composition

A dipeptide is a compound formed by the dehydration and condensation of two amino acids. Here, it can be viewed as any combination of two amino acids. Since there are 20 kinds of amino acids, a total of 400 dipeptide compositions are possible [49,50]. Dipeptide composition is obtained by calculating the ratio of the number of occurrences of dipeptides in the sequence to the sequence length. Dipeptides are spaced differently in protein sequences. In this paper, three common distributions (0, 1, and 2) were selected [51]. Eventually, each protein will generate a vector of 1,200 dimensions. Dipeptides probabilities are defined as
p a b [ i ] = D a b [ i ] N 1   ( a , b = G , R , L ;   i = 0 , 1 , 2 )
where D a b [ i ] represents the number of dipeptides formed by two amino acids a and b at an interval of i   , and N is the length of the protein sequence.

3.2.6. PSSM

In this work, the practical significance of the position-specific scoring matrix (PSSM) is to find the conserved features of specific conserved positions from the sequences of DSBs and SSBs that can be used for the classification of the two types of proteins [52]. The PSSM of the residues is implemented by the PSI-BLAST [53] program, which contains essential evolution information through three iterations. A 20-dimensional vector with integer values represents each residue. These values represent the frequency of mutations at various locations in the sequence, and the PSSM can be expressed as
P S S M S = [ G 1 , 1 G 1 , 20 G L , 1 G L , 20 ]
where P S S M S represents a 20 × L matrix of protein S   , and L represents the length of the protein sequence. Gi,j is the probability score of the amino acid at position i of the S protein sequence being replaced by the basic amino acid encoding j during evolution.

3.2.7. Physicochemical Properties

The physicochemical properties of proteins are intuitive and straightforward basic characteristics with reliable physical and biological meanings [54,55]. We selected 28 typical numerical properties [56] commonly used for DNA binding protein classification in the database AAindex [57] to encode amino acids. A protein sequence of length L can be expressed as a matrix of 28 × L dimensions, where each row represents the attribute value of the residue at that location. The list of AAindex physicochemical properties we used can be found in Supplementary Table S4.

3.3. Feature Transformation

Protein sequences usually have different lengths. However, machine learning-based methods such as GTB require fixed-length vectors for training. Here, we introduce the autocross-covariance (ACC) transformation to transform protein sequences into fixed-length vectors by measuring the correlation of two properties along the protein sequence [58]. The ACC method contains two variables, AC and CC. AC is used to calculate the correlation of two residues with a distance of lg in the same attribute. It is defined as
A C ( i , l g ) = j = 1 L l g ( s i j s i ¯ ) ( s i j + l g s i ¯ )   /   ( L lg )
where i is one of the columns corresponding to a residue, l g is the distance between the two residues, L is the number of residues in the protein sequence, s i j is the value of the i th row and the j th column in the matrix, and s i ¯ is the average score for L columns
s i ¯ = j = 1 L s i j / L
Therefore, the number of AC variables is obtained by multiplying the number of attributes by the number of LG values. LG is the maximum of the intervals lg.
The CC variable calculates the relationship between two different attributes. The specific calculation is
CC ( i 1 , i 2 , l g ) = j = 1 L l g ( S i 1 , j S i 1 ¯ ) ( S i 2 , j + l g S i 2 ¯ ) / ( L l g )
where i 1 and i 2 represent the columns corresponding to two different attributes, and S i 1 ¯ ( S i 2 ¯   ) is the average value of the i th column. We finally choose the ACC variable, which is the result of the combination of the AC variable and the CC variable.
In this work, ACC transformation was used on the matrix obtained by physicochemical, PSSM, DisEMBL, and NetSurfP. The feature matrix of each protein is converted into a vector, where the parameter LG is set to 5, and the visual description is shown in Figure 9.

3.4. Feature Selection

A feature selection algorithm can help us understand the characteristics of features, and it plays a vital role in further optimizing the algorithm and improving classification accuracy [59]. The candidate feature space selected by minimal-redundancy-maximal-relevance criterion (mRMR) is more representative [60,61]. Therefore, based on the candidate features of 1520 columns, we further select the optimal feature subset using the mRMR algorithm. The general process of mRMR feature selection is as follows: first, the data are processed and stored with the appropriate data structure. Second, the distribution and mutual information between features and between features and response variables are calculated. Finally, the features are scored and sorted by mRMR. A score file maximum correlation and minimum redundancy for each column feature are then obtained.
Specifically, the correlation between feature subset and category is calculated by the mean of the information gain of each feature and category
max D ( S , c ) ,   D = 1 | S | x i S I ( x i ; c )
where S is a feature set, c is the target category, x i represents a feature in the feature set S, and I ( x i ; c ) indicates all the mutual information values between a single feature x i and class c . max D ( S , c ) represents that x i in S has the highest dependence on the target class c .
The redundancy between two features is calculated by the sum of mutual information between two features and then divided by the square of the feature number in the subset
min R ( S ) ,   R = 1 | S | 2 x i , x j S I ( x i ; x j )
where I ( x i ; x j ) is the mutual information between two classes. If the two classes are highly dependent on each other, removing one of them will not affect classification performance.

3.5. Classification Model and Performance Evaluation

Gradient tree boosting (GTB) [62] is an integrated base classifier decision tree algorithm that can be used for classification and regression problems [63,64,65,66,67]. In this study, it is assumed that SSBs and DSBs belong to a binary classification problem. We finally chose the gradient tree boosting of ‘sklearn.ensemble’ as the classification method, because it can better address mixed types of data and is more robust to outliers. GTB produces a decision tree composed of J leaf nodes by reducing the gradient direction of each sample point and its residuals [68,69,70]. In the experiment, the optimal parameters of GTB were selected by 10-fold cross-validation on the benchmark dataset using a grid search strategy. These performance evaluations we use are defined as
SN = TP / ( TP + FN )
SP = TN / ( TN + FP )
F 1 = 2 × Recall × Precision Recall + Precision
Accuracy = TP + TN TP + TN + FP + FN
MCC = TP × TN FP × FN ( TP + FP ) ( TP + FN ) ( TN + FP ) ( TN + FN )  
In these equations, TP (the number of SSBs correctly classified), TN (the number of DSBs correctly classified), FP (the number of DSBs that are misclassified as SSBs) and FN (the number of SSBs that are misclassified as DSBs) represent true positives, true negatives, false positives and false negatives, respectively. Here, the category of SSBs is called a positive class, and the category of DSBs is called a negative class. Among these evaluation indexes, MCC and F1 reflect the overall performance of the classifier and can better evaluate the performance of the classifier in the case of unbalanced data. The classification performance was evaluated by 10-fold cross-validation.

4. Conclusions

In this work, we have proposed a novel method PredPSD for the classification prediction of SSBs and DSBs. The method is based on the gradient tree boosting (GTB) algorithm [71], and the model was trained on 1271 SSBs and 2252 DSBs non-redundant datasets. In the course of the experiment, we introduced three feature extraction algorithms that have never been used in this research problem that can extract a variety of features derived from multiple sequences. At the same time, we also combined four commonly used algorithms. We comprehensively evaluated the effects of different sequence extraction methods on prediction performance. Then, we used the ACC transformation algorithm to solve the problem of inconsistent feature dimensions. The mRMR method was used to obtain a set of optimal features with maximum correlation and minimum redundancy, and the strong specificity of the new features to distinguish different types of proteins was verified.
PredPSD resulted in prediction accuracy of 91.2% and an AUC of 0.956 on the training set through 10-fold cross-validation. Furthermore, in the independent dataset, PredPSD can achieve an accuracy of 77.0% and an SP of 0.855. The results show that the prediction performance of PredPSD is better than that of previous methods. Our study provides a complementary and effective method to predict SSBs and DSBs more accurately.

Supplementary Materials

The following are available online at https://www.mdpi.com/1420-3049/25/1/98/s1.

Author Contributions

Conceptualization, C.T., T.W., and L.D.; Methodology, C.T., T.W., W.Y., and L.D.; Validation, T.W. and W.Y.; Investigation, C.T., T.W., and L.D.; Writing—original draft preparation, T.W.; Writing—review and editing, C.T. and L.D.; Supervision, L.D.; Funding acquisition, L.D. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funding by National Natural Science Foundation of China (grant nos. 61972422 and 61672541).

Acknowledgments

The authors acknowledge the Experimental Center of School of Computer Science and Engineering of Central South University for providing computing resources.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Laetitia, A.; Audrey, O.; Isabelle, M.B.; Anne-Lise, S.; Chantal, G.; Bernard, M.; Patrice, P.; Jean-Pierre, C. Role of the single-stranded DNA-binding protein SsbB in pneumococcal transformation: Maintenance of a reservoir for genetic plasticity. PLoS Genet. 2011, 7, e1002156. [Google Scholar]
  2. Edsö, J.R. Single- and double-stranded DNA binding proteins act in concert to conserve a telomeric DNA core sequence. Genome Integr. 2011, 2, 2. [Google Scholar] [CrossRef] [Green Version]
  3. Richard, D.J.; Emma, B.; Liza, C.; Wadsworth, R.I.M.; Kienan, S.; Sharma, G.G.; Nicolette, M.L.; Sergie, T.; Mcilwraith, M.J.; Pandita, R.K. Single-stranded DNA-binding protein hSSB1 is critical for genomic stability. Nature 2008, 453, 677–681. [Google Scholar] [CrossRef]
  4. Olga, S.N.; Lue, N.F. Telomere DNA recognition in Saccharomycotina yeast: Potential lessons for the co-evolution of ssDNA and dsDNA-binding proteins and their target sites. Front. Genet. 2015, 6, 162. [Google Scholar]
  5. Croft, L.V.; Bolderson, E.; Adams, M.N.; El-Kamand, S.; Kariawasam, R.; Cubeddu, L.; Gamsjaeger, R.; Richard, D.J. Human single-stranded DNA binding protein 1 (hSSB1, OBFC2B), a critical component of the DNA damage response. Semin. Cell Dev. Biol. 2018, 86, 121–128. [Google Scholar] [CrossRef] [Green Version]
  6. Emmanuelle, D.; Amélie, H.M.; Giuseppe, B. Single-stranded DNA binding proteins unwind the newly synthesized double-stranded DNA of model miniforks. Biochemistry 2011, 50, 932–944. [Google Scholar]
  7. Doan, P.H.; Pitter, D.R.G.; Kocher, A.; Wilson, J.N.; Iii, G. A New Design Strategy and Diagnostic to Tailor the DNA-Binding Mechanism of Small Organic Molecules and Drugs. ACS Chem. Biol. 2016, 11, 3202–3213. [Google Scholar] [CrossRef]
  8. Dasgupta, D.; Howard, F.B.; Sasisekharan, V.; Miles, H.T. Drug-DNA binding specificity: Binding of netropsin and distamycin to poly(d2NH2A-dT). Biopolymers 2010, 30, 223–227. [Google Scholar] [CrossRef]
  9. Gao, Y.G.; Priebe, W.; Wang, A.H. Substitutions at C2’ of daunosamine in the anticancer drug daunorubicin alter its DNA-binding sequence specificity. Eur. J. Biochem. 2010, 240, 331–335. [Google Scholar] [CrossRef]
  10. Liu, H.; Zhang, W.; Zou, B.; Wang, J.; Deng, Y.; Deng, L. DrugCombDB: A comprehensive database of drug combinations toward the discovery of combinatorial therapy. Nucleic Acids Res. 2019. [Google Scholar] [CrossRef]
  11. Wang, W.; Liu, J.; Xiong, Y.; Zhu, L.; Zhou, X. Analysis and classification of DNA-binding sites in single-stranded and double-stranded DNA-binding proteins using protein information. IET Syst. Biol. 2014, 8, 176. [Google Scholar] [CrossRef] [Green Version]
  12. Tchurikov, N.A.; Fedoseeva, D.M.; Sosin, D.V.; Snezhkina, A.V.; Melnikova, N.V.; Kudryavtseva, A.V.; Kravatsky, Y.V.; Kretova, O.V. Hot spots of DNA double-strand breaks and genomic contacts of human rDNA units are involved in epigenetic regulation. J. Mol. Cell Biol. 2015, 7, 366–382. [Google Scholar] [CrossRef]
  13. Zhu, X.; Ericksen, S.S.; Mitchell, J.C. DBSI: DNA-binding site identifier. Nucleic Acids Res. 2013, 41, e160. [Google Scholar] [CrossRef] [Green Version]
  14. Yan, C.; Terribilini, M.; Wu, F.; Jernigan, R.L.; Dobbs, D.; Honavar, V. Predicting DNA-binding sites of proteins from amino acid sequence. BMC Bioinform. 2006, 7, 262. [Google Scholar] [CrossRef] [Green Version]
  15. Nagarajan, R.; Ahmad, S.; Gromiha, M.M. Novel approach for selecting the best predictor for identifying the binding sites in DNA binding proteins. Nucleic Acids Res. 2013, 41, 7606–7614. [Google Scholar] [CrossRef] [Green Version]
  16. Qu, K.; Wei, L.; Zou, Q. A Review of DNA-binding Proteins Prediction Methods. Curr. Bioinform. 2019, 14, 246–254. [Google Scholar] [CrossRef]
  17. Wei, L.; Tang, J.; Zou, Q. Local-DPP: An improved DNA-binding protein prediction method by exploring local evolutionary information. Inf. Sci. 2017, 384, 135–144. [Google Scholar] [CrossRef]
  18. Song, L.; Li, D.; Zeng, X.; Wu, Y.; Guo, L.; Zou, Q. nDNA-prot: Identification of DNA-binding proteins based on unbalanced classification. BMC Bioinform. 2014, 15, 298. [Google Scholar] [CrossRef] [Green Version]
  19. Shula, S.; Gershon, E.; Yael, M.G. From face to interface recognition: A differential geometric approach to distinguish DNA from RNA binding surfaces. Nucleic Acids Res. 2011, 39, 7390. [Google Scholar]
  20. Nimrod, G.; Szilágyi, A.; Leslie, C.; Ben-Tal, N. Identification of DNA-binding proteins using structural, electrostatic and evolutionary features. J. Mol. Biol. 2009, 387, 1040–1053. [Google Scholar] [CrossRef] [Green Version]
  21. Szabóová, A. Prediction of DNA-binding propensity of proteins by the ball-histogram method using automatic template search. BMC Bioinform. 2012, 13, 1–11. [Google Scholar] [CrossRef] [Green Version]
  22. Zou, Q.; Wan, S.; Ju, Y.; Tang, J.; Zeng, X. Pretata: Predicting TATA binding proteins with novel features and dimensionality reduction strategy. BMC Syst. Biol. 2016, 10, 114. [Google Scholar] [CrossRef] [Green Version]
  23. Jolma, A.; Yan, J.; Whitington, T.; Toivonen, J.; Nitta, K.; Rastas, P.; Morgunova, E.; Enge, M.; Taipale, M.; Wei, G. DNA-Binding Specificities of Human Transcription Factors. Cell 2013, 152, 327–339. [Google Scholar] [CrossRef] [Green Version]
  24. Wei-Zhong, L.; Jian-An, F.; Xuan, X.; Kuo-Chen, C. iDNA-Prot: Identification of DNA binding proteins using random forest with grey model. PLoS ONE 2011, 6, e24756. [Google Scholar]
  25. Morgan, H.; Estibeiro, P.; Wear, M.; Heinemann, U.; Cubeddu, L.; Gallagher, M.; Sadler, P.J.; Walkinshaw, M.D. Sequence specificity of single-stranded DNA-binding proteins: A novel DNA microarray approach. Nucleic Acids Res. 2007, 35, e75. [Google Scholar] [CrossRef] [Green Version]
  26. Kresten, L.L.; Best, R.B.; Depristo, M.A.; Dobson, C.M.; Michele, V. Simultaneous determination of protein structure and dynamics. Nature 2005, 433, 128–132. [Google Scholar]
  27. Wang, W.; Liu, J.; Zhou, X. Identification of single-stranded and double-stranded dna binding proteins based on protein structure. Bioinformatics 2013, 15, S4. [Google Scholar] [CrossRef] [Green Version]
  28. Francesco, R.; Bonham, A.J.; Mason, A.C.; Reich, N.O.; Plaxco, K.W. Reagentless, electrochemical approach for the specific detection of double- and single-stranded DNA binding proteins. Anal. Chem. 2009, 81, 1608–1614. [Google Scholar]
  29. Cai, Y.D.; Doig, A.J. Prediction of Saccharomyces cerevisiae protein functional class from functional domain composition. Bioinformatics 2004, 20, 1292–1300. [Google Scholar] [CrossRef] [Green Version]
  30. Yu, E.; Wang, F.; Lei, M.; Lue, N. A proposed OB-fold with a protein-interaction surface in Candida albicans telomerase protein Est3. Nat. Struct. Mol. Biol. 2008, 15, 985. [Google Scholar]
  31. Zasedateleva, O.A.; Mikheikin, A.L.; Turygin, A.Y.; Prokopenko, D.V.; Chudinov, A.V.; Belobritskaya, E.E.; Chechetkin, V.R.; Zasedatelev, A.S. Gel-based oligonucleotide microarray approach to analyze protein-ssDNA binding specificity. Nucleic Acids Res. 2008, 36, e61. [Google Scholar] [CrossRef] [Green Version]
  32. Wang, W.; Liu, J.; Sun, L. Surface shapes and surrounding environment analysis of single- and double-stranded DNA-binding proteins in protein-DNA interface. Proteins-Struct. Funct. Bioinform. 2016, 84, 979–989. [Google Scholar] [CrossRef]
  33. Remo, R.; West, S.M.; Alona, S.; Peng, L.; Mann, R.S.; Barry, H. The role of DNA shape in protein-DNA recognition. Nature 2009, 461, 1248–1253. [Google Scholar]
  34. Rim, K.I.W.; Cubeddu, L.; Blankenfeldt, W.; Naismith, J.H.; White, M.F. Insights into ssDNA recognition by the OB fold from a structural and thermodynamic study of Sulfolobus SSB protein. EMBO J. 2014, 22, 2561–2570. [Google Scholar]
  35. Yi, X.; Juan, L.; Dong-Qing, W. An accurate feature-based method for identifying DNA-binding residues on protein surfaces. Proteins-Struct. Funct. Bioinform. 2011, 79, 509–517. 79 2011, 79, 509–517. [Google Scholar]
  36. Taisuke, W.; Yoshiaki, K.; Yutaro, K.; Noriko, N.; Seiki, K.; Ryoji, M. Structure of RecJ exonuclease defines its specificity for single-stranded DNA. J. Biol. Chem. 2010, 285, 9762–9769. [Google Scholar]
  37. Wang, W.; Sun, L.; Zhang, S.; Zhang, H.; Shi, J.; Xu, T.; Li, K. Analysis and prediction of single-stranded and double-stranded DNA binding proteins based on protein sequences. BMC Bioinform. 2017, 18, 300. [Google Scholar] [CrossRef] [Green Version]
  38. Linding, R.; Jensen, L.J.; Diella, F.; Bork, P.; Gibson, T.J.; Russell, R.B. Protein Disorder Prediction: Implications for Structural Proteomics. Structure 2003, 11, 1453–1459. [Google Scholar] [CrossRef] [Green Version]
  39. Dickey, T.H.; Altschuler, S.E.; Wuttke, D.S. Single-stranded DNA-binding proteins: Multiple domains for multiple functions. Structure 2013, 21, 1074–1084. [Google Scholar] [CrossRef] [Green Version]
  40. Li, W. Fast Program for Clustering and Comparing Large Sets of Protein or Nucleotide Sequences. Bioinformatics 2006, 22, 1658. [Google Scholar] [CrossRef] [Green Version]
  41. Zhu, X.J.; Feng, C.Q.; Lai, H.Y.; Chen, W.; Lin, H. Predicting protein structural classes for low-similarity sequences by evaluating different features. Knowl. Based Syst. 2019, 163, 787–793. [Google Scholar] [CrossRef]
  42. Tan, J.X.; Li, S.H.; Zhang, Z.M.; Chen, C.X.; Chen, W.; Tang, H.; Lin, H. Identification of hormone binding proteins based on machine learning methods. Math. Biosci. Eng. 2019, 16, 2466–2480. [Google Scholar] [CrossRef]
  43. Chan, C.H.; Liang, H.K.; Hsiao, N.W.; Ko, M.T.; Lyu, P.C.; Hwang, J.K. Relationship between local structural entropy and protein thermostabilty. Proteins Struct. Funct. Bioinform. 2004, 57, 684–691. [Google Scholar] [CrossRef]
  44. Deng, L.; Guan, J.; Wei, X.; Yi, Y.; Zhang, Q.C.; Zhou, S. Boosting prediction performance of protein-protein interaction hot spots by using structural neighborhood properties. J. Comput. Biol. J. Comput. Mol. Cell Biol. 2013, 20, 878–891. [Google Scholar] [CrossRef] [Green Version]
  45. Agnew, H.D.; Coppock, M.B.; Idso, M.N.; Lai, B.T.; Liang, J.; McCarthy-Torrens, A.M.; Warren, C.M.; Heath, J.R. Protein-catalyzed capture agents. Chem. Rev. 2019, 119, 9950–9970. [Google Scholar] [CrossRef]
  46. Zhang, W.; Liu, J.; Zhao, M.; Li, Q. Predicting linear B-cell epitopes by using sequence-derived structural and physicochemical features. Int. J. Data Min. Bioinform. 2012, 6, 557–569. [Google Scholar] [CrossRef]
  47. Kuang, L.; Yan, X.; Tan, X.; Li, S.; Yang, X. Predicting Taxi Demand Based on 3D Convolutional Neural Network and Multi-task Learning. Remote Sens. 2019, 11, 1265. [Google Scholar] [CrossRef] [Green Version]
  48. Feng, Z.P. Prediction of the subcellular location of prokaryotic proteins based on a new representation of the amino acid composition. Biopolymers 2015, 58, 491–499. [Google Scholar] [CrossRef]
  49. Garg, A.; Raghava, G.P. ESLpred2: Improved method for predicting subcellular localization of eukaryotic proteins. BMC Bioinform. 2008, 9, 1–10. [Google Scholar] [CrossRef] [Green Version]
  50. Tang, H.; Zhao, Y.W.; Zou, P.; Zhang, C.M.; Chen, R.; Huang, P.; Lin, H. HBPred: A tool to identify growth hormone-binding proteins. Int. J. Biol. Sci. 2018, 14, 957–964. [Google Scholar] [CrossRef]
  51. Hao, L.; Qian-Zhong, L. Using pseudo amino acid composition to predict protein structural class: Approached by incorporating 400 dipeptide components. J. Comput. Chem. 2010, 28, 1463–1466. [Google Scholar]
  52. Ahmad, S.; Sarai, A. PSSM-based prediction of DNA binding sites in proteins. BMC Bioinform. 2005, 6, 1–6. [Google Scholar] [CrossRef] [Green Version]
  53. Altschul, S.F.; Madden, T.L.; Schäffer, A.A.; Zhang, J.; Zhang, Z.; Miller, W.; Lipman, D.J. Gapped BLAST and PSI-BLAST—A new generation of protein database search programs. Nucleic Acids Res. 1997, 25, 3389–3402. [Google Scholar] [CrossRef] [Green Version]
  54. Yang, W.; Zhu, X.J.; Huang, J.; Ding, H.; Lin, H. A brief survey of machine learning methods in protein sub-Golgi localization. Curr. Bioinform. 2019, 14, 234–240. [Google Scholar] [CrossRef]
  55. Tang, H.; Chen, W.; Lin, H. Identification of immunoglobulins using Chou’s pseudo amino acid composition with feature selection technique. Mol. Biosyst. 2016, 12, 1269–1275. [Google Scholar] [CrossRef]
  56. Huang, H.L.; Lin, I.C.; Liou, Y.F.; Tsai, C.T.; Hsu, K.T.; Huang, W.L.; Ho, S.J.; Ho, S.Y. Predicting and analyzing DNA-binding domains using a systematic approach to identifying a set of informative physicochemical and biochemical properties. BMC Bioinform. 2011, 12, S47. [Google Scholar] [CrossRef] [Green Version]
  57. Kawashima, S.; Kanehisa, M. AAindex: Amino Acid index database. Nucleic Acids Res. 1999, 27, 368–369. [Google Scholar] [CrossRef]
  58. Dong, Q.; Zhou, S.; Guan, J. A new taxonomy-based protein fold recognition approach based on autocross-covariance transformation. Bioinformatics 2009, 25, 2655–2662. [Google Scholar] [CrossRef] [Green Version]
  59. Zhang, J.; Liu, B. A Review on the Recent Developments of Sequence-based Protein Feature Extraction Methods. Curr. Bioinform. 2019, 14, 190–199. [Google Scholar] [CrossRef]
  60. Hanchuan, P.; Fuhui, L.; Chris, D. Feature selection based on mutual information: Criteria of max-dependency, max-relevance, and min-redundancy. IEEE Trans. Pattern Anal. Mach. Intell. 2005, 27, 1226–1238. [Google Scholar] [CrossRef]
  61. Wang, S.P.; Zhang, Q.; Lu, J.; Cai, Y.D. Analysis and Prediction of Nitrated Tyrosine Sites with the mRMR Method and Support Vector Machine Algorithm. Curr. Bioinform. 2018, 13, 3–13. [Google Scholar] [CrossRef]
  62. Friedman, J.H. Stochastic gradient boosting. Comput. Stat. Data Anal. 2002, 38, 367–378. [Google Scholar] [CrossRef]
  63. Hoque, M.T.; Chetty, M.; Lewis, A.; Sattar, A. Twin Removal in Genetic Algorithms for Protein Structure Prediction Using Low-Resolution Model. IEEE/ACM Trans. Comput. Biol. Bioinform. 2011, 8, 234–245. [Google Scholar] [CrossRef] [Green Version]
  64. Liu, D.; Tang, Y.; Chao, F.; Chen, Z.; Lei, D. PredRBR: Accurate Prediction of RNA-Binding Residues in proteins using Gradient Tree Boosting. In Proceedings of the IEEE International Conference on Bioinformatics & Biomedicine, Shenzhen, China, 15–18 December 2016. [Google Scholar]
  65. He, T.; Heidemeyer, M.; Ban, F.; Cherkasov, A.; Ester, M. SimBoost: A read-across approach for predicting drug–target binding affinities using gradient boosting machines. J. Cheminform. 2017, 9, 24. [Google Scholar] [CrossRef]
  66. Li, Y.; Niu, M.; Zou, Q. ELM-MHC: An improved MHC Identification method with Extreme Learning Machine Algorithm. J. Proteome Res. 2019, 18, 1392–1401. [Google Scholar] [CrossRef]
  67. Dou, K.; Guo, B.; Kuang, L. A privacy-preserving multimedia recommendation in the context of social network based on weighted noise injection. Multimed. Tools Appl. 2019, 78, 26907–26926. [Google Scholar] [CrossRef]
  68. Fan, C.; Liu, D.; Huang, R.; Chen, Z.; Deng, L. PredRSA: A gradient boosted regression trees approach for predicting protein solvent accessibility. BMC Bioinform. 2016, 17 (Suppl. 1), 8. [Google Scholar] [CrossRef] [Green Version]
  69. Pan, Y.; Wang, Z.; Zhan, W.; Deng, L. Computational identification of binding energy hot spots in protein-RNA complexes using an ensemble approach. Bioinformatics 2018, 34, 1473–1480. [Google Scholar] [CrossRef]
  70. Wen, Z.; Hua, Z.; Luo, L.; Liu, Q.; Wu, W.; Xiao, W. Predicting potential side effects of drugs by recommender methods and ensemble learning. Neurocomputing 2016, 173, 979–987. [Google Scholar]
  71. Deng, L.; Li, W.; Zhang, J. LDAH2V: Exploring meta-paths across multiple networks for lncRNA-disease association prediction. IEEE/ACM Trans. Comput. Biol. Bioinform. 2019. [Google Scholar] [CrossRef]
Sample Availability: Samples of the compounds are not available from the authors.
Figure 1. Workflow of PredPSD.
Figure 1. Workflow of PredPSD.
Molecules 25 00098 g001
Figure 2. Taking the maximum correlation score as a threshold, the number of feature columns changes with the threshold value.
Figure 2. Taking the maximum correlation score as a threshold, the number of feature columns changes with the threshold value.
Molecules 25 00098 g002
Figure 3. Probability of each feature being selected.
Figure 3. Probability of each feature being selected.
Molecules 25 00098 g003
Figure 4. The top five maximum correlation scores of features obtained by different feature extraction methods were calculated.
Figure 4. The top five maximum correlation scores of features obtained by different feature extraction methods were calculated.
Molecules 25 00098 g004
Figure 5. Average scores of DisEMBL’s three properties on SSBs and DSBs.
Figure 5. Average scores of DisEMBL’s three properties on SSBs and DSBs.
Molecules 25 00098 g005
Figure 6. ROC curves of the GTB model on the training set. Different colors represent different types of features.
Figure 6. ROC curves of the GTB model on the training set. Different colors represent different types of features.
Molecules 25 00098 g006
Figure 7. Performance comparison between our method and three existing classification methods on SSBs and DSBs from our training set. The existing classification methods are Wang 2017_SVM (based on sequence information and an SVM algorithm) and Wang 2017_RF (based on sequence information and an RF algorithm), all proposed by Wang W. et al. To our knowledge only these two methods are all relevant to this study.
Figure 7. Performance comparison between our method and three existing classification methods on SSBs and DSBs from our training set. The existing classification methods are Wang 2017_SVM (based on sequence information and an SVM algorithm) and Wang 2017_RF (based on sequence information and an RF algorithm), all proposed by Wang W. et al. To our knowledge only these two methods are all relevant to this study.
Molecules 25 00098 g007
Figure 8. ROC curve based on the independent dataset.
Figure 8. ROC curve based on the independent dataset.
Molecules 25 00098 g008
Figure 9. Application of the ACC transformation in this study. For example, protein1 refers to the name of a protein, and features extracted based on its sequence information are vectors, while others are matrices. In this case, the matrix can be converted to a vector.
Figure 9. Application of the ACC transformation in this study. For example, protein1 refers to the name of a protein, and features extracted based on its sequence information are vectors, while others are matrices. In this case, the matrix can be converted to a vector.
Molecules 25 00098 g009
Table 1. Performance of classification models derived from different types of features
Table 1. Performance of classification models derived from different types of features
FeaturesAccuracySNSPAUCMCCF1
PSSM0.9130.8140.9680.9680.8090.870
AAindex0.7590.5140.8990.8100.4570.604
OAAC0.7780.5720.8950.8430.5030.649
NetSurfP0.8740.7800.9270.9380.7230.817
LSE0.6680.4300.7820.6460.2190.456
Dipeptide0.8380.6450.9480.9120.6440.741
DisEMBL0.6820.4210.8180.6700.2580.477
Table 2. Performance of all feature descriptors with GTB algorithms based on our training set and comparisons with existing methods
Table 2. Performance of all feature descriptors with GTB algorithms based on our training set and comparisons with existing methods
FeaturesAccuracySNSPAUCMCCF1
PredPSD0.9120.7840.9750.9560.7990.854
Wang 2017_SVM0.8550.5850.9720.9270.6430.709
Wang 2017_RF0.7840.5780.9020.8270.5180.661
Table 3. Performance of all feature descriptors with GTB algorithms based on the independent dataset and comparisons with existing methods.
Table 3. Performance of all feature descriptors with GTB algorithms based on the independent dataset and comparisons with existing methods.
FeaturesAccuracySNSPAUCMCCF1
PredPSD0.7700.5120.8550.7080.3730.525
Wang 2017_SVM0.7030.3660.8140.6920.1850.380
Wang 2017_RF0.7210.3410.8470.6200.2030.378

Share and Cite

MDPI and ACS Style

Tan, C.; Wang, T.; Yang, W.; Deng, L. PredPSD: A Gradient Tree Boosting Approach for Single-Stranded and Double-Stranded DNA Binding Protein Prediction. Molecules 2020, 25, 98. https://doi.org/10.3390/molecules25010098

AMA Style

Tan C, Wang T, Yang W, Deng L. PredPSD: A Gradient Tree Boosting Approach for Single-Stranded and Double-Stranded DNA Binding Protein Prediction. Molecules. 2020; 25(1):98. https://doi.org/10.3390/molecules25010098

Chicago/Turabian Style

Tan, Changgeng, Tong Wang, Wenyi Yang, and Lei Deng. 2020. "PredPSD: A Gradient Tree Boosting Approach for Single-Stranded and Double-Stranded DNA Binding Protein Prediction" Molecules 25, no. 1: 98. https://doi.org/10.3390/molecules25010098

Article Metrics

Back to TopTop