Next Article in Journal
Role of Vitronectin and Its Receptors in Neuronal Function and Neurodegenerative Diseases
Next Article in Special Issue
Effect of Silica Microparticles on Interactions in Mono- and Multicomponent Membranes
Previous Article in Journal
Quenching of Protein Fluorescence by Fullerenol C60(OH)36 Nanoparticles
Previous Article in Special Issue
Interspecies Comparison of Interaction Energies between Photosynthetic Protein RuBisCO and 2CABP Ligand
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

PD-BertEDL: An Ensemble Deep Learning Method Using BERT and Multivariate Representation to Predict Peptide Detectability

College of Information and Computer, Taiyuan University of Technology, Taiyuan 030024, China
*
Author to whom correspondence should be addressed.
Int. J. Mol. Sci. 2022, 23(20), 12385; https://doi.org/10.3390/ijms232012385
Submission received: 1 September 2022 / Revised: 11 October 2022 / Accepted: 12 October 2022 / Published: 16 October 2022
(This article belongs to the Collection Computational Studies of Biomolecules)

Abstract

:
Peptide detectability is defined as the probability of identifying a peptide from a mixture of standard samples, which is a key step in protein identification and analysis. Exploring effective methods for predicting peptide detectability is helpful for disease treatment and clinical research. However, most existing computational methods for predicting peptide detectability rely on a single information. With the increasing complexity of feature representation, it is necessary to explore the influence of multivariate information on peptide detectability. Thus, we propose an ensemble deep learning method, PD-BertEDL. Bidirectional encoder representations from transformers (BERT) is introduced to capture the context information of peptides. Context information, sequence information, and physicochemical information of peptides were combined to construct the multivariate feature space of peptides. We use different deep learning methods to capture the high-quality features of different categories of peptides information and use the average fusion strategy to integrate three model prediction results to solve the heterogeneity problem and to enhance the robustness and adaptability of the model. The experimental results show that PD-BertEDL is superior to the existing prediction methods, which can effectively predict peptide detectability and provide strong support for protein identification and quantitative analysis, as well as disease treatment.

1. Introduction

Peptide detectability is defined as the probability of detecting a peptide from a standard sample analyzed by a proteomics program [1]. It is used to measure the relationship between the amount of protein in the sample and the peptide detected. In a shotgun protein assay, proteins in the mixture sample are enzymatically decomposed into peptides, and high-throughput peptide analysis is performed by liquid chromatography-tandem mass spectrometry (LC-MS/MS) and other techniques to determine the composition and content of proteins in the sample [1,2,3]. In this process, non-site cleavage of sequences, loss of peptide generated during enzymatic hydrolysis, and other abnormalities may lead to the deviation of the probability of peptide detection [2], thus restricting the identification and quantitative calculation of proteins. Studies have shown that peptide detectability is crucial to detection, analysis, and differential expression of proteins in proteomics [4,5,6]. Therefore, accurate prediction of peptide detectability helps protein detection and expression analysis provides reference value for the discovery of disease biomarkers and clinical research and thus helps us to more deeply understand cell biology and the underlying mechanisms of human diseases.
There is a strong correlation between amino acids and amino acids in the sequence. Researchers introduce contextual information as the feature representation of the peptide sequence level to better describe the global information of the sequence. Charoenkwan et al. [7] captured the context information of bitter peptides and used it as a feature, and used BiLSTM and DNN to identify bitter peptides. The results showed that the context information could effectively improve the recognition accuracy of bitter peptides. In peptide detectability, Serrano et al. [8] regarded amino acids as words, utilized word2vec to calculate the embedding vector of each word, and input the convolutional neural network to predict peptide detectability. Cheng et al. [9] used transformer and bidirectional gated neural network (BiGRU) to capture the context information of peptides, and improved the prediction performance of peptide detectability. These methods capture the context information of peptides and reflect the global information of peptides better. However, word2vec regards amino acids as words to calculate the embedding vector of each amino acid, but the vector is a fixed value. For the same amino acid, the embedding vector is the same regardless of its context, so it cannot fully reflect the context information of the peptides [9,10,11]. Transformer uses the attention mechanism to capture the association between the target word and other words in the text in order to enhance the target word’s semantic representation, but the target word itself will consist of the semantics of the main parts, leading to the information contained in each word after encoding being more inclined to the meaning of the word itself [12,13], which could not obtain high quality context features.
With the continuous development of natural language processing (NLP), the researchers used a new NLP technology-BERT to capture the context information. BERT adopts mask language model (MLM) to strengthen the dependence on context and uses output vector as the semantic representation of the whole sentence, thus integrating the semantic information of each word in the text more “fairly” and solve the problem of self-bias [14]. It has been successfully applied to protein post-translational modification site prediction [15], peptide recognition [7,16] and gene sequence study [11,17,18]. Therefore, BERT can effectively fuse the information of each amino acid to capture the context information of peptides and enhance the prediction performance of peptide detectability.
There are numerous factors that affect peptide detectability, including sequence information, physicochemical properties [19,20,21,22], context information derived from sequences, etc. [8,9]. Tang et al. [1] encoded the sequence derived information of peptides and input the feedforward neural network to predict the detection ability of peptides. Li et al. [19] considered the 292-dimensional physicochemical properties, such as peptide length, constructed artificial neural networks to predict peptide detectability. Guruceaga et al. [21] screened the 106-dimensional physicochemical properties of peptides, such as molecular weight and theoretical isoelectric point, constructed random forest classifier (RF) for peptide detection. Serrano et al. [8] extracted context information derived from sequences and input it to a convolutional neural network for feature learning, which improved the prediction accuracy of peptide detectivity. In the above studies, sequence information, physicochemical properties, and context information derived from sequences were used to predict peptide detectivity. However, these methods only considered information at a certain level of peptide, resulting in the simplification of the constructed feature space and the absence of some information. Zhang et al. [23] considered sequence information, sequence derived pseudo-amino acid composition information, and physicochemical properties, using LightGBM to predict non-classical secreted proteins. Xu et al. [24] used evolutionary information, the physicochemical properties of proteins, and sequence derived K-spaced amino acid pairs information to predict lysine succinylation sites by support vector machine (SVM), indicating that multivariate representation could make up for the unicity of feature space and effectively improve the accuracy of site prediction. The sequence information, physicochemical properties, and sequence derived information of peptides represent peptides from different perspectives. Considering three kinds of information simultaneously to predict peptide detectability can help the model to learn diversified feature representations, obtain richer feature information of peptide, and produce more reliable prediction results. Therefore, this paper intends to use sequence information, physicochemical properties, and context information derived from sequences to construct a multivariate feature space, enrich the embedded feature representation of peptides, and help the model to better predict peptide detectability.
In classification problems, heterogeneity among information is an important data characteristic of information itself [25]. Fully considering the differences between different information can help the model fully explore the characteristics of each information representation, and learn the high-quality features of specific information. The sequence information, physicochemical properties and context information describe the peptide from different perspectives. After encoding the information, the data dimension and embedding vector represented by each information are completely different, and there is great heterogeneity among the three types of information. Gao et al. [26] collected the physicochemical properties and peptide digestibility, and linearly linear fuse these information to obtain 588-dimensional vectors, used RF to predict peptide detectability. Yu et al. [27] considered the physicochemical properties and sequence information, used same CapsNet network to learn the features of the two types of information, respectively, and thus achieved the prediction of peptide detectability. These methods considered the information of different types of peptides and effectively predicted peptide detectability. However, when linear fusion or the same network is used to process different categories of information, the heterogeneity between different categories of information is not considered, which leads to the captured features losing the unique properties of the corresponding category information [28,29] and then affecting the prediction performance of the model. Ensemble deep learning strategy “specialized” a single kind of specific information [30]. Different deep learning technologies are adopted to learn the representation of specific information, which can capture the high-quality features of specific information, solve the heterogeneity problem among different information, and finally ensemble the classification results of multiple independent deep learning models. It has good fault tolerance and reliability. At present, ensemble deep learning has been successfully applied in biological sequence studies [31,32,33,34], genome analysis [35], medication adherence [36] and other fields, providing strong support for the prediction of peptide detectability.
Based on this, we propose PD-BertEDL, an ensemble deep learning method to predict peptide detectability. We introduced the dynamic bidirectional word embedding model BERT to capture the context information of peptides, and combined the context information with the sequence information and physicochemical information of peptides to construct multivariate feature space of peptides. Aiming at three types of different information of peptides, the ensemble deep learning strategy was introduced, and different deep learning methods were selected to capture the high-quality features of specific information: convolutional neural network (CNN) and bidirectional long short-term memory network (BiLSTM) were used to extract the local and global features of sequence information; CNN and BiLSTM were used to study the physicochemical characteristics of amino acids; and BiLSTM was used to learn the context characteristics of the peptides and to describe the global information of the sequence better. Then, the average fusion strategy was utilized to integrate the model prediction results based on the three types of features as the final prediction result to achieve the prediction of peptide detectability.

2. Results and Discussion

2.1. Performance of Different Encoding Schemes Based on Context Information

In order to verify whether the BERT vector adopted can capture the context information of peptide more effectively, we compared BERT with other two methods: word2vec and transformer. Among them, word2vec used CBOW algorithm, transformer used multi-head self-attention mechanism (head nums = 16), BERT used BERT-mini (which had 11.3 million parameters with 4 transformer layers and 256 hidden embedding sizes). For the three methods, Dense layer was used for classification after coding, and Sn, Sp, ACC, MCC, AUC, and AUPR values were calculated. The results are shown in Table 1.
From Table 1, compared with transformer and word2vec, the model achieves the highest value in all indicators when BERT is used to encode context information. Specifically, the ACC, MCC, AUC, and AUPR values for the BERT method are 68.95%, 37.89%, 74.99%, and 70.54%, respectively, improving by 9.48%, 18.93%, 11.50%, and 9.35% over the word2vec method, respectively. The word2vec method takes several words around the target word in the form of sliding window as samples for training. After training, the word vector of the corresponding word is obtained by looking up table [10]. No matter the amino acid context in the sequence, the word vector in the table is fixed. Therefore, the word2vec method does not take the context information of the sequence into account. In the pre-training process, BERT can calculate the association between words at each position and other words in the sequence, which can fully consider the relationship between amino acids and amino acids in the peptide sequence and can capture the context information of the peptide more effectively. Compared with transformer, the ACC, MCC, AUC and AUPR values of the BERT method are increased by 1.88%, 3.72%, 2.23%, and 2.55%, respectively. Since the information contained in each word after transformer encoding is more biased to its own meaning, the context information is not comprehensive. However, after multi-layer transformer of BERT, each output vector fairly and effectively integrates the information of each word in the entire sequence, which can better describe the global information of the peptide sequence and better predict the performance that can be obtained.

2.2. Ablation Experiment

2.2.1. Feature Combination Ablation Experiment

In this paper, we combined sequence information, physicochemical information, and context information of peptides to characterize peptides. To verify whether the multivariate feature space constructed using three kinds of peptide information could help to predict peptide detectability, we conducted validation on the Homo sapiens training set based on single and multiple features of peptides, and the experimental results are shown in Table 2.
Different columns in Table 2 represent different single features or feature combinations. Among them, Feature A means sequence feature, Feature B means physical and chemical feature, and Feature C means context feature, columns 2–4 represent single feature information. Columns 5–7 indicate that integrating the two kinds of feature using average fusion strategy. Last column indicates that integrate the sequence feature, physical and chemical feature and context feature of peptide using average fusion strategy, which is the feature combination method adopted in this paper. Table 2 records the ACC, MCC, AUC and AUPR values obtained under different feature combinations. As can show in Table 2, when the three features of peptide were integrated, the ACC, MCC, AUC and AUPR values were 82.58%, 66.52%, 88.32%, and 84.01%, respectively, which were better than other conditions. Compared with the case of a single feature, the feature combination method in this paper increases by 1.43%, 1.41%, and 2.16% in ACC, respectively. Compared with the two feature combinations, the feature combinations in this paper are improved by 0.98%, 1.44%, and 1.38% in ACC, respectively. This is because the case of a single feature and two feature combinations only considered part of the feature of the peptide, resulting in the absence of some features, which cannot describe the peptide information more comprehensively. Peptide detectability is influenced by many factors. Considering the sequence information, physicochemical information, and context information derived from sequences meanwhile, the model can learn diversified feature representations, obtain rich feature information of peptides, and significantly improve the prediction performance of the model [37].

2.2.2. Performance of Ensemble Deep Learning

In order to verify whether the ensemble deep learning strategy of PD-BertEDL method can effectively solve the heterogeneity problem among different information and improve the prediction performance of peptide detectability, we designed and implemented three methods: Linear Fusion [26], Same Network [27], and Hybrid [38,39,40], respectively. Among them, Linear Fusion linearly spliced the three kinds of peptide information used in this paper, input the spliced vector into CNN+BiLSTM network to extract features and used softmax to classify. The Same network used the same CNN+BiLSTM network to extract the features of the three kinds of peptide information and used the ensemble strategy based on average fusion to achieve peptide detectability prediction. Hybrid used CNN+BiLSTM, CNN+BiLSTM, BiLSTM to extract the sequence information, physicochemical information and features of context information derived from the sequence, respectively. The three obtained features were fused by matrix concatenation, and BiLSTM was used to learn the fused features. Finally, softmax was used for classification. In order to guarantee the fairness of the experiment, the coding work based on different peptide information and the parameter setting of the model in the above three methods is consistent with the proposed method. On the Homo sapiens dataset, we trained the four models, respectively, and evaluated the models with independent test set. The experimental results are shown in Table 3.
From Table 3, compared with Linear Fusion, Same Network, and Hybrid methods, the proposed method PD-BertEDL achieves the highest ACC, MCC, AUC, and AUPR values. Compared with the Linear Fusion method, PD-BertEDL method can extract features of different information, avoid the interference between different information, and effectively improve the quality of features. Compared with Same Network method, PD-BertEDL method uses different deep learning models to extract the features of three kinds of peptide information and learns the specific features of different information. Meanwhile, the ACC, MCC, AUC, and AUPR values are improved by 0.78%, 1.66%, 0.55% and 0.41%, respectively. Compared with Hybrid method, PD-BertEDL method adopts the ensemble method based on average fusion to integrate models based on three kinds of peptide information, give play to the advantages of each model, solve the constraints of different model training, and improve the overall prediction performance of the model. For Hybrid and PD-BertEDL, the difference is only in the final classification method. The MCC value of PD-BertEDL increases by 2.64% compared with Hybrid, and the MCC value can be used to measure the classification quality of a binary classifier [41]. This indicates that in the prediction of peptide detectability, after using different models to automatically obtain corresponding high-level features from specific information, ensemble classification method is more helpful to improve the prediction performance of peptide detectability than mixed model classification method. In summary, this paper adopts different deep learning methods to extract features of different information and adopts the ensemble strategy based on average fusion to design the ensemble deep learning method architecture reasonably and effectively, which can better predict peptide detectability.

2.3. Performance of Machine Learning and Deep Learning Methods

In this paper, we used sequence information, physicochemical information, and context information derived from sequences as three kinds of information representation of peptides. In order to explore effective learning methods based on these three kinds of information, exploit the unique properties of each information fully, and capture high-quality features, we adopted four classical machine learning algorithms: K-Nearest Neighbor (KNN), Logistic Regression (LR), RF and Gradient Boosting Decision Tree (GBDT) and two deep learning algorithms: CNN and BiLSTM, and the combination model of two algorithms. Training was conducted on the Homo sapiens dataset, respectively, and cross-validation mean results and independent test results were obtained, as shown in Figure 1 (see Tables S3 and S4 and Figures S1–S4 of Supplementary Materials for specific experimental results).
As can be seen from Figure 1, in the 5-fold cross-validation and independent test experiments, the three deep learning methods achieve higher accuracy compared with the traditional machine learning algorithms. For the sequence information encoded of one-hot coding, when CNN+BiLSTM combined model was used to learn features, the ACC value of cross validation reaches 81.47%, which is 1.10% higher than that of CNN model only and 0.06% higher than that of BiLSTM model only. For the physicochemical information of AP3-A coding, the ACC value of cross validation reached 80.58% when CNN+BiLSTM combined model was used to learn features, which was 0.90% and 1.50% higher than that of CNN model and BiLSTM model alone, respectively. This indicates that the combination model of CNN and BiLSTM can simultaneously extract the local information of amino acid residues and the order dependent relationship between different residues and obtain higher quality sequence features and physicochemical characteristics. However, we found that for the context information of peptides, only BiLSTM was used to learn features, which achieved the optimal result; its cross-validated ACC value was 9.87% higher than that of CNN model alone and 0.12% higher than that of CNN+BiLSTM model. This is because BERT uses the above and below of the target word to capture the context information of the sequence and fully reflect the global information of the sequence. While CNN uses the sliding window to form a specific scale matrix to capture features, which will weaken the context information extracted by BERT and even lead to losing part of the information in the process of noise reduction, reducing the prediction performance of the model [42]. BiLSTM can well capture the bidirectional semantic dependence, strengthen the context information extracted by BERT, to improve the prediction accuracy of peptide detectivity.

2.4. Evaluation of PD-BertEDL Prediction Ability

To evaluate the prediction performance of the proposed method PD-BertEDL, we compared the prediction performance of the proposed method PD-BertEDL with other existing methods based on the independent test set. In this part of the experiment, we chose four comparison methods: DNN [22], CapsNet [27], DeepMS [8], and PepFormer [9]. Among them, DNN method only used the single information representation of the peptide to predict peptide detectability. The CapsNet method used peptide sequence information and physicochemical information to describe the peptide, and used the same network to extract features from different information to achieve the prediction of peptide detectability. DeepMS and PepFormer are meaningful attempts of deep learning and NLP technology in predicting peptide detectability: DeepMS regarded peptide sequence as a sentence and used two-layer CNN to extract the context features of peptides encoded by word2vec for further prediction; PepFormer used transformer to capture the context features of peptides and used BiGRU to predict peptide detectability, which opens the idea for the application of deep learning and NLP technology in predicting peptide detectability. In order to ensure the fairness of the experiment, training is carried out on the training set of Homo sapiens, and the model is evaluated on the independent test set. The experimental results are as shown in Table 4.
From Table 4, we know that compared with other predictors, the Sn, ACC, and MCC values of the proposed PD-BertEDL method are 92.38%, 82.58%, and 66.52%, respectively, but the Sp value is relatively low. Sn and Sp are antagonistic, and Sn value represents the percentage of data predicted as positive cases in all positive cases [41]. In protein identification analysis and peptide detection, it is important to know which theory digested peptides can be identified. This means that higher Sn value is more helpful in predicting peptide detectability [26]. Through analysis, it was found that the proposed method PD-BertEDL achieves the highest Sn value, which indicates that our method obtains better prediction performance due to its ability to accurately predict positive samples. In addition, the ACC value of PD-BertEDL is enhanced by 13.88%, 1.27%, 10.74%, and 0.41% compared with DNN, CapsNet, DeepMS, and PepFormer, respectively. Through analysis, the DNN method uses only the physicochemical information of peptides, while PD-BertEDL method uses the sequence information, physicochemical information, and context information of peptides to enrich the characteristic representation of peptides and solve the problem that the information representation is simple due to the use of physicochemical information only. Compared with the CapsNet method, PD-BertEDL adopts different deep learning methods for feature learning of three different kinds of peptide information, which can extract the important features of each information more effectively. At the same time, PD-BertEDL combines with the ensemble strategy to effectively integrate the prediction results of three kinds of feature information and solve the heterogeneity problem among different information and enhances peptide detectability. Compared with the deep learning and NLP methods (DeepMS and PepFormer), PD-BertEDL introduces BERT to capture the context information of peptide more fully, and combines BiLSTM to capture the bidirectional semantic dependence well, strengthens the context information extracted by BERT, and effectively improves the overall prediction accuracy of the model.
Since the ROC curve and PR curve can more intuitively compare the performance of each predictor, the ROC curve and PR curve of PD-BertEDL and other predictors on the independent test set of Homo sapiens were plotted in this paper, as shown in Figure 2.
The AUC and AUPR of PD-BertEDL are 88.32% and 84.01%. Compared with DNN, CapsNet, DeepMS, and PepFormer, the AUC (AUPR) value of the PD-BertEDL method increased by 13.61% (14.18%), 2.65% (2.63%), 2.59% (4.09%), and 1.05% (2.19%), respectively. In conclusion, these results demonstrate that our method retains a better predictive power than the existing methods for the peptide detectability prediction.
In order to prove the stability of PD-BertEDL, we conducted 5-fold cross-validation experiment on the Homo sapiens dataset for DNN, CapsNet, DeepMS, Pepformer, and PD-BertEDL, and recorded the results in Table 5.
It can be seen from Table 5 that the standard deviations of Sn, Sp, ACC, MCC, AUC, and AUPR of PD-BertEDL method are 0.0134, 0.0113, 0.0042, 0.0137, 0.0049, and 0.0051, respectively. Based on the results of 5-fold cross-validation, we calculated the variance of Sn, Sp, ACC, MCC, AUC, and AUPR, which were 2.24 × 10−4, 1.59 × 10−4, 2.17 × 10−5, 2.32 × 10−4, 3.00 × 10−5, and 3.19 × 10−5, respectively. (The variance of indicators for all models in Table S5). The results show that the model can learn the characteristics of peptide sequence well when processing each fold of samples, then effectively realize the prediction of peptide detectability, and it has a certain stability.
In order to further verify the generalization ability of the proposed method PD-BertEDL, the DNN, CapsNet, DeepMS, and PepFormer methods were trained on the Musculus dataset and independently tested and verified. The prediction results are shown in Table 6.
As can be seen from Table 6, on the Musculus dataset, the proposed method PD-BertEDL still achieved the optimal results on a whole. This indicates that PD-BertEDL has good generalization performance and is suitable for data of different species. In addition, we find that the three methods DNN, DeepMS, and PepFormer reach the same conclusion as the proposed method. However, prediction accuracy obtained by these three methods are lower than that of the PD-BertEDL method. For the CapsNet method, the accuracy is significantly lower than other methods. The underlying reason is that the learning ability of this model is weak and it cannot fully learn the unique characteristics of data of different species [9]. Further confirmed, even from the perspective of different species, our method is more effective for predicting peptide detectability but also proves that the method has a higher complexity and a stronger ability to learn; it can learn the unique features of different species data; and it provides a useful reference for the prediction of peptide detectability in other species.

2.5. t-SNE Visualization

In this paper, we combined sequence information, physicochemical information, and context information derived from sequences to characterize peptides for the prediction of peptide detectability. To verify that all three kinds of information contribute to the prediction of peptide detectability, we performed visual verification using t-SNE on the independent test set of Homo sapiens, based on two modes of single information and three combinations of information. The initial coding vectors of all samples and high-level abstract features extracted from models based on different information are projected into a two-dimensional space, and the space is scaled to the interval [−1, 1]. The results are shown in Figure 3.
For each information of peptide, after feature learning by corresponding models, each model generates a better high-level feature discrimination representation. Compared with the single information to distinguish whether a peptide can be detected or not, the multivariate information of the integrated peptide shows better discrimination. Therefore, the proposed model can learn the abstract deep representation of peptides from the sequence information, physicochemical information, and context information, enhance the distinguishing ability of features, and help to predict peptide detectability.

3. Materials and Methods

Prediction of peptide detectability can be abstracted as a binary classification problem. The collected sequence can be classified as: the peptides detected by MS and the non-detected peptides [8]. We propose an ensemble deep learning method to investigate the underlying mechanisms of peptide detectability. The architecture of PD-BertEDL is shown in Figure 4.

3.1. Dataset Collection and Preprocessing

The dataset applied for training and testing originates from the GPMDB database [43], involving mass spectrometry data and detection frequencies for proteins identified by mass spectrometry [27]. From Cheng et al. [9], we obtained data of two species: Homo sapiens and Musculus. In order to avoid the deviation of the model due to the high sequence homology, we used the CD-HIT tool [44] with the threshold of 0.9 to delete sequences containing non-standard amino acids and redundant information. In addition, we calculated the peptide sequence lengths of the two datasets, and the maximum sequence lengths of Homo sapiens and Musculus after CD-HIT were 46 and 63, respectively. For the integrity of the sequence, we set the maximum length to the sequence length L. For sequence fragments containing fewer than L amino acids, we filled them with pseudo amino acids (represented by ‘X’). After that, we randomly selected 20% of them as an independent test set. The remaining peptide sequences were used as a training set. The statistical information of Homo sapiens and Musculus dataset is shown in Table 7.

3.2. Information Encoding

3.2.1. Sequence Information

One-of-21 encoding was used to encode sequence information of the peptide, which is a discrete representation of value 1 at the index corresponding to the amino acid in the peptide and 0 at all other positions [41]. For example, the one-of-21 encoding of a sequence fragment ‘KFVICHLKGK’ is [[001000000000][000001000000]……[000000010000]]. Thus, for a sequence fragment with length L, L × 21-dimensional vector representation would be obtained after one-hot-21 coding.

3.2.2. Physicochemical Information

Gao et al. [26] selected 15 amino acid indices from AAindex [45] as the physicochemical properties of peptides, including 10 structural indices, positive charge, 2 energy indices, hydrophobicity indices and amino acid composition (Table S1 of Supplementary Materials). In this part, the 15 amino acid indices were used to encode the physicochemical properties of the peptide, which is hereby referred to as AP3-A for the convenience of the following description. For motifs of length L, we obtained corresponding values according to the correspondence between amino acids contained in them and Table S2. For example, the corresponding physicochemical properties vector representation of sequence fragment ‘KFVICHLKGK’ is [[4.358,1.7,0.61…1,6.31], [4.663,1,0.8…0, 7.09], ……, [4.224,1.8,0.66…1,4.32]]. Thus, for a sequence fragment with length L, L × 15-dimensional vector representation would be obtained, which can represent physicochemical information.

3.2.3. Sequence-derived Context Information

BERT jointly adjusts the context of the text through the self-attention mechanism in all coding layers to obtain the deep bidirectional representation [46]. In the BERT model, the input vector of each word consists of three embeddings: token embedding, segment embedding and position embedding. Position embedding can represent the absolute position information of each word in a sentence [47]. A visualization of the word embedding structure is shown in Figure 5.
In order to fully capture the context information of the peptide, we regarded the peptide sequence as a sentence, and took k amino acids as a group, which was called “word” [47]. Since the parameter scale of BERT model is up to 100 million, it is very demanding for the experimental environment to use our own corpus to build a word list and retrain the BERT model. In addition, in the study of classification prediction using BERT, some research proved that the optimal results could be obtained when k = 1 was used [15,16]. Therefore, k was set to 1, and the peptide sequence of length L was divided from beginning to end to obtain L “words”. In this paper, the token sequence is an array composed of a single amino acid in the sequence. The token sequence is input into the BERT model to obtain the embedding vector representation of each amino acid. Given a sequence P = ( P 1 , P 2 ,…, P L ) ( P i refers to each amino acid), the tokenizer transforms the sequence P and obtains the mapping ids table of the sequence, where the first CLS is the sentence vector and the last SEP is the spacer. For sequence P, we can obtain token embeddings (1, L, 256), segment embeddings (1, L, 256), and position embeddings (1, L, 256) after three layers of coding representation. A synthetic representation of size (1, L, 256) is obtained using equations (1) and fed into the BERT model.
Embedding = embedding token + embedding segment + embedding position
Finally, we extracted intermediate features using a pre-trained language model (Bert-mini) adopted by Qiao et al. [15], used output vector of the last layer as the context information of sequence P and used it to predict peptide detectability.

3.3. The PD-BertEDL Architecture

3.3.1. Ensemble Deep Learning

Ensemble deep learning strategy combines ensemble learning and deep learning technology, and can “specialized” a single feature set [30], using different deep learning methods can learn the representation of a single feature set. The ensemble deep learning strategy integrates the learning and classification results of multiple deep learning models based on a single feature set, to solve the problem of heterogeneous data combination and realize the multi-faceted abstraction of data [23]. This strategy is currently widely used to predict different biological data, such as anti-tuberculate peptides [33], antimicrobial peptides [34], recognition of non-classical secreted proteins [23], DNA recombination points [35], etc., and has been shown to help improve prediction performance and model generalization.
Considering the heterogeneity among different information, we adopted different methods to extract the features of the three kinds of peptide information, and integrated the prediction results of the model based on the three features, thus learn its unique properties from the specific information and solve the heterogeneity problem among different categories of information. We used four models with different machine learning characteristics (LR, RF, KNN, GBDT [48]) and three deep learning models (CNN, BiLSTM, CNN and BiLSTM combination model) to learn the three kinds of peptide information and predict peptide detectability. Then, we obtained the prediction results of seven models based on each kind of peptide information. Corresponding to the sequence information of peptide, the optimal model (with the highest accuracy) was selected from the seven models according to the prediction accuracy of the model as the prediction model based on the sequence information of the peptide. Similarly, corresponding to the physicochemical information of peptides and the context information derived from sequences, the optimal prediction model based on the physicochemical information of peptides and the context information derived from sequences can be obtained according to the above screening methods. Finally, predicted output probable values of the three optimal prediction models were equal-weighted averaged with average fusion strategy to achieve ensemble prediction. After training and screening, the overall prediction result Y of the ensemble model based on CNN + BiLSTM, CNN + BiLSTM and BiLSTM was obtained, as showed in Equation (2).
Y = f y i = 1 3 i = 1 3 y i
where y i is the predicted probability value based on the i th information of the peptide, respectively. In this way, we effectively make use of the characteristics of each information without being affected by the number and dimension of features encoded by different information, thus enhance the performance and generalization of the model [23].

3.3.2. Convolutional Neural Networks

CNN is used to extract sequence information and local features of physicochemical information of peptides. CNN carries out convolution operations through convolution kernels with shared parameters to extract local spatial information, while different convolution kernels extract different local information, and the combination of these information constitutes spatial feature, which is more efficient than fully connected deep networks. In this part, the basic CNN unit is a one-dimensional convolutional network (1D-CNN), which contains a convolution layer and a relu layer. The unit is added in a nested way to adjust the number of convolutional layers in the network. The convolution layer extracts the hidden features in the original data by summing the dot product between the convolution kernel matrix and the input data matrix. After the convolution operation, the relu activation function is used to remove the negative information and retain the useful information for classification.
Taking motifs of length L as an example, the feature extraction process of 1D-CNN is shown in Equations (3)–(5).
X = f c o n v I
X i , j = F = 1 F r = 1 R W F , r i I F , r j + b i
X C N N = f r e l u X = r e l u X = max 0 , X
Here, I F × L is the coding vector of the input layer, and F 21 , 15 is the coding length of the sequence information and physicochemical information of the peptide. X i , j is the characteristic representation of the i th convolution kernel sliding to the j th amino acid, i 1 , , N f i l t e r ( N f i l t e r is the number of convolution kernels), j 1 , , L (R is the size of the convolution kernel, and S is the sliding step. W N f i l t e r × F × R is the weight matrix, and b is the bias term. X C N N N f i l t e r × L is the output of 1D-CNN. This process is repeated in a nested way to construct multi-layer CNN, so as to capture deeper feature information.

3.3.3. Bi-Directional Long Short-Term Memory Network

Considering the sequential correlation between different amino acids, BiLSTM is used to capture the interdependence information and global information among amino acids in peptide sequences, so as to enhance the information flow in the process of peptide sequence feature learning and to improve the discriminative ability of the network. BiLSTM is composed of LSTM units, and the structure of a single LSTM unit is shown in Figure 6.
After the LSTM unit receives the input data, the first step was to decide which information to keep through the “forgetting gate” and the output of the last moment. The second step was divided into two parts. Firstly, generating new information through the “input gate”; Second, new information was added through the tanh layer to update the current unit state. The third step was to get the output of LSTM unit through “output gate” and tanh layer. The calculation process of LSTM element at time step t   is shown in Equations (6)–(10).
f t = σ W x f x t + W h f h t 1 + b f
i t = σ W x i x t + W h i h t 1 + b i
C t = f t C t 1 + i t tanh W x C x t + W h C h t 1 + b C
o t = σ W x o x t + W h o h t 1 + b o
h t = o t tanh C t
f t ,     i t ,     C t ,     o t and h t represent forgotten gate, input gate, unit state, output gate and hidden state, respectively. x t represents the LSTM unit input at time step t ; W and b represent the weight matrix and the bias term, respectively. Therefore, the LSTM unit adjusts the internal information flow through the gating mechanism and controls the historical information through the “forgetting gate”, to ensure that the network learns the dependence relationship between amino acid residues. BiLSTM network has outputs from two directions, which are connected in series.

3.4. Evaluation Metrics

Several statistical measures were considered to evaluate the performance of the proposed model and other predictors. They were sensitivity (Sn), specificity (Sp), accuracy (ACC), and Matthew’s correlation coefficient (MCC). The definitions are as follows.
S n = T P T P + F N
S p = T N T N + F P  
A C C = T P + T N T P + T N + F P + F N
M C C = T P × T N F P × F N T P + F P × T P + F N × T N + F N × T N + F P
where TP, TN, FP, and FN represent true positives, true negatives, false positives, and false negatives, respectively. In addition, we also used the area under the receiver operating characteristic (ROC) curve (AUC) and the area under the precision recall rate (PR) curve (AUPR) to further access the overall performance of the model.

4. Conclusions

In this paper, we proposed an ensemble deep learning method, PD-BertEDL, to achieve effective prediction of peptide detectability. We used the sequence information, physicochemical information, and context information derived from the sequence of the peptide, and input the three kinds of information into the corresponding deep learning model after encoding. In view of these three different kinds of information, CNN and BiLSTM were used to extract the peptide sequence information and the characteristics of physicochemical information. BiLSTM was used to capture the long-distance dependence information of the peptide sequence and the deep features of the context information of the peptide. Finally, we ensemble model prediction results based on a single feature representation to improve prediction performance and model generalization ability. The results of a k-fold cross-validation experiment and an independent testing experiment suggest that our method has potential to be a useful tool for the prediction of peptide detectability.
Although our method achieves promising performance in peptide detectability prediction, there is still some room for improvement. In the future, we will try to reduce the memory consumption of the model. At the same time, we will consider developing an online platform aimed at providing an effective means of predicting peptide detectability.

Supplementary Materials

The following are available online at: https://www.mdpi.com/article/10.3390/ijms232012385/s1.

Author Contributions

H.W.: methodology, investigation, and writing–review and editing; J.W.: methodology, visualization, and writing–original draft; Z.F.: validation and writing–review and editing; Y.L.: methodology and writing–review and editing; H.Z.: formal analysis and writing–review and editing. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Youth Project of Shanxi Province, grant number 20210302123092.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

References

  1. Tang, H.; Arnold, R.J.; Alves, P.; Xun, Z.; Clemmer, D.E.; Novotny, M.V.; Reilly, J.P.; Radivojac, P. A computational approach toward label-free protein quantification using predicted peptide detectability. Bioinformatics 2006, 22, e481–e488. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Jarnuczak, A.F.; Lee, D.C.-H.; Lawless, C.; Holman, S.W.; Eyers, C.E.; Hubbard, S.J. Analysis of intrinsic peptide detectability via integrated label-free and SRM-based absolute quantitative proteomics. J. Proteome Res. 2016, 15, 2945–2959. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Mallick, P.; Schirle, M.; Chen, S.S.; Flory, M.R.; Lee, H.; Martin, D.; Ranish, J.; Raught, B.; Schmitt, R.; Werner, T.; et al. Computational prediction of proteotypic peptides for quantitative proteomics. Nat. Biotechnol. 2007, 25, 125–131. [Google Scholar] [CrossRef] [PubMed]
  4. Lu, P.; Vogel, C.; Wang, R.; Yao, X.; Marcotte, E.M. Absolute protein expression profiling estimates the relative contributions of transcriptional and translational regulation. Nat. Biotechnol. 2007, 25, 117–124. [Google Scholar] [CrossRef] [PubMed]
  5. Bihan, T.L.; Robinson, M.D.; Stewart, I.I.; Figeys, D. Definition and characterization of a “trypsinosome” from specific peptide characteristics by nano-HPLC-MS/MS and in silico analysis of complex protein mixtures. J. Proteome Res. 2004, 3, 1138–1148. [Google Scholar] [CrossRef] [PubMed]
  6. Qeli, E.; Omasits, U.; Goetze, S.; Stekhoven, D.J.; Frey, J.E.; Basler, K.; Wollscheid, B.; Brunner, E.; Ahrens, C.H. Improved prediction of peptide detectability for targeted proteomics using a rank-based algorithm and organism-specific data. J. Proteomics 2014, 108, 269–283. [Google Scholar] [CrossRef]
  7. Charoenkwan, P.; Nantasenamat, C.; Hasan, M.M.; Manavalan, B.; Shoombuatong, W. BERT4Bitter: A bidirectional encoder representations from transformers (BERT)-based model for improving the prediction of bitter peptides. Bioinformatics 2021, 37, 2556–2562. [Google Scholar] [CrossRef]
  8. Serrano, G.; Guruceaga, E.; Segura, V. DeepMSPeptide: Peptide detectability prediction using deep learning. Bioinformatics 2020, 36, 1279–1280. [Google Scholar] [CrossRef]
  9. Cheng, H.; Rao, B.; Liu, L.; Cui, L.; Xiao, G.; Su, R.; Wei, L. PepFormer: End-to-End Transformer-Based Siamese Network to Predict and Enhance Peptide Detectability Based on Sequence Only. Anal. Chem. 2021, 93, 6481–6490. [Google Scholar] [CrossRef]
  10. Church, K.W. Word2Vec. Nat. Lang. Eng. 2017, 23, 155–162. [Google Scholar] [CrossRef]
  11. Le, N.Q.-K.; Ho, Q.T.; Nguyen, T.T.-D.; Ou, Y.Y. A transformer architecture based on BERT and 2D convolutional neural network to identify DNA enhancers from sequence information. Brief. Bioinform. 2021, 22, bbab005. [Google Scholar] [CrossRef]
  12. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention is all you need. NIPS 2017, 30. [Google Scholar] [CrossRef]
  13. Acheampong, F.A.; Nunoo-Mensah, H.; Chen, W. Transformer models for text-based emotion detection: A review of BERT-based approaches. Artif. Intell. Rev. 2021, 54, 5789–5829. [Google Scholar] [CrossRef]
  14. Cui, Y.; Huang, C. A Chinese Text Classification Method Based on BERT and Convolutional Neural Network. In Proceedings of the 2021 7th International Conference on Systems and Informatics (ICSAI), Chongqing, China, 13–15 November 2021; pp. 1–6. [Google Scholar]
  15. Qiao, Y.; Zhu, X.; Gong, H. BERT-Kcr: Prediction of lysine crotonylation sites by a transfer learning method with pre-trained BERT models. Bioinformatics 2021, 38, 648–654. [Google Scholar] [CrossRef]
  16. Zhang, Y.; Lin, J.; Zhao, L.; Zeng, X.; Liu, X. A novel antibacterial peptide recognition algorithm based on BERT. Brief. Bioinform. 2021, 22, bbab200. [Google Scholar] [CrossRef]
  17. Ji, Y.; Zhou, Z.; Liu, H.; Davuluri, R.V. DNABERT: Pre-trained Bidirectional Encoder Representations from Transformers model for DNA-language in genome. Bioinformatics 2021, 37, 2112–2120. [Google Scholar] [CrossRef]
  18. Zhang, L.; Qin, X.; Liu, M.; Liu, G.; Ren, Y. BERT-m7G: A Transformer Architecture Based on BERT and Stacking Ensemble to Identify RNA N7-Methylguanosine Sites from Sequence Information. Comput. Math. Methods Med. 2021, 2021, 7764764. [Google Scholar] [CrossRef]
  19. Li, Y.F.; Arnold, R.J.; Tang, H.; Radivojac, P. The Importance of Peptide Detectability for Protein Identification, Quantification, and Experiment Design in MS/MS Proteomics. J. Proteome Res. 2010, 9, 6288–6297. [Google Scholar] [CrossRef] [Green Version]
  20. Zohora, F.T.; Rahman, M.Z.; Tran, N.H.; Xin, L.; Shan, B.; Li, M. DeepIso: A deep learning model for peptide feature detection from LC-MS map. Sci. Rep. 2019, 9, 17168. [Google Scholar] [CrossRef] [Green Version]
  21. Guruceaga, E.; Garin-Muga, A.; Prieto, G.; Bejarano, B.; Marcilla, M.; Marín-Vicente, C.; Casal, J.; Vizcaíno, J.A.; Corrales, F.J.; Segura, V. Enhanced missing proteins detection in NCI60 cell lines using an integrative search engine approach. J. Proteome Res. 2017, 16, 4374–4390. [Google Scholar] [CrossRef]
  22. Zimmer, D.; Schneider, K.; Sommer, F.; Schroda, M.; Mühlhaus, T. Artificial intelligence understands peptide observability and assists with absolute protein quantification. Front. Plant Sci. 2018, 9, 1559. [Google Scholar] [CrossRef] [Green Version]
  23. Zhang, Y.; Yu, S.; Xie, R.; Li, J.; Leier, A.; Marquez-Lago, T.T.; Akutsu, T.; Smith, A.I.; Ge, Z.; Wang, J.; et al. PeNGaRoo, a combined gradient boosting and ensemble learning framework for predicting non-classical secreted proteins. Bioinformatics 2020, 36, 704–712. [Google Scholar] [CrossRef]
  24. Xu, H.D.; Shi, S.P.; Wen, P.P.; Qiu, J.D. SuccFind: A novel succinylation sites online prediction tool via enhanced characteristic strategy. Bioinformatics 2015, 31, 3748–3750. [Google Scholar] [CrossRef] [Green Version]
  25. Xiao, H.; Wang, Y. A systematical approach to classification problems with feature space heterogeneity. Kybernetes 2019, 48, 2006–2029. [Google Scholar] [CrossRef]
  26. Gao, Z.; Chang, C.; Yang, J.; Zhu, Y.; Fu, Y. AP3: An Advanced Proteotypic Peptide Predictor for Targeted Proteomics by Incorporating Peptide Digestibility. Anal. Chem. 2019, 91, 8705–8711. [Google Scholar] [CrossRef]
  27. Yu, M.; Duan, Y.; Li, Z.; Zhang, Y. Prediction of Peptide Detectability Based on CapsNet and Convolutional Block Attention Module. Int. J. Mol. Sci. 2021, 22, 12080. [Google Scholar] [CrossRef] [PubMed]
  28. Wang, J.; Yang, B.; Leier, A.; Marquez-Lago, T.T.; Hayashida, M.; Rocker, A.; Zhang, Y.; Akutsu, T.; Chou, K.C.; Strugnell, R.A.; et al. Bastion6: A bioinformatics approach for accurate prediction of type VI secreted effectors. Bioinformatics 2018, 34, 2546–2555. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  29. Wang, J.; Li, J.; Yang, B.; Xie, R.; Marquez-Lago, T.T.; Leier, A.; Hayashida, M.; Akutsu, T.; Zhang, Y.; Chou, K.C.; et al. Bastion3: A two-layer ensemble predictor of type III secreted effectors. Bioinformatics 2019, 35, 2017–2028. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  30. Cao, Y.; Geddes, T.A.; Yang, J.Y.-H.; Yang, P. Ensemble deep learning in bioinformatics. Nat. Mach. Intell. 2020, 2, 500–508. [Google Scholar] [CrossRef]
  31. Ma, Y.; Guo, Z.; Xia, B.; Zhang, Y.; Liu, X.; Yu, Y.; Tang, N.; Tong, X.; Wang, M.; Ye, X.; et al. Identification of antimicrobial peptides from the human gut microbiome using deep learning. Nat. Biotechnol. 2022, 40, 921–931. [Google Scholar] [CrossRef]
  32. Xu, Z.; Luo, M.; Lin, W.; Xue, G.; Wang, P.; Jin, X.; Xu, C.; Zhou, W.; Cai, Y.; Yang, W.; et al. DLpTCR: An ensemble deep learning framework for predicting immunogenic peptide recognized by T cell receptor. Brief. Bioinform. 2021, 22, bbab335. [Google Scholar] [CrossRef]
  33. Akbar, S.; Ahmad, A.; Hayat, M.; UrRehman, A.; Khan, S.; Ali, F. iAtbP-Hyb-EnC: Prediction of antitubercular peptides via heterogeneous feature representation and genetic algorithm based ensemble learning model. Comput. Biol. Med. 2021, 137, 104778. [Google Scholar] [CrossRef]
  34. Zarayeneh, N.; Hanifeloo, Z. Antimicrobial peptide prediction using ensemble learning algorithm. arXiv 2020, arXiv:2005.01714. [Google Scholar]
  35. Liu, B.; Wang, S.; Long, R.; Chou, K.C. iRSpot-EL: Identify recombination spots with an ensemble learning approach. Bioinformatics 2017, 33, 35–41. [Google Scholar] [CrossRef] [Green Version]
  36. Gu, Y.; Zalkikar, A.; Liu, M.; Kelly, L.; Hall, A.; Daly, K.; Ward, T. Predicting medication adherence using ensemble learning and deep learning models with large scale healthcare data. Sci. Rep. 2021, 11, 18961. [Google Scholar] [CrossRef]
  37. Wang, H.; Zhao, H.; Yan, Z.; Zhao, J.; Han, J. MDCAN-Lys: A Model for Predicting Succinylation Sites Based on Multilane Dense Convolutional Attention Network. Biomolecules 2021, 11, 872. [Google Scholar] [CrossRef]
  38. Ning, W.; Xu, H.; Jiang, P.; Cheng, H.; Deng, W.; Guo, Y.; Xue, Y. HybridSucc: A hybrid-learning architecture for general and species-specific succinylation site prediction. Genom. Proteom. Bioinform. 2020, 18, 194–207. [Google Scholar] [CrossRef]
  39. Ahmed, S.; Muhammod, R.; Khan, Z.H.; Adilina, S.; Sharma, A.; Shatabda, S.; Dehzangi, A. ACP-MHCNN: An accurate multi-headed deep-convolutional neural network to predict anticancer peptides. Sci. Rep. 2021, 11, 23676. [Google Scholar] [CrossRef]
  40. Rifaioglu, A.S.; Cetin, A.R.; Cansen, K.D.; Doğan, T.; Martin, M.; Atalay, V. MDeePred: Novel multi-channel protein featurization for deep learning-based binding affinity prediction in drug discovery. Bioinformatics 2021, 37, 693–704. [Google Scholar] [CrossRef]
  41. Wang, H.; Zhao, H.; Zhang, J.; Han, J.; Liu, Z. A parallel model of DenseCNN and ordered-neuron LSTM for generic and species-specific succinylation site prediction. Biotechnol. Bioeng. 2022, 119, 1755–1767. [Google Scholar] [CrossRef]
  42. Li, W.; Gao, S.; Zhou, H.; Huang, Z.; Zhang, K.; Li, W. The automatic text classification method based on bert and feature union. In Proceedings of the 2019 IEEE 25th International Conference on Parallel and Distributed Systems (ICPADS), Tianjin, China, 4–6 December 2019; pp. 774–777. [Google Scholar]
  43. Craig, R.; Cortens, J.P.; Beavis, R.C. Open source system for analyzing, validating, and storing protein identification data. J. Proteome Res. 2004, 3, 1234–1242. [Google Scholar] [CrossRef]
  44. Huang, Y.; Niu, B.; Gao, Y.; Fu, L.; Li, W. CD-HIT Suite: A web server for clustering and comparing biological sequences. Bioinformatics 2010, 26, 680–682. [Google Scholar] [CrossRef]
  45. Kawashima, S.; Pokarowski, P.; Pokarowska, M.; Kolinski, A.; Katayama, T.; Kanehisa, M. AAindex: Amino acid index database, progress report 2008. Nucleic Acids Res. 2007, 36, D202–D205. [Google Scholar] [CrossRef] [Green Version]
  46. Devlin, J.; Chang, M.W.; Lee, K.; Toutanova, K. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv 2018, arXiv:1810.04805. [Google Scholar]
  47. Li, Z.; Fang, J.; Wang, S.; Zhang, L.; Chen, Y.; Pian, C. Adapt-Kcr: A novel deep learning framework for accurate prediction of lysine crotonylation sites based on learning embedding features and attention architecture. Brief. Bioinform. 2022, 23, bbac037. [Google Scholar] [CrossRef]
  48. Ke, G.; Xu, Z.; Zhang, J.; Bian, J.; Liu, T.Y. DeepGBM: A deep learning framework distilled by GBDT for online prediction tasks. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, Anchorage, AK, USA, 4–8 August 2019; pp. 384–394. [Google Scholar]
Figure 1. Performance of different methods on the Homo sapiens dataset: (a) 5-fold cross-validation mean results of 21 methods; (b) Independent test results of 21 methods.
Figure 1. Performance of different methods on the Homo sapiens dataset: (a) 5-fold cross-validation mean results of 21 methods; (b) Independent test results of 21 methods.
Ijms 23 12385 g001
Figure 2. The results of PD-BertEDL and other predictors: (a) AUROC values; (b) AUPR values.
Figure 2. The results of PD-BertEDL and other predictors: (a) AUROC values; (b) AUPR values.
Ijms 23 12385 g002
Figure 3. t-SNE visualization results on the independent test set of Homo sapiens: (a) t-SNE of sequence information; (b) t-SNE of physicochemical information; (c) t-SNE of context information; (d) t-SNE of three information.
Figure 3. t-SNE visualization results on the independent test set of Homo sapiens: (a) t-SNE of sequence information; (b) t-SNE of physicochemical information; (c) t-SNE of context information; (d) t-SNE of three information.
Ijms 23 12385 g003
Figure 4. The architecture of PD-BertEDL.
Figure 4. The architecture of PD-BertEDL.
Ijms 23 12385 g004
Figure 5. BERT coding representation.
Figure 5. BERT coding representation.
Ijms 23 12385 g005
Figure 6. LSTM unit structure.
Figure 6. LSTM unit structure.
Ijms 23 12385 g006
Table 1. Performance of three encoding methods on the independent set of Homo sapiens.
Table 1. Performance of three encoding methods on the independent set of Homo sapiens.
MethodSn (%)Sp (%)ACC (%)MCC (%)AUC (%)AUPR (%)
BERT69.7968.0968.9537.8974.9970.54
transformer68.0866.0967.0734.1772.7667.99
word2vec53.4965.3359.4718.9663.4961.19
Table 2. Performance of single feature or multiple feature ensemble.
Table 2. Performance of single feature or multiple feature ensemble.
Feature A
Feature B
Feature C
ACC(%)81.1581.1780.4281.6081.1481.2082.58
MCC(%)62.6463.9761.1264.7364.4564.3566.52
AUC(%)87.0786.5787.0787.5887.5987.6288.32
AUPR(%)82.1181.1682.6982.7083.0583.0684.01
Table 3. The results of PD-BertEDL and other integration methods.
Table 3. The results of PD-BertEDL and other integration methods.
ModelACC (%)MCC (%)AUC (%)AUPR (%)
Linear fusion81.6165.3888.1283.71
Same network81.8064.8687.7783.60
Hybrid81.0263.8887.1882.26
PD-BertEDL(our)82.5866.5288.3284.01
Table 4. The results of PD-BertEDL and other predictors on the independent test set of Homo sapiens.
Table 4. The results of PD-BertEDL and other predictors on the independent test set of Homo sapiens.
ModelSn (%)Sp (%)ACC (%)MCC (%)
DNN [22]75.5961.9368.7037.86
CapsNet [27]88.3274.4381.3163.34
DeepMS [8]89.0271.8480.3561.72
PepFormer [9]81.5679.6182.1761.17
PD-BertEDL(our)92.3872.9682.5866.52
Table 5. The results of 5-fold cross-validation with different models on Homo sapiens dataset.
Table 5. The results of 5-fold cross-validation with different models on Homo sapiens dataset.
ModelSn (%)Sp (%)ACC (%)MCC (%)AUC (%)AUPR (%)
DNN [22] 73.92 ± 1.7162.10 ± 1.7067.96 ± 0.1436.28 ± 0.3273.77 ± 0.4268.88 ± 0.62
CapsNet [27]89.98 ± 0.7173.02 ± 1.1181.50 ± 0.4863.94 ± 0.9385.97 ± 0.3381.50 ± 0.50
DeepMS [8]90.79 ± 1.0368.86 ± 0.3879.73 ± 0.3861.08 ± 0.7085.25 ± 0.2879.25 ± 0.39
PepFormer [9]82.85 ± 2.1376.20 ± 3.3681.96 ± 0.3259.03 ± 3.6887.25 ± 0.3182.36 ± 0.79
PD-BertEDL (our)90.25 ± 1.3471.76 ± 1.1382.13 ± 0.4265.10 ± 1.3788.35 ± 0.4984.29 ± 0.51
Table 6. The results of PD-BertEDL and other predictors on the independent test set of Musculus.
Table 6. The results of PD-BertEDL and other predictors on the independent test set of Musculus.
ModelSn (%)Sp (%)ACC (%)MCC (%)AUC (%)AUPR (%)
DNN [22]58.3465.2962.0523.6767.3460.39
CapsNet [27]35.3782.3860.4320.2264.6458.88
DeepMS [8]72.0666.3569.0138.3475.5466.62
PepFormer [9]73.2573.7773.8846.9480.7872.20
PD-BertEDL(our)85.2965.5374.7651.4181.2172.99
Table 7. Dataset information of Homo sapiens and Musculus, where the threshold of CD-HIT is 0.9.
Table 7. Dataset information of Homo sapiens and Musculus, where the threshold of CD-HIT is 0.9.
Dataset TypePositive SampleNegative Sample
Homo sapiensTraining22,40422,813
Independent test56015703
MusculusTraining73208356
Independent test18292089
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Wang, H.; Wang, J.; Feng, Z.; Li, Y.; Zhao, H. PD-BertEDL: An Ensemble Deep Learning Method Using BERT and Multivariate Representation to Predict Peptide Detectability. Int. J. Mol. Sci. 2022, 23, 12385. https://doi.org/10.3390/ijms232012385

AMA Style

Wang H, Wang J, Feng Z, Li Y, Zhao H. PD-BertEDL: An Ensemble Deep Learning Method Using BERT and Multivariate Representation to Predict Peptide Detectability. International Journal of Molecular Sciences. 2022; 23(20):12385. https://doi.org/10.3390/ijms232012385

Chicago/Turabian Style

Wang, Huiqing, Juan Wang, Zhipeng Feng, Ying Li, and Hong Zhao. 2022. "PD-BertEDL: An Ensemble Deep Learning Method Using BERT and Multivariate Representation to Predict Peptide Detectability" International Journal of Molecular Sciences 23, no. 20: 12385. https://doi.org/10.3390/ijms232012385

APA Style

Wang, H., Wang, J., Feng, Z., Li, Y., & Zhao, H. (2022). PD-BertEDL: An Ensemble Deep Learning Method Using BERT and Multivariate Representation to Predict Peptide Detectability. International Journal of Molecular Sciences, 23(20), 12385. https://doi.org/10.3390/ijms232012385

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop