Next Article in Journal
Dietary Walnuts Prevented Indomethacin-Induced Gastric Damage via AP-1 Transcribed 15-PGDH, Nrf2-Mediated HO-1, and n-3 PUFA-Derived Resolvin E1
Next Article in Special Issue
Functional Insights into the Sphingolipids C1P, S1P, and SPC in Human Fibroblast-like Synoviocytes by Proteomic Analysis
Previous Article in Journal
miR-9-5p and miR-221-3p Promote Human Mesenchymal Stem Cells to Alleviate Carbon Tetrachloride-Induced Liver Injury by Enhancing Human Mesenchymal Stem Cell Engraftment and Inhibiting Hepatic Stellate Cell Activation
Previous Article in Special Issue
Altered Serum Proteins Suggest Inflammation, Fibrogenesis and Angiogenesis in Adult Patients with a Fontan Circulation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

ProPept-MT: A Multi-Task Learning Model for Peptide Feature Prediction

1
Postgraduate Training Base Alliance, Wenzhou Medical University, Wenzhou 325000, China
2
Wenzhou Institute, University of Chinese Academy of Sciences, Wenzhou 325000, China
3
Department of Physics, and Fujian Provincial Key Laboratory for Soft Functional Materials Research, Xiamen University, Xiamen 361005, China
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
Int. J. Mol. Sci. 2024, 25(13), 7237; https://doi.org/10.3390/ijms25137237
Submission received: 28 May 2024 / Revised: 26 June 2024 / Accepted: 28 June 2024 / Published: 30 June 2024
(This article belongs to the Special Issue Proteomics and Its Applications in Disease 3.0)

Abstract

:
In the realm of quantitative proteomics, data-independent acquisition (DIA) has emerged as a promising approach, offering enhanced reproducibility and quantitative accuracy compared to traditional data-dependent acquisition (DDA) methods. However, the analysis of DIA data is currently hindered by its reliance on project-specific spectral libraries derived from DDA analyses, which not only limits proteome coverage but also proves to be a time-intensive process. To overcome these challenges, we propose ProPept-MT, a novel deep learning-based multi-task prediction model designed to accurately forecast key features such as retention time (RT), ion intensity, and ion mobility (IM). Leveraging advanced techniques such as multi-head attention and BiLSTM for feature extraction, coupled with Nash-MTL for gradient coordination, ProPept-MT demonstrates superior prediction performance. Integrating ion mobility alongside RT, mass-to-charge ratio (m/z), and ion intensity forms 4D proteomics. Then, we outline a comprehensive workflow tailored for 4D DIA proteomics research, integrating the use of 4D in silico libraries predicted by ProPept-MT. Evaluation on a benchmark dataset showcases ProPept-MT’s exceptional predictive capabilities, with impressive results including a 99.9% Pearson correlation coefficient (PCC) for RT prediction, a median dot product (DP) of 96.0% for fragment ion intensity prediction, and a 99.3% PCC for IM prediction on the test set. Notably, ProPept-MT manifests efficacy in predicting both unmodified and phosphorylated peptides, underscoring its potential as a valuable tool for constructing high-quality 4D DIA in silico libraries.

1. Introduction

Mass spectrometry (MS) technology finds extensive application in proteomics research. The amalgamation of liquid chromatography (LC) and MS stands as a potent approach for the separation of complex compounds. However, the traditional proteomics methodologies that rely on DDA often encounter challenges stemming from the pace of MS sequencing and the semi-stochastic sampling of precursor ions, resulting in restricted throughput and diminished reproducibility [1]. DIA represents a progression and enhancement of the DDA mode, operating independently of prior MS scan outcomes to facilitate a relatively unbiased MS/MS analysis of all precursor ions within the defined isolation windows, thereby circumventing data imbalances stemming from randomness [2,3]. Nevertheless, a significant constraint remains in the current DIA proteomics approach, requiring the prior establishment of project-specific, high-quality spectral libraries through DDA analysis of extensively pre-fractionated or repetitively injected samples, leading to substantial time consumption and sample wastage [4,5].
In recent years, artificial intelligence has made considerable strides and found widespread application in the biomedical field, spanning medical image analysis [6,7,8], bioinformatics [9,10], disease diagnosis and treatment [11,12], and proteomics [13,14,15,16,17]. In proteomics research, we have reported several deep learning methods applicable to DIA data analysis, such as Dear-DIAXMBD [13], an untargeted analysis method for directly analyzing DIA data; AttnPep [14], used for rescoring peptide-spectra match scores; and SeFilterDIA [15], designed to automatically identify high-confidence peptides. Deep learning approaches can also be leveraged to generate in silico libraries by predicting the fragment ion intensity, RT, and IM of given peptide sequences [18]. Noteworthy examples include DeepMass:Prism [19], Prosit [20], DeepDIA [21], pDeep [22], DeepPhospho [23], Guan et al. [24], MS2CNN [25], Predfull [26], and Deep4D [27]. Among these, DeepDIA employs a model based on bidirectional long short-term memory networks (BiLSTM) [28], encoding each amino acid into a 20-dimensional one-hot vector representing the 20 amino acids. Deep4D utilizes a deep learning model based on the self-attention [29] module, encoding each amino acid into a 23-dimensional one-hot vector, with an initial 20 dimensions representing the 20 amino acids and the subsequent three dimensions representing acetylation, oxidation, and phosphorylation modifications. Another example, DeepPhospho, also harnesses a model based on the self-attention module and exhibits outstanding performance in predicting phosphorylated peptide RT.
The combination of ion mobility spectrometry with mass spectrometry holds promise for enhancing sensitivity and simplifying spectra [30,31,32], particularly demonstrating unique advantages in the analysis of proteins, peptides, and complex compound isomers, thus propelling proteomics into a new era of 4D analysis. Calculating the inherent collisional cross-section (CCS) values of peptides based on experimentally derived ion mobility can significantly enhance the reliability of their identification [33]. Additionally, 4D DIA-based proteomics has shown higher precursor ion sampling rates and increased specificity in precursor identification [34]. Despite the notable progress achieved by deep learning methods in constructing in silico libraries, the construction of high-quality in silico libraries for 4D DIA proteomics and phosphoproteomics remains a challenge.
Here, we propose ProPept-MT, a new deep learning-based multi-task learning model, designed for the precise prediction of RT, ion intensity, and IM of both unmodified and phosphorylated peptides to construct 4D DIA in silico libraries. Through ProPept-MT, a workflow for 4D DIA proteomics analysis has been developed, based on the predicted multidimensional in silico libraries. ProPept-MT features a hybrid network architecture that merges self-attention modules and BiLSTM modules, bolstered by Nash-MTL [35] for gradient aggregation, thereby ensuring coordinated parameter updates across all tasks. Prominently, ProPept-MT indicates superior prediction performance with fewer training parameters compared to existing deep learning-based single-task prediction models.

2. Results

2.1. Development of Model Structure

ProPept-MT is a cutting-edge multi-task deep learning architecture designed to predict multiple peptide features simultaneously. We concatenated the multi-head attention module with the BiLSTM module to fully leverage their advantages in capturing global sequence information. The multi-head attention module serves as the core structure, establishing direct associations between different positions, and focusing on the interactions among individual amino acids within the sequence to enhance feature extraction capabilities. Its robust sequence modeling ability allows it to synchronously process information from various parts of the sequence, capture long-range dependencies, and improve computational efficiency through parallel processing across multiple attention heads, thereby reducing the likelihood of information loss. The BiLSTM, as an auxiliary structure, further augments the model’s performance by capturing longer-term temporal features. It learns from potential information overlooked by preceding subnetworks, generating rich and comprehensive context-aware representations through bidirectional sequence processing. This enhances the accuracy of predictions and robustness to different types of protein sequences.
The hybrid design demonstrates superior performance in deep learning methodologies. Compared to traditional approaches, this innovative combination can analyze the proteomics mass spectrometry data more comprehensively and in detail, which provides more reliable and precise data support for scientific research. Additionally, a series of single-task (ST) assessments were performed on a designated dataset using identical model specifications and hyperparameters, denoted as ProPept-ST, to enable comparative analysis with the multi-task model. The conclusive experimental findings manifest the exceptional performance of ProPept-MT over ProPept-ST.

2.2. Performance of ProPept-ST in Predicting Retention Time

Liquid chromatography is integral to mass spectrometry analysis in bottom-up proteomics [36], with RT playing a crucial role in DIA proteomics. To assess ProPept-MT’s performance, we initially validated ProPept-ST’s capacity for predicting RT. We compared ProPept-ST with existing single-task models for RT prediction, which can be applied to unmodified or phosphorylated peptides, thereby illustrating the sturdiness of its model architecture.
The study commenced with an examination of the predictive efficacy of RT for unmodified peptides, juxtaposing the performance of ProPept-ST against three contemporary deep learning models: Deep4D, DeepDIA, and DeepLC. DeepLC incorporates a deep convolutional neural network (CNN) architecture with an encoding approach grounded in atom composition. The evaluation hinged on 15 unmodified peptide datasets as delineated in the DeepLC study, where peptides were characterized by experimental RT or indexed RT (iRT) [37]. These datasets spanned three distinct LC modalities: reversed-phase LC (RPLC), hydrophilic interaction LC (HILIC), and strong cation exchange chromatography (SCX). ProPept-ST underwent preliminary pre-training on the SWATH library dataset, followed by fine-tuning on the remaining 14 datasets.
Across all 15 datasets, ProPept-ST consistently outperformed extant models in terms of mean absolute error (MAE) (Figure 1A,B) and Δ t 95 % (Supplementary Figure S1). Importantly, the HeLa HF dataset, constructed using 15-min short gradients, posed a potential challenge due to reduced resolution and peak capacity, potentially affecting the predictive accuracy of apex peptide RT [38]. Despite these challenges, ProPept-ST exhibited superior performance compared to the leading DeepLC model on the HeLa HF dataset, achieving MAE values of 0.27 vs. 0.31 and Δ t 95 % values of 1.46 vs. 1.62. In the case of the SWATH library test set, the predicted iRT values demonstrated high precision, with a PCC of 0.997 (Figure 1C).
Furthermore, we subjected ProPept-ST to retraining on the 14 datasets excluding the SWATH library. Figure 1D and Figure S2 indicate that, among the 14 test datasets, the proportion of cases where the fine-tuned median absolute error (MedAE) exceeds that of the non-fine-tuned results is 71.4%, with only 0.07% of cases showing a decrease in performance. This highlights the utility of fine-tuning in enhancing model adaptability to diverse LC types and gradients, thereby underscoring its considerable benefits in RT prediction. Collectively, these findings showcase the superior predictive performance of ProPept-ST over other established models for unmodified peptide RT prediction.
To evaluate ProPept-ST’s performance in predicting RT for phosphopeptides, we conducted a comparative analysis against DeepPhospho and Deep4D, utilizing three distinct phosphopeptide RT datasets: RPE1 DDA, RPE1 DIA, and U2OS DDA, following the methodology described in the DeepPhospho study. ProPept-ST undergoes pre-training on the RPE1 DDA dataset, followed by fine-tuning on the RPE1 DIA and U2OS DIA datasets. ProPept-ST exhibited superior performance in terms of MedAE across all three datasets, as depicted in Figure 2A. Specifically, on the RPE1 DDA test set, ProPept-ST achieved a MedAE of 1.57, surpassing Deep4D and DeepPhospho, which scored 1.62 and 1.74, respectively. Moreover, on the U2OS_DDA test set, the predicted iRT values closely matched the experimentally observed iRT values, presenting high precision with a PCC of 0.997 (Figure 2B). These findings highlight ProPept-ST’s remarkable capability in accurately predicting phosphopeptide RT.

2.3. Ablation Studies

To elucidate the excellent design of our model, we executed a comparative analysis contrasting ProPept-ST with six alternative models (Figure 2C,D). These models predominantly feature combinations of self-attention modules, LSTM networks, and CNN modules. We meticulously designed and tested various parameter combinations, with Table 1 showcasing the main combinations. Each module’s layer count was carefully adjusted to ensure optimal performance. In the final stages of the model, we integrated either attention modules or multi-layer perceptron (MLP) layers to further enhance performance. Through this detailed tuning and optimization, we aimed to achieve the highest possible effectiveness in all aspects. Performance evaluation was carried out on the H4 DDAp’s RT dataset, with reported MedAE values. The finding reveals that in our model architecture, integrating attention modules at the end supersedes the direct use of fully connected layers, yielding MedAE values of 0.715 and 0.727 (Figure 2C), respectively. Interestingly, despite possessing the largest parameter count among these models, the model utilizing solely the self-attention modules exhibited the poorest performance. Noticeably, ProPept-ST showed outstanding performance with the fewest parameters (Figure 2D).

2.4. Performance of ProPept-MT on Benchmark Datasets

Based on our research, it is argued that fine-tuning a pre-trained model generally yields superior performance compared to retraining from scratch. As such, we initially pre-trained on the H5 DDAp dataset and subsequently fine-tuned on eight other specific datasets. Simultaneously, ProPept-ST trained individually for each task on the benchmark datasets is regarded as the evaluation baseline.
For the RT prediction task, ProPept-MT surpasses previously reported models on five datasets in terms of MedAE (Figure 3A). Specifically, on the H1 DDA test set, ProPept-MT achieves a MedAE of 0.598, surpassing the performance of both ProPept-ST and DeepPhospho, which achieve MedAEs of 0.643 and 1.157, respectively. In all benchmark datasets, all values are presented in minutes within the original RT dimension. Additionally, the PCC value on the H2 DIA test set is exceptionally high, reaching 0.999 (Figure 3B and Figure S3). Furthermore, in accordance with the primary indicator highlighted in the DeepPhospho study for RT prediction, we compared Δ t 95 % values (Supplementary Figure S4A). ProPept-MT outperforms both ProPept-ST and DeepPhospho on nine datasets, demonstrating performance on the H1 DDA test set of 4.97 compared to 5.20 and 6.59, respectively. Strikingly, for DeepPhospho, five Transformer encoder layers (4–8) of varying sizes were trained and integrated for testing.
For the IM prediction task, Table 2 showcases that across five datasets, ProPept-MT’s PCC values exceed those of ProPept-ST, achieving a performance of 0.992 compared to 0.981 on the M1 DDAp test set. Moreover, Figure 3C portrays the distribution of IM absolute errors between ProPept-MT and ProPept-ST. The high accuracy of IM prediction, with a PCC value of 0.993, is further revealed on the H2 DIA test set (Figure 3D and Figure S5).
For the fragment ion intensity prediction task, ProPept-MT outperforms ProPept-ST and DeepPhospho on nine datasets in terms of median PCC, as shown in Table 2, achieving a performance of 0.941 vs. 0.927 vs. 0.918 on the M2 DDAp test set. Of note is ProPept-MT’s performance on the H1 DDA and H5 DDAp test sets, where it achieves PCC values exceeding 0.75 for 93.54% and 86.32% of peptides, respectively, with median PCC values of 0.97 and 0.94 (Figure 4A). Mirror representations for specific peptides evidenced robust concordance between our prognostication and authentic measurement, with PCC values of 0.985 and 0.979, respectively (Figure 4B). In comparison to existing models, our multi-task model also exhibits improved overall consistency between experimental and predicted fragment ion intensities for the test set (Figure 4C). For the H4 DDAp dataset, ProPept-MT achieves a median PCC of 0.945, median spectral angle (SA) of 0.835, and median dot product (DP) of 0.967. Similarly, for the H6 DDAp dataset, ProPept-MT achieves a median PCC of 0.940, median SA of 0.824, and median DP of 0.962. Additionally, ProPept-MT outperforms DeepPhospho with respect to median SA on six datasets (Supplementary Figure S4B), which serves as the primary indicator for fragment ion intensity prediction in the DeepPhospho study.
Figure 3E presents the loss of ProPept-MT on both the training and validation sets of the H5 DDAp dataset, elucidating the model’s robust training and lack of overfitting. In Figure 3F, we display the fine-tuning loss of ProPept-MT on the H7 DDAp training set, comparing it with the loss of the ProPept-ST model trained separately for each of the three tasks. These results indicate that the loss for each task can be rapidly minimized by ProPept-MT during a brief training period. Detailed performance metrics are available in Table 2. Our findings disclose that, compared to the reported models, ProPept-MT can achieve optimal performance with the fewest parameters, underscoring its superior ability to predict peptide features.

2.5. Performance Comparison between ProPept-MT and Other Models

To assess the accuracy and fairness of the ProPept-MT model, the identical datasets employed for training, validation, and testing in the comparative models are utilized. Two deep learning models, DeepDIA and DeepPhospho, which are proficient in predicting features of unmodified peptides and phosphorylated peptides, will be compared. Although both models employ similar or slightly enhanced network structures for predicting two or more peptide features, their training methodologies involve training each task separately.
The data processing approach of comparative models is being followed, with the corresponding datasets being obtained for each model and the results being reported according to the evaluation metrics of the respective tasks. For instance, the capability of DeepDIA to predict RT and fragment ion intensity will be utilized. Peptides with lengths less than seven or greater than fifty, or those containing variable modifications, are filtered out. For the task of predicting fragment ion intensity, the selection is further refined to include only sub-ions with intensity values greater than zero and charge states of 1+ or 2+. Additionally, peptides where the number of sub-ions contained in each parent ion is greater than or equal to six are selected. Subsequently, peptides with precursor charges of 2+ and 3+ are segregated for DeepDIA training. In contrast, they will be treated as two separate tasks for joint training by ProPept-MT. One-third of the dataset is allocated for testing, while the remaining two-thirds are further divided into two-thirds for training and one-third for validation.
Similarly, DeepPhospho, like DeepDIA, can predict RT and fragment ion intensity but extends its support to phosphopeptides. The dataset used by DeepPhospho mirrors benchmark datasets, divided into training, validation, and testing sets in an 8:1:1 ratio. Consequently, the comparison results between ProPept-MT and DeepPhospho can be found in Section 2.4. Notably, in comparison with DeepDIA, we continue to fine-tune the filtered dataset using the parameters pre-trained by ProPept-MT on the benchmark dataset H5 DDAp. Conversely, DeepPhospho and DeepDIA undergo retraining on each dataset using their default parameters.
For the fragment ion intensity prediction task, a nuanced comparison with the DeepDIA model reveals 12 distinctive combinations of fragment ion types. These combinations are characterized by varying factors, including the charge states of sub-ions (1+ or 2+), the presence or absence of b/y ions, and the potential for two neutral losses (NH3 or H2O). Similarly, for the DeepPhospho model, eight unique combinations of fragment ion types are unveiled by our analysis. These combinations stem from key factors such as the charge states of sub-ions (1+ or 2+), the presence or absence of b/y ions, and the potential for the loss of H3PO4. Furthermore, our methodology involves the deliberate exclusion of loss items from the phosphate salt component that are deemed implausible. This rigorous approach includes a filtering mechanism to exclude the intensity predictions of these ions, ensuring the robustness and accuracy of our findings.
To ensure consistent comparisons, datasets were initially filtered to exclude those with insufficient data volume, which could potentially affect the performance of single-task training in DeepDIA. For result evaluation, the primary metrics outlined in the DeepDIA study are followed, with PCC and median DP being utilized as key evaluation metrics for RT and ion intensity, respectively. Concerning RT prediction, ProPept-MT outperforms both ProPept-ST and DeepDIA in PCC on five datasets, achieving 0.987 compared to 0.971 and 0.958 on the H6 DDA test set (Table 3). Figure 5A visualizes the distribution of absolute errors for RT prediction across six datasets, showing that ProPept-MT’s MedAE consistently exceeds that of DeepDIA and ProPept-ST on five datasets.
Furthermore, ProPept-MT showcases exceptional predictive performance in IM prediction, exhibiting a superior PCC compared to ProPept-ST across five datasets, achieving 0.971 versus 0.963 on the H6 DDAp test set (Table 3). In addition, ProPept-MT consistently outperforms various alternative baselines in IM prediction (Figure 5B). Moreover, in terms of median DP, ProPept-MT consistently surpasses the models reported in previous studies (Figure 5C). During fine-tuning, ProPept-MT takes advantage of fragment ions with H3PO4 loss from the pre-training set and fragment ions with two types of neutral losses (NH3 or H2O) in the fine-tuning set, achieving better performance than single-task approaches and highlighting the model’s generalization capability. Figure 5D portrays the distribution of DP values for each peptide in the test sets of the H1 DDA dataset, stratified by precursor charges of 2+ and 3+. For peptides with a precursor charge of 2+, the percentage surpassing a DP value of 0.75 is 98.53%, with a median DP of 0.978. Subsequently, for peptides with a precursor charge of 3+, the percentage is 96.35%, with a median DP of 0.959. For a detailed analysis of performance metrics, please refer to Table 3, where ProPept-MT’s performance across six datasets exceeds that of the reported models, underscoring its superior capability in predicting unmodified peptide features.

3. Materials and Methods

3.1. Dataset Collection and Pre-Processing

Constructing an effective benchmark dataset stands as a critical endeavor for training and assessing deep learning models. Curating multiple recently released raw mass spectrometry datasets from esteemed repositories, including ProteomeXchange [39], PRIDE [40], iProX [41], and jPOST [42], was initially undertaken. These MS data, acquired using timsTOF Pro or timsTOF Pro 2 mass spectrometers, provided crucial ion mobility information. Following that, two software packages, MaxQuant (version 2.4.8.0) [43] and DIA-NN (version 1.8.1) [44], were deployed to analyze MS data, thereby resulting in the final peptide identification outcomes. This meticulous process led to the compilation and assembly of nine benchmark datasets, comprising a total of 353,052 entries (Table 4). These datasets encompass samples from both humans and mice, featuring a range of variable modifications, such as phosphorylation at serine, threonine, and tyrosine sites, oxidation of methionine, and N-terminal acetyl modification. Moreover, carbamidomethylation of cysteine served as a fixed modification.
Table 4. Dataset Structure.
Table 4. Dataset Structure.
Data NameSpeciesInstrumentPeptidesIdentifier
H1_DDAHumantimsTOF Pro64,358PXD041421 [45]
H2_DIAHumantimsTOF Pro125,360PXD041391 [45]
H3_DIApHumantimsTOF Pro42,351PXD034709 [46]
H4_DDApHumantimsTOF Pro31,599PXD034709
H5_DDApHumantimsTOF Pro42,677PXD027834 [47]
H6_DDApHumantimsTOF Pro16,784PXD042842 [48]
H7_DDApHumantimsTOF Pro 29495PXD043026 [49]
M1_DDApMousetimsTOF Pro12,132PXD028051 [50]
M2_DDApMousetimsTOF Pro 28296PXD043026
Benchmark datasets exclusively comprise the 20 common amino acids. In the identification results of MaxQuant and DIA-NN, peptide data with q values exceeding 0.01 were filtered out. For the output files of MaxQuant, msms.txt and evidence.txt were specifically selected, and peptides with phosphorylation modification site probabilities below 0.75 were excluded to ensure dataset quality. Furthermore, given that these data originate from multiple Liquid Chromatography-Tandem Mass Spectrometry (LC-MS/MS) runs, the highest-scoring data point from the same peptide was selected for the fragment ion intensity prediction task, while fragment ions with intensities less than or equal to zero were excluded. The median of the corresponding target values was utilized for the RT and IM prediction tasks.

3.2. The Model Architecture of ProPept-MT

Figure 6A depicts the workflow of ProPept-MT. ProPept-MT is a blended network structure that adopts multi-task deep learning to map peptide sequences into high-dimensional vectors. This intricate process involves self-attention modules, nonlinear transformations, and BiLSTM networks integrated into four main modules: an input layer, an embedding layer, a sequence modeling layer, and an output layer (Figure 6B). The embedding layer encodes both the input amino acid sequence and the precursor charge into feature vectors. Subsequently, the sequence modeling layer learns representations of peptide features, culminating in the generation of prediction values by the output layer. This integration enables the model to focus on inter-amino-acid correlations, capture long-term temporal features and latent information, and generate richer contextual information representations.
Each submodule is described as follows:
Input layer. A peptide consists of an amino acid sequence, with the precursor charge represented as a scalar. The 20 common amino acids are denoted in uppercase letters, such as “G” for glycine and “A” for alanine. If the N-terminal of the peptide contains an acetylation modification, “a” is prepended to the sequence; otherwise, “_” is prepended. Variable modifications in the sequence are indicated by “s,” “t,” and “y” for phosphorylation modifications at serine, threonine, and tyrosine sites, respectively, and “m” for methionine oxidation. “$” is appended at the end of the sequence to aid the model in determining when to cease processing the sequence. The maximum peptide length is set to 52, with any portion of the sequence shorter than 52 padded with “#”.
Embedding layer. For the RT prediction task, each amino acid is directly embedded into a 256-dimensional tensor. Conversely, for the IM and fragment ion intensity prediction tasks, each amino acid is first embedded into a 192-dimensional tensor and each precursor charge into a 64-dimensional tensor, which are then concatenated into a 256-dimensional tensor. To incorporate the positional information of amino acids, standard sine and cosine functions are used as positional encoding [29], resulting in a 54 × 256 tensor.
Sequence modeling layer. Serving as the backbone of ProPept-MT, this layer comprises a series of Transformer encoders and BiLSTM subnetworks. The Transformer encoder subnetwork enlists more efficient self-attention modules to capture correlations between amino acids at different positions in the peptide sequence. It consists of six stacked encoder layers, each containing a multi-head attention module and a fully connected feed-forward network, with residual connections and layer normalization. The number of attention heads and hidden layer dimensions of the feed-forward network are eight and 1024, respectively. The goal of this subnetwork is to extract the initial representation of the peptide and feed it to the next subnetwork. The BiLSTM subnetwork consists of a single bidirectional LSTM layer with 512 hidden dimensions and its goal is to capture longer distance dependencies more effectively. Furthermore, as a shared layer, the sequence modeling layer shares its learning parameters across tasks. The attention module is described as:
A t t e n t i o n ( Q , K , V ) = S o f t m a x ( Q K T d k ) V
where Q , K , V is derived from the dot product between the input matrix and three parameter matrices. The operation Q K T generates a similarity matrix between each amino acid’s position and other amino acids’ positions. Subsequently, each element in the matrix is divided by a scalar d k , followed by the application of the softmax function to generate probabilities. Finally, the result is multiplied by V to obtain the context vector representation for each amino acid. d k represents the size of the hidden layer.
Output layer. A linear layer is used to project the features of each amino acid position into an n-dimensional vector, which acts as the output for predicting fragment ion intensities. Here, n represents the number of fragment ion types to be predicted. For the RT and IM prediction tasks, the hidden layer outputs of BiLSTM are used for generating instance-specific weights for sequence features, and a weighted averaging approach is exercised to produce the final RT and IM predictions.

3.3. Loss Function

Two distinct loss functions were exploited to minimize training errors, the mean squared error (MSE) loss function for predicting fragment ion intensity and the L1 norm for predicting RT and IM. These functions are expressed as follows:
MSE = 1 n i = 1 n ( y i y i ^ ) 2
L 1   Loss = 1 n i = 1 n | y i y i ^ |
where n represents the number of training samples, y i is the experimental value, and y i ^ is the predicted value.

3.4. Model Training

The experiment was conducted using Python 3.9 and implemented within the Torch deep learning framework (version 1.10.0) [51] (https://pytorch.org/). Applying multi-task learning during model training offers a potential avenue for reducing computational costs, albeit accompanied by the challenge of potential conflicts arising in the gradients of distinct tasks. To tackle this issue, Nash-MTL, as delineated in Algorithm 1, approaches the gradient aggregation step as a bargaining game [35]. This methodology facilitates task negotiation, aiming to achieve consensus on the direction of updating shared parameters, thus effectively alleviating this challenge.
Algorithm 1. Nash-MTL
Input:  θ 0 - initial parameter vector, { l i } i = 1 K –differentiable loss functions, μ –learning rate
Output:  θ T
for t = 1,…, T do
 Compute task gradients g i t = θ ( t 1 ) l i
 Set G ( t ) the matrix with columns g i ( t )
 Solve for α :   ( G t ) T ( G t ) α = 1 / α to obtain α t
 Update the parameters θ ( t ) = θ ( t ) μ G ( t ) α ( t )
end for
return  θ T
For the multi-task learning paradigm involving parameters θ , this methodology postulates the existence of a sphere B centered at the origin with a radius . The objective is to locate the update vector Δ θ within this defined sphere. This scenario is framed as a bargaining problem, where the center of the sphere represents the point of disagreement, while B signifies the set of agreements. The utility function for each participant is defined as u i ( Δ θ ) = g i T Δ θ , where g i signifies the gradient vector of task i loss at θ .
A unified training approach was implemented, wherein each training iteration sequentially addresses distinct tasks: fragment ion intensity, RT, and IM. For the fragment ion intensity prediction task, should the need arise to train peptides of varying precursor charges separately, the sequence follows 3+, 2+, RT, and IM. The Adam gradient descent algorithm was applied uniformly across all tasks, characterized by a batch size of 128, beta1 of 0.9, beta2 of 0.999, epsilon of 1e-8, and a learning rate of 1e-4. Moreover, extensive exploration of hyperparameters was conducted, accompanied by model simplification. This endeavor facilitated ProPept-MT in capturing intricate features among amino acid sequences, thereby enhancing prediction precision.

3.5. Evaluation Metrics

For the fragment ion intensity prediction task, the median PCC was selected as the ultimate evaluation metric. In addition, to promote comparisons with other established models, we adhered to their evaluation criteria, incorporating normalized SA and DP as two supplementary metrics, and reported their respective medians. The definition of SA is as follows:
S A = 1 2 arccos ( y y ^ ) Π
where y ^ and y are the predicted and experimental vectors, respectively, with L2 norm equal to 1.
For the RT prediction task, the MedAE served as the primary evaluation metric, complemented by the coefficient of determination ( R 2 ), inter quartile range (IQR), PCC, and Δ t 95 % for comparative analysis across models. Here, Δ t 95 % denotes the minimum time window accommodating 95% of peptides, reflecting the disparity between experimentally observed and predicted RT. Regarding IM prediction, we focused on R 2 , PCC, and Δ t 95 % as key metrics, with PCC assuming the role of the principal evaluation criterion. The definitions of MedAE and Δ t 95 % are outlined as follows:
M e d A E = m e d i a n ( | y 1 y 1 ^ | , , | y i y i ^ | )
Δ t 95 % = 2 | y y ^ | 95 %
where n represents the number of training samples, y i is the actual value, and y i ^ is the predicted value. The subscript 95% indicates that the deviation covers 95%.
In the context of ProPept-MT training, where multiple evaluation metrics are relevant to each task, a specific formula is used to gauge model performance, quantifying the extent of performance enhancement between successive training epochs. The formula is as follows:
Δ p = 100 % × 1 T t = 1 T 1 M t m = 1 M t ( 1 ) w t , m ( B t , m N t , m ) N t , m
where T represents the number of tasks; W represents the number of metrics; w t , m represents the optimization direction of the m metric of the t task, with a binary value where 0 indicates that the smaller the metric, the better, and 1 indicates that the larger the metric, the better; B represents the list of metrics for the first training epoch; N represents the list of metrics for the current training epoch.

4. Discussion

In this study, we introduced ProPept-MT, a new multi-task deep learning model designed to enhance the accurate prediction of peptide features and expedite 4D DIA proteomics by precisely predicting the RT, fragment ion intensity, and IM of unmodified peptides or phosphopeptides. First, the evaluation of ProPept-ST’s RT prediction performance was conducted using 15 unmodified RT datasets and three phosphopeptide RT datasets, comparing its performance with the existing advanced RT prediction models. Subsequently, the multi-task prediction performance of ProPept-MT on nine benchmark datasets was assessed and compared with ProPept-ST and the existing advanced models. Evidently, ProPept-MT demonstrates superior prediction performance across all datasets and can predict peptide features such as ion mobility, which other models cannot predict.
In evaluating ProPept-ST, we assessed the ability of single-task models to predict retention time for both unmodified and phosphorylated peptide data. ProPept-ST was compared with Deep4D, DeepDIA, and DeepLC for unmodified peptides, and with Deep4D and DeepPhospho for phosphorylated peptides. To ensure fairness, we used datasets from the respective studies: unmodified peptide data from DeepLC and phosphorylated peptide data from DeepPhospho. This approach was necessary due to the varying nature of the input data supported by these models. Notably, Deep4D, DeepPhospho, and DeepLC support modifications such as methionine oxidation, phosphorylation at STY sites, and N-terminal acetylation, whereas DeepDIA does not.
In comparing ProPept-MT, we evaluated the multi-task model’s ability to predict retention time, ion intensity, and ion mobility for both unmodified and phosphorylated peptides. ProPept-MT was compared with ProPept-ST and DeepPhospho for phosphorylated peptides, and with ProPept-ST and DeepDIA for unmodified peptides. Unlike DeepPhospho and DeepDIA, which train and predict each task separately, ProPept-MT employs a multi-task approach. Due to the lack of ion mobility information in the datasets from comparative model studies, we used our curated datasets for this comparison. Overall, the use of different datasets and comparison modes ensures a fair and comprehensive assessment, meeting the requirements of each model.
Critical to peptide analysis are features such as fragment ion intensity, IM, and RT, which are effectively captured by our model through a nuanced training strategy that utilizes input data and noise from various tasks. This combined training approach helps mitigate the divergent noise inherent in different tasks, thereby improving learning outcomes and enhancing model robustness. In other words, by leveraging the inherent information contained within peptide sequences and precursor charge states, ProPept-MT was effectively trained on fragment ion intensity, RT, and IM within a multi-task learning framework. This advanced approach achieved a form of data augmentation, enabling each task to learn from a richer set of information rather than being confined to the specific data each task individually possessed, while also preventing overfitting. Additionally, ProPept-MT engages a hard parameter-sharing mechanism to embed the data representations of each task into a unified semantic space, followed by the application of a task-specific layer to extract task-specific representations for each task [52,53]. This approach significantly reduces memory consumption and eliminates redundant learning of information in shared layers, ultimately leading to higher inference speed and shorter training times.
Experimental results indicated that ProPept-MT outperformed single-task training, demonstrating robust enhancements in prediction accuracy. This suggested a certain degree of task interrelatedness among the three predicted peptide attributes, allowing each task to update parameters in a similar direction. By simultaneously optimizing multiple related tasks, ProPept-MT leveraged the interdependencies among them, ensuring that the intrinsic correlations within the data were effectively preserved and utilized. Furthermore, the fine-tuning process showcased the model’s flexibility, enabling ProPept-MT to seamlessly adapt to various types of LC and gradient lengths. This adaptability ensured that ProPept-MT could be applied to different experimental setups, enhancing its utility and performance across diverse proteomics research scenarios.
ProPept-MT manifested extensive potential applications. It accurately predicted the ion intensity, RT, and IM of peptides, facilitating the construction of 4D DIA spectral libraries. This capability aided in protein identification and quantitative analysis, enhancing the reliability of data and experimental efficiency. Moreover, we believe that ProPept-MT, through its precise analysis of proteomic mass spectrometry data, can uncover additional peptides and proteins, thereby providing valuable tools for fundamental scientific research, drug development, and disease treatment. We will continue to expand upon this research in the future.
Despite ProPept-MT’s improved prediction performance, negative transfer occurs during training. Analysis of the training loss curve reveals a step-like distribution of the losses for each task, with fragment ion intensity showing the smallest loss and IM the largest. This bias causes the model to prioritize reducing the loss of fragment ion intensity over RT and IM, potentially leading to sustained outstanding performance with regard to fragment ion intensity but the gradual deterioration of RT and IM performance. Furthermore, evaluating the performance improvement of each training epoch reveals challenges in selecting an optimal set of model parameters that perform best on each task. This issue underscores the need for further refinement. As a result, ongoing research focuses on alternative deep learning approaches to address these challenges and improve the prediction of additional peptide features. Future work also involves integrating spatial protein structures into training data to predict specific modification sites.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/ijms25137237/s1.

Author Contributions

Conceptualization, J.S., Q.H. and G.H.; methodology, G.H.; software, G.H.; validation, G.H. and Q.H.; formal analysis, G.H. and Q.H.; investigation, G.H., J.C. and R.Y.; data curation, G.H., J.C. and R.Y.; writing—original draft preparation, G.H.; writing—review and editing, J.S. and Q.H.; visualization, G.H.; supervision, J.S. and Y.C.; project administration, J.S. and Y.C.; funding acquisition, J.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Ministry of Science and Technology of the People’s Republic of China (STI2030-Major Projects2021ZD0201900), and the National Natural Science Foundation of China under Grant 12090052.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Open source after receiving articles such as source programs and data.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Bekker-Jensen, D.B.; Bernhardt, O.M.; Hogrebe, A.; Martinez-Val, A.; Verbeke, L.; Gandhi, T.; Kelstrup, C.D.; Reiter, L.; Olsen, J.V. Rapid and site-specific deep phosphoproteome profiling by data-independent acquisition without the need for spectral libraries. Nat. Commun. 2020, 11, 787. [Google Scholar] [CrossRef] [PubMed]
  2. Gillet, L.C.; Navarro, P.; Tate, S.; Rost, H.; Selevsek, N.; Reiter, L.; Bonner, R.; Aebersold, R. Targeted data extraction of the MS/MS spectra generated by data-independent acquisition: A new concept for consistent and accurate proteome analysis. Mol. Cell. Proteom. 2012, 11, 016717. [Google Scholar] [CrossRef] [PubMed]
  3. Ludwig, C.; Gillet, L.; Rosenberger, G.; Amon, S.; Collins, B.C.; Aebersold, R. Data-independent acquisition-based SWATH-MS for quantitative proteomics: A tutorial. Mol. Syst. Biol. 2018, 14, e8126. [Google Scholar] [CrossRef] [PubMed]
  4. Ting, Y.S.; Egertson, J.D.; Payne, S.H.; Kim, S.; MacLean, B.; Kall, L.; Aebersold, R.; Smith, R.D.; Noble, W.S.; MacCoss, M.J. Peptide-Centric Proteome Analysis: An Alternative Strategy for the Analysis of Tandem Mass Spectrometry Data. Mol. Cell. Proteom. 2015, 14, 2301–2307. [Google Scholar] [CrossRef] [PubMed]
  5. Searle, B.C.; Swearingen, K.E.; Barnes, C.A.; Schmidt, T.; Gessulat, S.; Kuster, B.; Wilhelm, M. Generating high quality libraries for DIA MS with empirically corrected peptide predictions. Nat. Commun. 2020, 11, 1548. [Google Scholar] [CrossRef]
  6. Xing, X.; Li, X.; Wei, C.; Zhang, Z.; Liu, O.; Xie, S.; Chen, H.; Quan, S.; Wang, C.; Yang, X.; et al. DP-GAN+B: A lightweight generative adversarial network based on depthwise separable convolutions for generating CT volumes. Comput. Biol. Med. 2024, 174, 108393. [Google Scholar] [CrossRef] [PubMed]
  7. Zhang, S.; Yang, B.; Yang, H.; Zhao, J.; Zhang, Y.; Gao, Y.; Monteiro, O.; Zhang, K.; Liu, B.; Wang, S. Potential rapid intraoperative cancer diagnosis using dynamic full-field optical coherence tomography and deep learning: A prospective cohort study in breast cancer patients. Sci. Bull. 2024, 69, 1748–1756. [Google Scholar] [CrossRef] [PubMed]
  8. Wang, Z.; Liu, Z.; Yu, J.; Gao, Y.; Liu, M. Multi-scale nested UNet with transformer for colorectal polyp segmentation. J. Appl. Clin. Med. Phys. 2024, 25, e14351. [Google Scholar] [PubMed]
  9. Hu, H.; Feng, Z.; Lin, H.; Cheng, J.; Lyu, J.; Zhang, Y.; Zhao, J.; Xu, F.; Lin, T.; Zhao, Q.; et al. Gene function and cell surface protein association analysis based on single-cell multiomics data. Comput. Biol. Med. 2023, 157, 106733. [Google Scholar] [CrossRef]
  10. Xu, F.; Li, X.; Wu, R.; Qi, H.; Jin, J.; Liu, Z.; Wu, Y.; Lin, H.; Shen, C.; Shuai, J. Incoherent feedforward loop dominates the robustness and tunability of necroptosis biphasic, emergent, and coexistent dynamics. Fundam. Res. 2024. [Google Scholar] [CrossRef]
  11. Li, X.; Chen, G.; Zhou, X.; Peng, X.; Li, M.; Chen, D.; Yu, H.; Shi, W.; Zhang, C.; Li, Y.; et al. Roles of Akirin1 in early prediction and treatment of graft kidney ischemia–reperfusion injury. Smart Med. 2024, 3, e20230043. [Google Scholar] [CrossRef]
  12. Zhu, F.; Niu, Q.; Li, X.; Zhao, Q.; Su, H.; Shuai, J. FM-FCN: A Neural Network with Filtering Modules for Accurate Vital Signs Extraction. Research 2024, 7, 0361. [Google Scholar] [CrossRef] [PubMed]
  13. He, Q.; Zhong, C.Q.; Li, X.; Guo, H.; Li, Y.; Gao, M.; Yu, R.; Liu, X.; Zhang, F.; Guo, D.; et al. Dear-DIA(XMBD): Deep Autoencoder Enables Deconvolution of Data-Independent Acquisition Proteomics. Research 2023, 6, 0179. [Google Scholar] [CrossRef] [PubMed]
  14. Li, Y.; He, Q.; Guo, H.; Shuai, S.C.; Cheng, J.; Liu, L.; Shuai, J. AttnPep: A Self-Attention-Based Deep Learning Method for Peptide Identification in Shotgun Proteomics. J. Proteome Res. 2024, 23, 834–843. [Google Scholar] [CrossRef] [PubMed]
  15. He, Q.; Guo, H.; Li, Y.; He, G.; Li, X.; Shuai, J. SeFilter-DIA: Squeeze-and-Excitation Network for Filtering High-Confidence Peptides of Data-Independent Acquisition Proteomics. Interdiscip Sci. 2024. [Google Scholar] [CrossRef]
  16. Jiang, T.T.; Fang, L.; Wang, K. Deciphering “the language of nature”: A transformer-based language model for deleterious mutations in proteins. Innovation 2023, 4, 100487. [Google Scholar] [CrossRef] [PubMed]
  17. Yang, Z.; Zeng, X.; Zhao, Y.; Chen, R. AlphaFold2 and its applications in the fields of biology and medicine. Signal Transduct. Target. Ther. 2023, 8, 115. [Google Scholar] [CrossRef] [PubMed]
  18. Zeng, W.F.; Zhou, X.X.; Zhou, W.J.; Chi, H.; Zhan, J.; He, S.M. MS/MS Spectrum Prediction for Modified Peptides Using pDeep2 Trained by Transfer Learning. Anal. Chem. 2019, 91, 9724–9731. [Google Scholar] [CrossRef]
  19. Tiwary, S.; Levy, R.; Gutenbrunner, P.; Salinas Soto, F.; Palaniappan, K.K.; Deming, L.; Berndl, M.; Brant, A.; Cimermancic, P.; Cox, J. High-quality MS/MS spectrum prediction for data-dependent and data-independent acquisition data analysis. Nat. Methods 2019, 16, 519–525. [Google Scholar] [CrossRef]
  20. Gessulat, S.; Schmidt, T.; Zolg, D.P.; Samaras, P.; Schnatbaum, K.; Zerweck, J.; Knaute, T.; Rechenberger, J.; Delanghe, B.; Huhmer, A.; et al. Prosit: Proteome-wide prediction of peptide tandem mass spectra by deep learning. Nat. Methods 2019, 16, 509–518. [Google Scholar] [CrossRef]
  21. Yang, Y.; Liu, X.; Shen, C.; Lin, Y.; Yang, P.; Qiao, L. In silico spectral libraries by deep learning facilitate data-independent acquisition proteomics. Nat. Commun. 2020, 11, 146. [Google Scholar] [CrossRef]
  22. Zhou, X.X.; Zeng, W.F.; Chi, H.; Luo, C.; Liu, C.; Zhan, J.; He, S.M.; Zhang, Z. pDeep: Predicting MS/MS Spectra of Peptides with Deep Learning. Anal. Chem. 2017, 89, 12690–12697. [Google Scholar] [CrossRef]
  23. Lou, R.; Liu, W.; Li, R.; Li, S.; He, X.; Shui, W. DeepPhospho accelerates DIA phosphoproteome profiling through in silico library generation. Nat. Commun. 2021, 12, 6685. [Google Scholar] [CrossRef]
  24. Guan, S.; Moran, M.F.; Ma, B. Prediction of LC-MS/MS Properties of Peptides from Sequence by Deep Learning. Mol. Cell. Proteom. 2019, 18, 2099–2107. [Google Scholar] [CrossRef]
  25. Lin, Y.M.; Chen, C.T.; Chang, J.M. MS2CNN: Predicting MS/MS spectrum based on protein sequence using deep convolutional neural networks. BMC Genom. 2019, 20, 906. [Google Scholar] [CrossRef] [PubMed]
  26. Liu, K.; Li, S.; Wang, L.; Ye, Y.; Tang, H. Full-Spectrum Prediction of Peptides Tandem Mass Spectra using Deep Neural Network. Anal. Chem. 2020, 92, 4275–4283. [Google Scholar] [CrossRef]
  27. Chen, M.; Zhu, P.; Wan, Q.; Ruan, X.; Wu, P.; Hao, Y.; Zhang, Z.; Sun, J.; Nie, W.; Chen, S. High-Coverage Four-Dimensional Data-Independent Acquisition Proteomics and Phosphoproteomics Enabled by Deep Learning-Driven Multidimensional Predictions. Anal. Chem. 2023, 95, 7495–7502. [Google Scholar] [CrossRef] [PubMed]
  28. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural. Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
  29. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. Adv. Neural. Inf. Process. Syst. 2017, 30. [Google Scholar] [CrossRef]
  30. Distler, U.; Kuharev, J.; Navarro, P.; Levin, Y.; Schild, H.; Tenzer, S. Drift time-specific collision energies enable deep-coverage data-independent acquisition proteomics. Nat. Methods 2014, 11, 167–170. [Google Scholar] [CrossRef]
  31. Demichev, V.; Szyrwiel, L.; Yu, F.; Teo, G.C.; Rosenberger, G.; Niewienda, A.; Ludwig, D.; Decker, J.; Kaspar-Schoenefeld, S.; Lilley, K.S.; et al. dia-PASEF data analysis using FragPipe and DIA-NN for deep proteomics of low sample amounts. Nat. Commun. 2022, 13, 3944. [Google Scholar] [CrossRef] [PubMed]
  32. Helm, D.; Vissers, J.P.; Hughes, C.J.; Hahne, H.; Ruprecht, B.; Pachl, F.; Grzyb, A.; Richardson, K.; Wildgoose, J.; Maier, S.K.; et al. Ion mobility tandem mass spectrometry enhances performance of bottom-up proteomics. Mol. Cell. Proteom. 2014, 13, 3709–3715. [Google Scholar] [CrossRef] [PubMed]
  33. Gabelica, V.; Shvartsburg, A.A.; Afonso, C.; Barran, P.; Benesch, J.L.P.; Bleiholder, C.; Bowers, M.T.; Bilbao, A.; Bush, M.F.; Campbell, J.L.; et al. Recommendations for reporting ion mobility Mass Spectrometry measurements. Mass Spectrom. Rev. 2019, 38, 291–320. [Google Scholar] [CrossRef] [PubMed]
  34. Meier, F.; Brunner, A.D.; Frank, M.; Ha, A.; Bludau, I.; Voytik, E.; Kaspar-Schoenefeld, S.; Lubeck, M.; Raether, O.; Bache, N.; et al. diaPASEF: Parallel accumulation-serial fragmentation combined with data-independent acquisition. Nat. Methods 2020, 17, 1229–1236. [Google Scholar] [CrossRef] [PubMed]
  35. Navon, A.; Shamsian, A.; Achituve, I.; Maron, H.; Kawaguchi, K.; Chechik, G.; Fetaya, E. Multi-task learning as a bargaining game. arXiv 2022, arXiv:2202.01017. [Google Scholar]
  36. Aebersold, R.; Mann, M. Mass spectrometry-based proteomics. Nature 2003, 422, 198–207. [Google Scholar] [CrossRef]
  37. Escher, C.; Reiter, L.; MacLean, B.; Ossola, R.; Herzog, F.; Chilton, J.; MacCoss, M.J.; Rinner, O. Using iRT, a normalized retention time for more targeted measurement of peptides. Proteomics 2012, 12, 1111–1121. [Google Scholar] [CrossRef] [PubMed]
  38. Wren, S.A. Peak capacity in gradient ultra performance liquid chromatography (UPLC). J. Pharm. Biomed. Anal. 2005, 38, 337–343. [Google Scholar] [CrossRef]
  39. Vizcaino, J.A.; Deutsch, E.W.; Wang, R.; Csordas, A.; Reisinger, F.; Rios, D.; Dianes, J.A.; Sun, Z.; Farrah, T.; Bandeira, N.; et al. ProteomeXchange provides globally coordinated proteomics data submission and dissemination. Nat. Biotechnol. 2014, 32, 223–226. [Google Scholar] [CrossRef]
  40. Cote, R.G.; Griss, J.; Dianes, J.A.; Wang, R.; Wright, J.C.; van den Toorn, H.W.; van Breukelen, B.; Heck, A.J.; Hulstaert, N.; Martens, L.; et al. The PRoteomics IDEntification (PRIDE) Converter 2 framework: An improved suite of tools to facilitate data submission to the PRIDE database and the ProteomeXchange consortium. Mol. Cell. Proteom. 2012, 11, 1682–1689. [Google Scholar] [CrossRef]
  41. Ma, J.; Chen, T.; Wu, S.; Yang, C.; Bai, M.; Shu, K.; Li, K.; Zhang, G.; Jin, Z.; He, F.; et al. iProX: An integrated proteome resource. Nucleic Acids Res. 2019, 47, D1211–D1217. [Google Scholar] [CrossRef] [PubMed]
  42. Okuda, S.; Watanabe, Y.; Moriya, Y.; Kawano, S.; Yamamoto, T.; Matsumoto, M.; Takami, T.; Kobayashi, D.; Araki, N.; Yoshizawa, A.C.; et al. jPOSTrepo: An international standard data repository for proteomes. Nucleic Acids Res. 2017, 45, D1107–D1111. [Google Scholar] [CrossRef] [PubMed]
  43. Cox, J.; Mann, M. MaxQuant enables high peptide identification rates, individualized p.p.b.-range mass accuracies and proteome-wide protein quantification. Nat. Biotechnol. 2008, 26, 1367–1372. [Google Scholar] [CrossRef] [PubMed]
  44. Demichev, V.; Messner, C.B.; Vernardis, S.I.; Lilley, K.S.; Ralser, M. DIA-NN: Neural networks and interference correction enable deep proteome coverage in high throughput. Nat. Methods 2020, 17, 41–44. [Google Scholar] [CrossRef] [PubMed]
  45. Wang, H.; Lim, K.P.; Kong, W.; Gao, H.; Wong, B.J.H.; Phua, S.X.; Guo, T.; Goh, W.W.B. MultiPro: DDA-PASEF and diaPASEF acquired cell line proteomic datasets with deliberate batch effects. Sci. Data 2023, 10, 858. [Google Scholar] [CrossRef] [PubMed]
  46. Lou, R.; Cao, Y.; Li, S.; Lang, X.; Li, Y.; Zhang, Y.; Shui, W. Benchmarking commonly used software suites and analysis workflows for DIA proteomics and phosphoproteomics. Nat. Commun. 2023, 14, 94. [Google Scholar] [CrossRef] [PubMed]
  47. Zhang, S.; Chen, H.; Li, C.; Chen, B.; Gong, H.; Zhao, Y.; Qi, R. Water-Soluble Tomato Extract Fruitflow Alters the Phosphoproteomic Profile of Collagen-Stimulated Platelets. Front. Pharmacol. 2021, 12, 746107. [Google Scholar] [CrossRef]
  48. Tan, D.; Lu, M.; Cai, Y.; Qi, W.; Wu, F.; Bao, H.; Qv, M.; He, Q.; Xu, Y.; Wang, X.; et al. SUMOylation of Rho-associated protein kinase 2 induces goblet cell metaplasia in allergic airways. Nat. Commun. 2023, 14, 3887. [Google Scholar] [CrossRef]
  49. Teschner, D.; Gomez-Zepeda, D.; Declercq, A.; Lacki, M.K.; Avci, S.; Bob, K.; Distler, U.; Michna, T.; Martens, L.; Tenzer, S.; et al. Ionmob: A Python package for prediction of peptide collisional cross-section values. Bioinformatics 2023, 39, btad486. [Google Scholar] [CrossRef] [PubMed]
  50. Lei, W.L.; Li, Y.Y.; Meng, T.G.; Ning, Y.; Sun, S.M.; Zhang, C.H.; Gui, Y.; Wang, Z.B.; Qian, W.P.; Sun, Q.Y. Specific deletion of protein phosphatase 6 catalytic subunit in Sertoli cells leads to disruption of spermatogenesis. Cell Death Dis. 2021, 12, 883. [Google Scholar] [CrossRef]
  51. Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury, J.; Chanan, G.; Killeen, T.; Lin, Z.; Gimelshein, N.; Antiga, L. Pytorch: An imperative style, high-performance deep learning library. Adv. Neural. Inf. Process. Syst. 2019, 32. [Google Scholar]
  52. Caruana, R. Multitask learning. Mach. Learn. 1997, 28, 41–75. [Google Scholar] [CrossRef]
  53. Sun, T.; Shao, Y.; Li, X.; Liu, P.; Yan, H.; Qiu, X.; Huang, X. Learning sparse sharing architectures for multiple tasks. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; pp. 8936–8943. [Google Scholar]
Figure 1. ProPept-ST evaluates the prediction performance of unmodified peptide retention time. (A,B) The prediction accuracy of different models for peptide RT (A) and iRT (B) is compared based on the mean absolute error on various datasets. (C) Scatter plot comparing ProPept-ST predicted RT values with experimentally observed RT values for the SWATH library dataset. (D) Distribution of absolute errors in peptide RT prediction by the ProPept-ST model, both fine-tuned and retrained.
Figure 1. ProPept-ST evaluates the prediction performance of unmodified peptide retention time. (A,B) The prediction accuracy of different models for peptide RT (A) and iRT (B) is compared based on the mean absolute error on various datasets. (C) Scatter plot comparing ProPept-ST predicted RT values with experimentally observed RT values for the SWATH library dataset. (D) Distribution of absolute errors in peptide RT prediction by the ProPept-ST model, both fine-tuned and retrained.
Ijms 25 07237 g001
Figure 2. ProPept-ST assesses the performance of predicting phosphorylated peptide RT and the ablation experiment on the ProPept-ST model. (A) Comparison of the prediction accuracy of various models for phosphorylated peptide RT based on the median absolute error on different datasets. (B) Scatter plot comparing ProPept-ST-predicted iRT values with experimentally observed iRT values for the U2OS_DDA dataset. (C) Median absolute error of RT prediction by ProPept-ST and six other models on the benchmark dataset H4 DDAp. (D) Parameter count comparison between ProPept-ST and six other models.
Figure 2. ProPept-ST assesses the performance of predicting phosphorylated peptide RT and the ablation experiment on the ProPept-ST model. (A) Comparison of the prediction accuracy of various models for phosphorylated peptide RT based on the median absolute error on different datasets. (B) Scatter plot comparing ProPept-ST-predicted iRT values with experimentally observed iRT values for the U2OS_DDA dataset. (C) Median absolute error of RT prediction by ProPept-ST and six other models on the benchmark dataset H4 DDAp. (D) Parameter count comparison between ProPept-ST and six other models.
Ijms 25 07237 g002
Figure 3. ProPept-MT’s performance in predicting RT and IM, as well as the loss curves for training three tasks on specific datasets. (A) Distribution of absolute errors for predicting RT on benchmark datasets for each model. (B,D) Scatter plots showing ProPept-MT’s predictions of RT (B) and IM (D) on the H2 DIA test set. (C) Distribution of absolute errors for predicting IM on benchmark datasets for ProPept-MT and ProPept-ST. (E) Loss curves for training and validation of the three tasks on dataset H5 DDAp for ProPept-MT. (F) On the H7 DDAp training set, the loss curves of ProPept-ST retrained on three tasks and the fine-tuned loss curves of ProPept-MT.
Figure 3. ProPept-MT’s performance in predicting RT and IM, as well as the loss curves for training three tasks on specific datasets. (A) Distribution of absolute errors for predicting RT on benchmark datasets for each model. (B,D) Scatter plots showing ProPept-MT’s predictions of RT (B) and IM (D) on the H2 DIA test set. (C) Distribution of absolute errors for predicting IM on benchmark datasets for ProPept-MT and ProPept-ST. (E) Loss curves for training and validation of the three tasks on dataset H5 DDAp for ProPept-MT. (F) On the H7 DDAp training set, the loss curves of ProPept-ST retrained on three tasks and the fine-tuned loss curves of ProPept-MT.
Ijms 25 07237 g003
Figure 4. Performance of ProPept-MT in predicting fragment ion intensity. (A) Histogram distribution of PCC for each peptide on the H1 DDA and H5 DDAp test sets. (B) Mirror plot showing the experimental and predicted values of fragment ion intensities for two specific peptides (unmodified peptide and phosphopeptide). (C) Box plots showing the distribution of PCC, DP, and SA for ProPept-MT on the H1 DDA and H6 DDAp test sets.
Figure 4. Performance of ProPept-MT in predicting fragment ion intensity. (A) Histogram distribution of PCC for each peptide on the H1 DDA and H5 DDAp test sets. (B) Mirror plot showing the experimental and predicted values of fragment ion intensities for two specific peptides (unmodified peptide and phosphopeptide). (C) Box plots showing the distribution of PCC, DP, and SA for ProPept-MT on the H1 DDA and H6 DDAp test sets.
Ijms 25 07237 g004
Figure 5. Comparing the performance of ProPept-MT and DeepDIA in predicting RT, IM, and fragment ion intensity. (A) Distribution of absolute errors for predicting RT on different datasets for each model. (B) Distribution of absolute errors for predicting IM on different datasets for ProPept-MT and ProPept-ST. (C) Assessing ProPept-MT’s performance in predicting fragment ion intensity for different precursor charges on benchmark datasets. (D) Distribution of dot product (DP) for predicting fragment ion intensity of 2+ and 3+ precursor charges on the H1 DDA test set for each model.
Figure 5. Comparing the performance of ProPept-MT and DeepDIA in predicting RT, IM, and fragment ion intensity. (A) Distribution of absolute errors for predicting RT on different datasets for each model. (B) Distribution of absolute errors for predicting IM on different datasets for ProPept-MT and ProPept-ST. (C) Assessing ProPept-MT’s performance in predicting fragment ion intensity for different precursor charges on benchmark datasets. (D) Distribution of dot product (DP) for predicting fragment ion intensity of 2+ and 3+ precursor charges on the H1 DDA test set for each model.
Ijms 25 07237 g005
Figure 6. The workflow and model architecture of ProPept-MT. (A) ProPept-MT employs a multi-task deep learning model to generate in silico prediction libraries from protein or peptide sequence databases. (B) ProPept-MT is used for predicting RT, IM, and fragment ion intensity for any given unmodified peptide or phosphopeptide. Given the peptide sequence and precursor charge as input, our model uses Transformer encoder modules and a BiLSTM network to calculate context representations for all amino acids, which it finally outputs through separately designed output layers for each task.
Figure 6. The workflow and model architecture of ProPept-MT. (A) ProPept-MT employs a multi-task deep learning model to generate in silico prediction libraries from protein or peptide sequence databases. (B) ProPept-MT is used for predicting RT, IM, and fragment ion intensity for any given unmodified peptide or phosphopeptide. Given the peptide sequence and precursor charge as input, our model uses Transformer encoder modules and a BiLSTM network to calculate context representations for all amino acids, which it finally outputs through separately designed output layers for each task.
Ijms 25 07237 g006
Table 1. Performance of ProPept-ST under different parameters. “d_model” is the dimension of the embedded vector. “n_head” is the number of attention heads calculated in parallel in the multi-head self-attention module. “d_ff_lstm” is a hidden layer dimension in the lstm layer. “n_lstm” is the number of layers of lstm.
Table 1. Performance of ProPept-ST under different parameters. “d_model” is the dimension of the embedded vector. “n_head” is the number of attention heads calculated in parallel in the multi-head self-attention module. “d_ff_lstm” is a hidden layer dimension in the lstm layer. “n_lstm” is the number of layers of lstm.
d_modeln_headd_ff_lstmn_lstmMedAE
256851210.715
256851220.732
5001051210.723
5001051220.720
256825610.724
256825620.768
5001025610.722
5001025620.755
Table 2. Performance metrics of ProPept-MT on benchmark datasets.
Table 2. Performance metrics of ProPept-MT on benchmark datasets.
Data NameMetrics/ModelRetention TimeIon IntensityIon Mobility
R 2 Δ t 95 % MedAEIQRPCCSAPCCDP R 2 PCC Δ t 95 %
H1_DDADeepPhospho0.9756.5881.1572.3150.9900.8520.9580.973---
ProPept-ST0.98055.2000.6431.3230.9900.8700.96570.9790.9770.98870.0961
ProPept-MT0.98104.9740.5981.2640.9910.8720.96630.9800.9780.98920.0963
H2_DIADeepPhospho0.9895.4811.0192.0460.9970.8050.8890.953---
ProPept-ST0.9982.3330.3190.6380.9990.8170.9000.9590.9860.99310.062
ProPept-MT0.9972.8070.3950.7720.99860.8200.9010.9600.9840.99280.063
H3_DIAPDeepPhospho0.98611.4282.1034.1110.9970.7860.8720.944---
ProPept-ST0.99527.0450.7531.5080.997640.7970.8820.94970.9850.99240.0614
ProPept-MT0.99516.9730.8701.7130.997630.7980.8890.95020.9840.99220.0617
H4_DDApDeepPhospho0.97610.5241.9153.6780.9900.8090.9280.955---
ProPept-ST0.98356.4470.7151.4240.99180.8310.9410.9650.9710.98560.102
ProPept-MT0.98396.2030.7301.4220.99190.8350.9450.9670.9720.98620.099
H5_DDApDeepPhospho0.98012.6582.4084.8220.9930.8190.9350.960---
ProPept-ST0.9879.2490.9451.8860.99350.83240.9400.9660.9610.9810.100
ProPept-MT0.9888.6991.0772.1420.99390.83170.9410.9650.9590.9800.102
H6_DDApDeepPhospho0.98012.8832.6004.2310.9930.8040.9280.953---
ProPept-ST0.9916.5190.8201.6600.9960.8140.9340.9580.9630.98150.117
ProPept-MT0.9906.1870.8021.5880.9950.8240.9400.9620.9600.98090.113
H7_DDApDeepPhospho0.9585.3520.8311.5980.9830.8070.9320.954---
ProPept-ST0.9772.8530.3180.6500.9880.8230.9410.9610.9820.9910.080
ProPept-MT0.9802.2550.2940.5870.9900.8380.9500.9680.9860.9940.067
M1_DDApDeepPhospho0.97611.7161.8093.5350.9910.8150.9380.958---
ProPept-ST0.9916.5190.8201.6600.9960.8140.9340.9580.9630.9810.117
ProPept-MT0.9895.4980.7021.4230.9950.8340.9490.9660.9820.9920.077
M2_DDApDeepPhospho0.9664.9440.8121.5180.9860.7920.9180.947---
ProPept-ST0.9803.0500.3670.7550.9900.8070.9270.9550.9410.9730.112
ProPept-MT0.9821.6680.2430.4830.9910.8270.9410.9630.9550.9780.090
Table 3. Comparison of performance metrics between ProPept-MT and DeepDIA.
Table 3. Comparison of performance metrics between ProPept-MT and DeepDIA.
Data NameMetrics/ModelRetention TimeIon Intensity(2+)Ion Intensity(3+)Ion Mobility
R 2 Δ t 95 % MedAEIQRPCCSAPCCDPSAPCCDP R 2 PCC Δ t 95 %
H1_DDADeepDIA0.9757.1320.9741.9620.9870.8020.9500.9520.7210.9000.905---
ProPept-ST0.9805.4220.6701.3990.99020.8470.9590.9710.7910.9250.9470.9740.9870.106
ProPept-MT0.9814.8230.6621.3400.99050.8660.9680.9780.8170.9430.9590.9760.9880.100
H2_DIADeepDIA0.9944.2140.5901.1780.9970.7380.9130.9170.7080.8930.897---
ProPept-ST0.99732.8170.3810.7700.9990.8120.8940.9570.7920.8650.9470.9840.99210.066
ProPept-MT0.99702.9930.4260.8520.9980.8210.9070.9610.7940.8750.9480.9830.99180.069
H3_DIApDeepDIA0.97915.4101.9643.9090.9900.6920.8800.8850.6730.8670.871---
ProPept-ST0.9929.4131.1342.2820.9960.7880.85070.9450.76600.8370.93320.9790.9900.073
ProPept-MT0.9938.3010.9391.8730.9970.7830.85130.9420.76640.8520.93350.9820.9910.068
H4_DDApDeepDIA0.97314.7201.9303.9100.9860.7460.9160.9210.6680.8610.867---
ProPept-ST0.9848.2430.9921.9720.9920.8050.9280.9540.7500.8910.9240.9600.9800.126
ProPept-MT0.9866.6850.8651.7290.9930.8190.9380.9600.7740.9130.9370.9670.9840.l13
H5_DDApDeepDIA0.98312.8502.3723.4190.9920.7710.9330.9360.6920.8800.885---
ProPept-ST0.9898.4840.9421.8880.99480.8190.9400.9600.7650.8970.9330.9600.9810.104
ProPept-MT0.9917.2250.8501.7050.99540.8390.9520.9680.7980.9280.9500.9690.9840.086
H6_DDApDeepDIA0.91821.3892.4564.9280.9580.6700.8610.8690.6570.8530.859---
ProPept-ST0.94216.9911.9243.8570.9710.7540.8730.9260.7650.9050.9330.9230.9630.159
ProPept-MT0.9736.4910.7051.3970.9870.7800.9280.9470.7910.9280.9470.9400.9710.145
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

He, G.; He, Q.; Cheng, J.; Yu, R.; Shuai, J.; Cao, Y. ProPept-MT: A Multi-Task Learning Model for Peptide Feature Prediction. Int. J. Mol. Sci. 2024, 25, 7237. https://doi.org/10.3390/ijms25137237

AMA Style

He G, He Q, Cheng J, Yu R, Shuai J, Cao Y. ProPept-MT: A Multi-Task Learning Model for Peptide Feature Prediction. International Journal of Molecular Sciences. 2024; 25(13):7237. https://doi.org/10.3390/ijms25137237

Chicago/Turabian Style

He, Guoqiang, Qingzu He, Jinyan Cheng, Rongwen Yu, Jianwei Shuai, and Yi Cao. 2024. "ProPept-MT: A Multi-Task Learning Model for Peptide Feature Prediction" International Journal of Molecular Sciences 25, no. 13: 7237. https://doi.org/10.3390/ijms25137237

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop