Next Article in Journal
Changes in Expression in BMP2 and Two Closely Related Genes in Guinea Pig Retinal Pigment Epithelium during Induction and Recovery from Myopia
Previous Article in Journal
Effects of Nitro-Oxidative Stress on Biomolecules: Part 1—Non-Reactive Molecular Dynamics Simulations
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Prediction of Cholecystokinin-Secretory Peptides Using Bidirectional Long Short-term Memory Model Based on Transfer Learning and Hierarchical Attention Network Mechanism

1
College of Information Engineering, Shanghai Maritime University, Shanghai 201306, China
2
School of Health Science and Engineering, University of Shanghai for Science and Technology, Shanghai 200093, China
3
Joint Center for Translational Medicine, Southern Medical University Affiliated Fengxian Hospital, Shanghai 201499, China
*
Authors to whom correspondence should be addressed.
Biomolecules 2023, 13(9), 1372; https://doi.org/10.3390/biom13091372
Submission received: 19 June 2023 / Revised: 30 July 2023 / Accepted: 16 August 2023 / Published: 11 September 2023

Abstract

:
Cholecystokinin (CCK) can make the human body feel full and has neurotrophic and anti-inflammatory effects. It is beneficial in treating obesity, Parkinson’s disease, pancreatic cancer, and cholangiocarcinoma. Traditional biological experiments are costly and time-consuming when it comes to finding and identifying novel CCK-secretory peptides, and there is an urgent need to develop a new computational method to predict new CCK-secretory peptides. This study combines the transfer learning method with the SMILES enumeration data augmentation strategy to solve the data scarcity problem. It establishes a fusion model of the hierarchical attention network (HAN) and bidirectional long short-term memory (BiLSTM), which fully extracts peptide chain features to predict CCK-secretory peptides efficiently. The average accuracy of the proposed method in this study is 95.99%, with an AUC of 98.07%. The experimental results show that the proposed method is significantly superior to other comparative methods in accuracy and robustness. Therefore, this method is expected to be applied to the preliminary screening of CCK-secretory peptides.

1. Introduction

Cholecystokinin (CCK) is a gastrointestinal hormone that causes gallbladder contraction and was discovered and named by Ivy and Oldbery in 1928. CCK is a peptide hormone comprising 33 amino acids, released by the small intestinal mucosa I cell. CCK plays various roles in the organism, such as producing satiety to inhibit food intake, slowing gastric emptying, and stimulating pancreatic and gallbladder secretion production [1,2,3,4]. Cholecystokinin-8 and nerve growth factor (NGF) work together to maintain and repair the nervous system [5]. Su et al. [6] found that CCK has neurotrophic and anti-inflammatory effects and improves adverse effects, such as the inflammatory response and neuronal damage, in Parkinson’s disease patients. Numerous experiments have confirmed that CCK is a vital tumor growth factor in the digestive tract and has an apparent promoting effect on pancreatic and bile duct cancer [7,8]. In recent years, therapeutic peptides have emerged as an advanced and novel cancer treatment strategy that can treat many diseases [9,10,11]. Compared with traditional recombinant antibody therapies, peptide-based therapies are highly targeted, productive, less toxic, and easy to synthesize and modify [12]. CCK as a therapeutic peptide will also offer broad research prospects. CCK-secretory peptides can target the intestinal endocrine cell membrane CaSR to activate the Gq signalling pathway, promoting the secretion of CCK by the intestinal endocrine cells. High-quality CCK-secretory peptides have the advantages of safety, no toxic side effects, gastrointestinal digestive enzyme hydrolysis tolerance, and easy absorption. CCK-secretory peptides are suitable for developing food and drugs as functional ingredients or food base materials. The prediction of CCK-secretory peptides has a significant application value in developing foods, health foods, and drugs with functions such as delaying gastric emptying, promoting satiety, and weight loss.
In recent decades, biological researchers have conducted numerous experiments to study the substances that stimulate CCK secretion. It has been demonstrated that the primary nutrients stimulating CCK release are ingested fats and proteins, especially amino acids, peptides, and protein hydrolysates [13]. Daly Kristian et al. [14] demonstrated that Phe, Leu, and Glu induce CCK secretion via small-intestinal tissue experiments in mice. In pig jejunal tissue experiments, Leu, Ile, or a mixture of amino acids significantly increased CCK secretion [15]. Santos-Hernández et al. [16] found that VLLPDEVSGL and a derivative fragment, VLLPD, induced CCK release in gastrointestinal egg-white digestion. The above experiments require a high amount of time from biological researchers, with the help of numerous devices, to speculate whether peptides or amino acids have a facilitative effect on CCK. These tasks are tedious, labor-intensive, and costly.
With the increasing therapeutic importance of CCK, there is an urgent need to develop efficient, accurate, and cost-effective predictive techniques for CCK-secretory peptides. The issues above have prompted us to propose a method to address the limitations of biological clinical experiments. Nowadays, numerous in silico techniques have been applied in various fields. Mol2vec [17] has been utilized in the field of chemistry to identify the vector representations of molecular substructures. Similarly, Jo et al. [18] employed message-passing neural networks for predicting SMILES chemical properties. In natural language processing (NLP), the word2vec [19] method has also been applied to the semantic search of relevant documents. Sunjaya et al. [20] used ARIMA and long short-term memory (LSTM) methods to predict the number of positive COVID-19 cases in Indonesia, and the results showed that the LSTM model outperformed the ARIMA model. Han et al. [21] proposed an LSTM model that combines input and hidden layer attention mechanisms for long-term streamflow prediction in the environmental domain. Liu et al. [22] introduced the MolRoPE–BERT framework, which integrates an effective position-encoding method for capturing sequence position information and a pre-trained BERT model for molecular property prediction. However, these in silico techniques still face challenges, as they require a large number of training data and may struggle to achieve a good prediction performance on small-scale datasets.
This paper describes a method to predict CCK-secretory peptides with high accuracy. The sequences of peptides associated with CCK secretion are efficiently characterized as SMILESs, and data augmentation is achieved via SMILES enumeration. Based on this, a BiLSTM model based on transfer learning and hierarchical attention networks was developed to predict CCK-secretory peptides. In the training phase, a two-stage transfer learning strategy was adopted to train the model, to prevent overfitting. The data augmentation and transfer learning strategy solves the problem of the scarcity of CCK-secretory peptide data, while improving the performance and generality of the model. The model has a more significant performance and accuracy compared with machine learning models based on the AAC and DPC representations and similar transfer learning models. Therefore, the proposed method is expected to be applied in the initial screening of CCK-secretory peptides. Molecular docking [23,24] helps us to understand the interactions between molecules, and the combination of the proposed method and molecular docking is conducive to developing CCK-secretory peptide foods and drugs.

2. Materials and Methods

2.1. Construction of Datasets

In this work, two datasets were used to construct the model, named as the source dataset and target dataset. The details of the datasets are shown in Table 1.
Positive samples are denoted by C + , and negative samples are denoted by C . Therefore, the whole dataset can be denoted as C .
C = C + C ,
In addition, there is no overlap between the positive and negative dataset.
= C + C ,
Source dataset: To train a source task prediction model, four high-quality absorption, distribution, metabolism, excretion, and toxicity (ADMET) datasets were collected from ADMETlab [25,26]. The source datasets contain pgb-sub, DILI, and SR-MMP, which have some similarities to the target dataset, in that they are all drug-related datasets. The CCK-secretory peptides can also be used in targeted therapy peptide drug development. The datasets with partial similarity are conducive to the transfer of transfer learning model knowledge. And these datasets have different sizes, ranging from 467 to 1840, which can better explore the differences between datasets.
Target dataset: This dataset contains 99 data obtained from 37 papers investigating CCK-secretory peptide activity. Among them, the number of samples with effective peptides was 54, and the number of samples with ineffective peptides was 45. The sample distribution in the dataset is relatively balanced, which can prevent the overfitting of the model caused by an unbalanced dataset, to some extent.
Each dataset is randomly divided into training and testing datasets, using a hierarchical strategy, with ratios of 0.8 and 0.2, respectively. Therefore, the sample distribution in the training and test datasets is the same for an original dataset. All experiments were repeated five times, using different random seeds, to minimize the impact of the sample distribution on the model performance. The model was trained on the training dataset, and the parameter selection was adjusted to achieve a better model performance. The average of the five repeated experiments on the test dataset was calculated as the final model performance. The proposed model and comparison method were evaluated according to sensitivity (Sen), specificity (Spe), accuracy (Acc), Matthews correlation coefficient (MCC), and the area under the curve (AUC).

2.2. Feature Representation of Peptides

Efficient feature representation is essential for constructing predictive models and, therefore, requires the generation of corresponding features for each peptide sequence. The feature representation method has an important impact on the performance of the prediction model. This study used three different features to develop models based on machine learning and deep learning techniques. The machine learning model uses the currently popular machine learning peptide representation, including amino acid composition and dipeptide composition. The deep learning model uses a SMILES to characterize the sequence of the peptide. Below is a brief description of the different feature representation methods and principles.

2.2.1. Amino Acid Composition

Characterizing amino acids or protein sequences as AAC is one of the simplest methods available for feature extraction. The AAC descriptor represents the frequency of standard amino acids occurring in protein sequences [27]. As all peptides are composed of 20 standard amino acids, the AAC can be expressed as a vector of dimension 20, and we calculate the AAC as follows.
A A C ( i ) = a a ( i ) L ,
where A A C ( i ) is the abundance of the ith amino acid, a a ( i ) is the number of occurrences of the ith amino acid, and L is the length of the peptide.

2.2.2. Dipeptide Composition

Dipeptide composition (DPC) can express more information than AAC and represent the local order of amino acid sequence pairs [28]. We calculated all possible dipeptides (i.e., amino acid sequence pairs, e.g., AA, AC, etc.) with a dimension 400 (20 × 20) vector, to represent the DPC. The DPC is calculated as follows.
D P C ( i ) = d p ( i ) L 1 ,
where D P C ( i ) is the frequency of occurrence of the ith dipeptide, d p ( i ) is the number of circumstances of the ith dipeptide, and L is the length of the peptide.

2.2.3. Simplified Molecular Input Line Entry System

The Simplified Molecular Input Line Entry System (SMILES) is a chemical notation system for modern chemical information processing [29]. The SMILES is based on the principles of molecular graph theory and allows for a strict structural specification, by using a minimal and natural syntax. The SMILES naming rules are simple and easy to understand for researchers, and the system is considered to be the best compromise between human and machine chemical notation to date. Computationally, the SMILES interprets chemical molecules speedily and compactly, thus meeting machine time and space-saving goals. The SMILES is based on a computer-language-parsing approach, and our conversion of peptide sequences to the SMILES will significantly improve the data processing efficiency.

2.3. Data Augmentation

Deep learning has developed rapidly in recent years, and the computational power has increased greatly. However, deep learning usually requires a large number of samples for training, to achieve better generalization. Data augmentation is a good idea to solve this problem. Data augmentation refers to the method of increasing the number of data by adding small changes to the existing data, or creating new synthetic data from the existing data. This strategy solves the data imbalance problem, while expanding the dataset, improves the performance and generalization of the model, and prevents the model from overfitting, to some extent [30].
This study used SMILES enumeration for data augmentation. SMILES enumeration uses a Python script to generate multiple SMILESs, by changing the order of atoms using the cheminformatics library RDKit, wherein different atom orders lead to different SMILESs, thus achieving data augmentation [31]. An example of the use of a SMILES for data augmentation is shown in Figure 1.

2.4. Machine Learning Models

To demonstrate the predictive power of a hierarchy-attention-network-based BiLSTM deep learning model for CCK-secretory peptides, we built some classical machine learning models for comparison. Specifically, in this study, we characterized the peptide sequences as AAC and DPC forms, respectively, and then built machine learning classifiers based on the AAC and DPC representations using SVM light and scikit-learn from the Python library, respectively, including support vector machines (SVM), random forest (RF), multilayer perceptron (MLP), k-nearest neighbor (KNN), gradient-boosting decision tree (GBDT), and extreme gradient boosting (XGBoost). The relevant parameters of different classifiers were adjusted in the modeling process, and the best parameters are reported.

2.5. Transfer Learning Models

The transfer learning model used in this work is a bidirectional long and short-term memory network model based on a hierarchical attention network. SMILESs encoded after data enhancement are fed into this model. A BiLSTM is used to aggregate the information and extract the best features, through a new hierarchical attention network. Following are brief descriptions of each part of the model.

2.5.1. Bidirectional Long and Short-Term Memory

The bidirectional long and short-term memory network comprises forward long and short-term networks and backwards long and short-term networks. BiLSTM and LSTM are variants of RNN. Researchers solved the problems of RNN gradient explosion and the poor ability to rely on information over long distances by adding gating mechanisms (forget gate, input gate, output gate) to preserve information long term, thus proposing LSTM [32]. The sequences of peptides are input to two LSTMs of BiLSTM in forward and reverse order, respectively, for feature extraction, and the word vector formed via the stitching of the two output vectors (i.e., the extracted feature vectors) is used as the final feature expression of the word. The model design concept of BiLSTM is to make the feature data obtained at the time t have the information between the past and future simultaneously. It is worth mentioning that the two LSTM neural network parameters in BiLSTM are independent, and they only share the word-embedding word vector list. The structure of BiLSTM is shown schematically in Figure 2.
The LSTM model can better capture the longer-distance dependencies. This is because LSTM uses a gating mechanism to determine which information should be remembered and which should be forgotten. However, modelling sentences with LSTM also has a problem: it cannot encode information from back to front. In this study, we used BiLSTM to aggregate the sequence information of peptides. BiLSTM allows for the better capture of bidirectional textual semantic dependencies.

2.5.2. Hierarchical Attention Network Mechanism

The attentional mechanism mimics the internal processes of biological observation behavior and is a mechanism that combines internal experience with external sensation, to improve the precision of observation in certain regions [33]. This mechanism provides a higher scalability and robustness, by focusing on important information based on the size of the weights, and continuously adjusting the weights so that important information can be selected, even under different circumstances. In this study, the hierarchical attention network mechanism is used to focus on more critical features gradually.
The hierarchical attention network (HAN) consists of several components: a word sequence encoder, word-level attention layer, sentence encoder, and sentence-level attention layer [34]. The embedding word layer first transforms the input text into vector form. The BiLSTM layer extracts the text features, the global information of the text is obtained using the self-attention layer and, finally, the fusion and pooling are performed by the fusion and pooling layers. Finally, the fusion and pooling layers are used to fuse and pool the text. After the splicing layer, the fully connected layer, and the SoftMax, the implication relationship in the text is obtained, to achieve the classification of the text. The attention layer is calculated as follows:
u p i = tanh ( W W h p i + b w ) ,
α p i = exp ( u P i T u w ) i exp ( u P i T u w ) ,
V P = i α p i h p i ,
where W w and b w are the adjustable weights and bias terms of the attention model, tanh is the activation function, h p i is the output of the BiLSTM layer, u w is the weight value, and the calculated result α p i indicates the important information of each word in the sentence; V P is the output vector of the text after the attention model calculation. Similarly, the output vector V h of the hypothetical text after the attention model calculation can also be calculated.

2.5.3. Design of Transfer Learning Models

Transfer learning is an approach to improving a learner by transferring information from one domain to a related domain, which can reduce training costs and overcome the problems of data scarcity and time-consuming training [35]. Transfer learning can be simply defined as: given a source domain D s , a learning task T s , and corresponding target domains D t and T t , the effect of improving the target prediction function f t ( · ) in D t by the learning knowledge already available in D s and T s .
The transfer learning method used in this study is parametric transfer, and the transfer learning model is built in two phases using this method. In the first phase, the source dataset is fed into the model for source task pre-training and prediction. In the second stage, the model parameters learned from the source domain are migrated to the target dataset, and the parameters are adjusted slightlym to make the model applicable to the target task. The transfer learning process in this study is shown in Figure 3.

2.6. Our Approach

The framework of our proposed method is shown in Figure 4. The amino acid and peptide sequences are randomly divided into training and test sets, in a 4:1 ratio. All peptide sequences are represented as SMILESs, and the target dataset is augmented via the implementation of a SMILES enumeration strategy. The SMILESs were then encoded, and the encoded features were used as the input for the transfer learning model mentioned previously. A BiLSTM prediction model, based on transfer learning and the hierarchical attention network mechanism, is used to accomplish the target task, which is named by us as TL-AHB.
Specifically, the target dataset was preprocessed using the SMILES enumeration technique, and the model was pre-trained on the source dataset using transfer learning. We solved the scarcity problem in the training data, while significantly improving the performance and accuracy of the model. Considering the large number of SMILESs of large length in the dataset, BiLSTM can better capture the textual semantic dependencies in both directions over longer distances. The hierarchical attention network, which progressively extracts features with higher weights, provides the most critical features for the model. Thus, the TL-AHB model combines the advantages of BiLSTM and HAN, which is important in improving the model’s performance. The use of transfer learning and data augmentation strategies allows us to obtain models with good predictive and generalization capabilities compared to traditional methods.

2.7. Performance Evaluation Metrics

This study investigates a prediction problem of CCK-secretory peptides, and this problem can be considered a binary classification problem. Therefore, five metrics widely used in binary classification tasks were used to evaluate the performance of the model, including the sensitivity ( S e n ), specificity ( S p e ), accuracy ( A c c ), Matthews correlation coefficient ( M C C ), and area under the curve (AUC). Each metric was calculated as follows:
S e n = T P T P + F N ,
S p e = T N T N + F P ,
A c c = T P + T N T P + F P + T N + F N   ,
M C C = ( T P × T N ) ( F P × F N ) ( T P + F P ) ( T P + F N ) ( T N + F P ) ( T N + F N ) ,
where T P represents the true positive number, T N represents the true negative number, F P represents the false positive number, and F N represents the false negative number. S e n , S p e , and A c c take values in the range (0, 1), and M C C takes values in the range (−1, 1). S e n and S p e measure the predictive ability of the classification predictor for positivity and negativity, respectively, and A c c and M C C are used to assess the overall performance of the predictor. ROC curves were plotted based on the false-positive and false-negative rates, and AUC was the area under the ROC curve.

3. Results and Discussion

To comprehensively evaluate the ability of our proposed method to predict CCK-secretory peptides, firstly, the amino acids of peptides were analyzed, and machine learning models based on AAC and DPC representations were built on the target dataset. Then, the performance of our proposed method was analyzed on the source and target tasks, respectively. Finally, some CCK-secretory peptide sequences with high similarity were used to verify the model’s predictive ability.

3.1. Compositional Analysis

Amino acid composition analysis is critical, because the two-terminal residues play a key role in the biological study of peptides. Firstly, the amino acid distributions in the active, the inactive, and all peptides need to be analyzed in the target dataset. The results of the analysis are shown in Figure 5, where the compositions of all 20 amino acids in the peptides associated with CCK secretion are compared and counted. Specific residues, including Glu, Phe, and Arg, were higher in the effective peptides than in the ineffective ones. In contrast, Ala, Gly, Leu, and Pro were higher in the ineffective peptides.

3.2. Analysis of Machine Learning Model Based on AAC and DPC Representations

Some peptide sequences are long, and their arrangement is complex, making it impractical to use the sequence directly for prediction; as commonly used characterization methods for peptide sequence prediction, AAC and DPC can simplify sequences, extract different features to replace complex sequences, and achieve interval prediction. Both have advantages in characterizing peptide sequences, and AAC features can represent a sequence with fewer feature dimensions and more straightforward calculations. On the other hand, dipeptide features contain the positional information of amino acids in protein sequences, making them more accurate in characterizing sequence information. We established machine learning models on the target dataset using six classic algorithms (SVM, RF, MLP, KNN, GBDT, XGBoost) to study the performance of different models, based on two representation methods. Each model has been adjusted and tested on the optimal parameters. These models were evaluated on the test set, and the experimental results are shown in Table 2 and Table 3.
Table 2 and Table 3 show the details of the performance comparison of the machine learning models based on AAC and DPC, respectively. It can be observed that the best performance is achieved using the GBDT algorithm, with 84.62% for Sen, 85.71% for Spe, 85.19% for Acc, 70.33% for MCC, and 85.16% for AUC in the machine learning prediction model with the AAC representation approach. In the machine learning prediction model with the DPC representation approach, the SVM algorithm model showed the best performance, but Acc was only 57.14%, and AUC was 59.09%. It is obvious that the performance of the machine learning prediction model based on the AAC representation approach is better than that of the prediction model based on the DPC representation approach, overall. Both representation methods have their limitations. The AAC feature disregards the positional information of the amino acids in the sequence, while the DPC feature has a higher dimensionality, making the calculations relatively complex. Due to data scarcity, both methods show a poor performance. Although the former has shown a better prediction ability, the model still has huge room for improvement when applied to predict CCK-secretory peptides. This also motivates the creation of our proposed method.

3.3. SMILES Augmentation Times on Model Impact Analysis

The SMILES of one molecule may obtain tens or even hundreds of different SMILES expressions via data augmentation, and the amount of data has a great impact on the model performance, so we analyzed the relationship between the number of SMILES augmentations and the model performance. The impact of different augmentation times (1, 5, 10, 20, 30, 40, and 50) on the model performance was evaluated on the source dataset using pgb-sub as an example. As shown in Figure 6, the AUC of the model is only about 81% when no data augmentation is applied to the SMILES. As the augmentation times increase, the AUC also increases, and the trend of performance improvement eventually levels off. In order to consider the performance and training time of the model, an optimal augmentation threshold needs to be determined for the SMILES. According to the results shown in Figure 6, the model levelled off at an augmentation time of 30, so the augmentation time of the SMILES was set to 30.

3.4. Analysis of Prediction Results of Source Tasks on Different Models

Then, we built four deep learning pre-trained models on the source dataset and used them for the prediction analysis of three datasets, DILI, pgb-sub, and SR-MMP. These four deep learning pre-trained models include HL, AHL, HB, and AHB. This set of experiments demonstrates the predictive power of the BiLSTM model based on SMILES representation and hierarchical attention network, where A denotes data enhancement using SMILES enumeration, and H denotes the use of the hierarchical attention network mechanism. The evaluation results of the source dataset on different pre-trained models are shown in Table 4.
As shown in Table 4, experimental control groups, with and without data augmentation, were set up in the pre-training experiments, to show the effect of SMILES enumeration on the pre-trained models. In the three source datasets, AHL improved the AUC by 2.04%, 4.99%, and 2.26%, respectively, compared with the HL model; AHB improved the AUC by 4.00%, 4.48%, and 1.66%, respectively, compared with the HB model. The performances of the models with data augmentation were all improved to some extent. On the other hand, BiLSTM also played a key role in model performance improvement. AHB improved the AUC by 2.17%, 1.74%, and 0.27% for the three datasets compared to the AHL model, respectively. As mentioned earlier, LSTM captures information through a gating mechanism and can only retain unidirectional information about the sequence, and cannot encode information from the reverse to the forward direction. BiLSTM, because of the introduction of reverse information, can be trained in both the forward and reverse directions. Despite the higher computational complexity, BiLSTM captures bidirectional textual semantic dependencies, learns global and local features, and can better aggregate peptide chain information. BiLSTM can provide a better performance in datasets containing a large number of long sequences.

3.5. Analysis of Prediction Results of Target Tasks on Different Models

After completing pre-training on the source dataset, four transfer learning models, including TL-HL, TL-AHL, TL-HB, and TL-AHB, were obtained by combining the transfer learning strategy and parameter fine-tuning, wherein TL denotes transfer learning. Finally, the four transfer learning models were used in the identification of CCK-secretory peptides. The performance evaluation of the models is shown in Table 5.
It is obvious that the BiLSTM model (TL-AHB) based on transfer learning and the hierarchical attention network mechanism showed the best performance in predicting CCK-secretory peptides. The sensitivity (Sen) was 90.97%, the specificity (Spe) was 98.92%, the accuracy (Acc) was 95.99%, the Matthews correlation coefficient (MCC) was 91.27% and, in particular, the area under the ROC curve reached 98.07%. Among the machine learning models based on AAC and DPC representations, the best model is GBDT, with Acc and AUC of 85.19% and 85.16%. TL-AHB improves Acc and AUC by 10.8% and 12.91% compared to the former. Among the pre-training models, the best-performing model is the AHB model on the SR-MMP dataset, with Acc and AUC of 83.81% and 91.13%, and TL-AHB has an Acc and AUC improvement of 12.18% and 6.94% over this model. According to the experimental results, it can be seen that TL-AHB shows a strong predictive ability and excellent stability in predicting CCK-secretory peptides after transfer learning.

3.6. Analysis of the Prediction Performance of Similar Sequences on the Model

To investigate the model’s predictive ability for some similar CCK-secretory peptide sequences, some similar CCK-secretory peptide sequences of known classes were selected to validate the model. Each peptide sequence was entered into the model and predicted 50 times, and the model prediction performance is shown in Table 6. In Table 6, some peptides exhibited completely different activities, despite their extreme sequence similarity, with some peptides even differing by only one amino acid sequence. The average accuracy of our proposed method on similar sequences is about 92.36%. HAN assigns different weights to the features of peptide sequences, which helps the model to focus on the most critical features step by step, so that the model still performs well for some sequences with extremely high similarity. However, some samples also have a poor predictive performance, with accuracy rates of only 86% for RYLG and 82% for RYPS, respectively. The amino acid sequence at both ends of the peptide has an important influence on the activity, so the activity prediction of similar sequences of CCK-secretory peptides is a difficult point in this study, and there is still room for improvement, to overcome this difficult point in the future.

4. Conclusions

In this study, a BiLSTM model (TL-AHB) based on transfer learning and a hierarchical attention network mechanism is proposed, for the first time, to predict CCK-secretory peptides. The SMILES enumeration method is used to enhance the SMILES, and the transfer learning strategy is combined to pre-train and fine-tune the model parameters. The information on the peptide chain is fully learned, and effective and ineffective peptides are recognized using the TL-AHB model. The experimental results demonstrate the excellent performance and robustness of the model. Of course, some difficulties still need to be overcome in the future. The model cannot measure the strength of the CCK-secretion-stimulating effect of peptides, and there are still broad research prospects for this topic. As an efficient and low-cost tool, the TL-AHB prediction model is expected to be used to predict and research CCK-secretory peptides. In addition, user-friendly and publicly accessible web servers are becoming a popular trend. This study aims to help biological researchers more easily identify CCK-secretory peptides, and we hope to provide a web server for the prediction method proposed in this paper in the future. A robust online web server can significantly enhance the influence of the TL-AHB model.

Author Contributions

J.L. and P.C. conceived the algorithm, carried out the analysis, carried out the experiments, and wrote the manuscript. H.S. and X.G. designed, performed, and analyzed the experiments, and reviewed and edited the manuscript. P.Z., M.W. and Z.S. prepared the dataset. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Key Research and Development Program (2022YFF1100102), the National Natural Science Foundation of China (32172247), and the Major Project of Inner Mongolia Science and Technology Department (2021ZD0002).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are available from the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Moran, T.H. Gut Peptides in the Control of Food Intake. Int. J. Obes. 2009, 33, S7–S10. [Google Scholar] [CrossRef] [PubMed]
  2. Crozier, S.J.; Williams, J.A. Dietary Protein Stimulates Pancreatic Growth via CCK-independent MTOR Activation. FASEB J. 2009, 23, 984.5. [Google Scholar] [CrossRef]
  3. Conwell, D. Pancreatic Acinar Cell Function in Minimal Change Pancreatitis (MCP): A Clinical Application of the Endoscopic Cholecystokinin (CCK) Stimulated Pancreatic Function Test (EPFT). Am. J. Gastroenterol. 2003, 98, S69–S70. [Google Scholar] [CrossRef]
  4. Myers, S.I.; Bartula, L.L.; Colvin, M.P.; Parkman, H.P. Cholecystokinin (CCK) down Regulates PGE2 and PGI2 Release in Inflamed Guinea Pig Gallbladder Smooth Muscle Cell Cultures. Prostaglandins Leukot. Essent. Fat. Acids 2005, 73, 121–126. [Google Scholar] [CrossRef] [PubMed]
  5. Tirassa, P.; Manni, L.; Aloe, L.; Lundeberg, T. Cholecystokinin-8 and Nerve Growth Factor: Two Endogenous Molecules Working for the Upkeep and Repair of the Nervous System. CDTCNSND 2002, 1, 495–510. [Google Scholar] [CrossRef]
  6. Su, Y.; Liu, N.; Zhang, Z.; Li, H.; Ma, J.; Yuan, Y.; Shi, M.; Liu, J.; Zhao, Z.; Zhang, Z.; et al. Cholecystokinin and Glucagon-like Peptide-1 Analogues Regulate Intestinal Tight Junction, Inflammation, Dopaminergic Neurons and α-Synuclein Accumulation in the Colon of Two Parkinson’s Disease Mouse Models. Eur. J. Pharmacol. 2022, 926, 175029. [Google Scholar] [CrossRef]
  7. Awad, C.; McGovern, C.; Matters, G.; Smith, J.P. Endogenous Cholecystokinin (CCK) Expression in Human Pancreatic Cancer. Pancreatology 2013, 13, e14–e15. [Google Scholar] [CrossRef]
  8. Evers, B.M.; Gomez, G.; Townsend, C.M.; Rajaraman, S.; Thompson, J.C. Endogenous Cholecystokinin Regulates Growth of Human Cholangiocarcinoma. Ann. Surg. 1989, 210, 317–323. [Google Scholar] [CrossRef]
  9. Mader, J.S.; Hoskin, D.W. Cationic Antimicrobial Peptides as Novel Cytotoxic Agents for Cancer Treatment. Expert. Opin. Investig. Drugs 2006, 15, 933–946. [Google Scholar] [CrossRef]
  10. Tzima, C.S.; Banti, C.N.; Hadjikakou, S.K. Assessment of the Biological Effect of Metal Ions and Their Complexes Using Allium Cepa and Artemia Salina Assays: A Possible Environmental Implementation of Biological Inorganic Chemistry. J. Biol. Inorg. Chem. 2022, 27, 611–629. [Google Scholar] [CrossRef]
  11. Barras, D.; Widmann, C. Promises of Apoptosis-Inducing Peptides in Cancer Therapeutics. CPB 2011, 12, 1153–1165. [Google Scholar] [CrossRef] [PubMed]
  12. Marqus, S.; Pirogova, E.; Piva, T.J. Evaluation of the Use of Therapeutic Peptides for Cancer Treatment. J. Biomed. Sci. 2017, 24, 21. [Google Scholar] [CrossRef] [PubMed]
  13. Conlon, J.M.; Walsh, J.H.; Dockray, G.J. Gut Peptides: Biochemistry and Physiology. Q. Rev. Biol. 1996, 71, 453–454. [Google Scholar] [CrossRef]
  14. Daly, K.; Al-Rammahi, M.; Moran, A.; Marcello, M.; Ninomiya, Y.; Shirazi-Beechey, S.P. Sensing of Amino Acids by the Gut-Expressed Taste Receptor T1R1-T1R3 Stimulates CCK Secretion. Am. J. Physiol. Gastrointest. Liver Physiol. 2013, 304, G271–G282. [Google Scholar] [CrossRef]
  15. Tian, M.; Heng, J.; Song, H.; Zhang, Y.; Chen, F.; Guan, W.; Zhang, S. Branched Chain Amino Acids Stimulate Gut Satiety Hormone Cholecystokinin Secretion through Activation of the Umami Taste Receptor T1R1/T1R3 Using an in Vitro Porcine Jejunum Model. Food Funct. 2019, 10, 3356–3367. [Google Scholar] [CrossRef]
  16. Santos-Hernández, M.; Amigo, L.; Recio, I. Induction of CCK and GLP-1 Release in Enteroendocrine Cells by Egg White Peptides Generated during Gastrointestinal Digestion. Food Chem. 2020, 329, 127188. [Google Scholar] [CrossRef]
  17. Jaeger, S.; Fulle, S.; Turk, S. Mol2vec: Unsupervised Machine Learning Approach with Chemical Intuition. J. Chem. Inf. Model. 2018, 58, 27–35. [Google Scholar] [CrossRef]
  18. Jo, J.; Kwak, B.; Choi, H.-S.; Yoon, S. The Message Passing Neural Networks for Chemical Property Prediction on SMILES. Methods 2020, 179, 65–72. [Google Scholar] [CrossRef]
  19. Savittri, S.A.; Amalia, A.; Budiman, M.A. A Relevant Document Search System Model Using Word2vec Approaches. J. Phys. Conf. Ser. 2021, 1898, 012008. [Google Scholar] [CrossRef]
  20. Sunjaya, B.A.; Permai, S.D.; Gunawan, A.A.S. Forecasting of Covid-19 Positive Cases in Indonesia Using Long Short-Term Memory (LSTM). Procedia Comput. Sci. 2023, 216, 177–185. [Google Scholar] [CrossRef]
  21. Han, D.; Liu, P.; Xie, K.; Li, H.; Xia, Q.; Cheng, Q.; Wang, Y.; Yang, Z.; Zhang, Y.; Xia, J. An Attention-Based LSTM Model for Long-Term Runoff Forecasting and Factor Recognition. Environ. Res. Lett. 2023, 18, 024004. [Google Scholar] [CrossRef]
  22. Liu, Y.; Zhang, R.; Li, T.; Jiang, J.; Ma, J.; Wang, P. MolRoPE-BERT: An Enhanced Molecular Representation with Rotary Position Embedding for Molecular Property Prediction. J. Mol. Graph. Model. 2023, 118, 108344. [Google Scholar] [CrossRef] [PubMed]
  23. Kamaraj, B.; Al-Subaie, A.M.; Ahmad, F.; Surapaneni, K.M.; Alsamman, K. Effect of Novel Leukemia Mutations (K75E & E222K) on Interferon Regulatory Factor 1 and Its Interaction with DNA: Insights from Molecular Dynamics Simulations and Docking Studies. J. Biomol. Struct. Dyn. 2021, 39, 5235–5247. [Google Scholar] [CrossRef] [PubMed]
  24. Al-Subaie, A.M.; Kamaraj, B. The Structural Effect of FLT3 Mutations at 835th Position and Their Interaction with Acute Myeloid Leukemia Inhibitors: In Silico Approach. IJMS 2021, 22, 7602. [Google Scholar] [CrossRef]
  25. Dong, J.; Wang, N.-N.; Yao, Z.-J.; Zhang, L.; Cheng, Y.; Ouyang, D.; Lu, A.-P.; Cao, D.-S. ADMETlab: A Platform for Systematic ADMET Evaluation Based on a Comprehensively Collected ADMET Database. J. Cheminform. 2018, 10, 29. [Google Scholar] [CrossRef]
  26. Xiong, G.; Wu, Z.; Yi, J.; Fu, L.; Yang, Z.; Hsieh, C.; Yin, M.; Zeng, X.; Wu, C.; Lu, A.; et al. ADMETlab 2.0: An Integrated Online Platform for Accurate and Comprehensive Predictions of ADMET Properties. Nucleic Acids Res. 2021, 49, W5–W14. [Google Scholar] [CrossRef]
  27. Bhasin, M.; Raghava, G.P.S. Classification of Nuclear Receptors Based on Amino Acid Composition and Dipeptide Composition. J. Biol. Chem. 2004, 279, 23262–23266. [Google Scholar] [CrossRef]
  28. Rao, H.B.; Zhu, F.; Yang, G.B.; Li, Z.R.; Chen, Y.Z. Update of PROFEAT: A Web Server for Computing Structural and Physicochemical Features of Proteins and Peptides from Amino Acid Sequence. Nucleic Acids Res. 2011, 39, W385–W390. [Google Scholar] [CrossRef]
  29. Weininger, D. SMILES, a Chemical Language and Information System. 1. Introduction to Methodology and Encoding Rules. J. Chem. Inf. Comput. Sci. 1988, 28, 31–36. [Google Scholar] [CrossRef]
  30. Zhai, J.; Qi, J.; Zhang, S. Imbalanced Data Classification Based on Diverse Sample Generation and Classifier Fusion. Int. J. Mach. Learn. Cyber. 2022, 13, 735–750. [Google Scholar] [CrossRef]
  31. Bjerrum, E.J. SMILES Enumeration as Data Augmentation for Neural Network Modeling of Molecules. arXiv Preprint 2017, arXiv:1703.07076. [Google Scholar] [CrossRef]
  32. Hochreiter, S.; Schmidhuber, J. Long Short-Term Memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
  33. Teng, F.; Yang, W.; Chen, L.; Huang, L.; Xu, Q. Explainable Prediction of Medical Codes with Knowledge Graphs. Front. Bioeng. Biotechnol. 2020, 8, 867. [Google Scholar] [CrossRef] [PubMed]
  34. Yang, Z.; Yang, D.; Dyer, C.; He, X.; Smola, A.; Hovy, E. Hierarchical Attention Networks for Document Classification. In Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, San Diego, CA, USA, 12–17 June 2016; pp. 1480–1489. [Google Scholar] [CrossRef]
  35. Weiss, K.; Khoshgoftaar, T.M.; Wang, D. A Survey of Transfer Learning. J. Big Data 2016, 3, 9. [Google Scholar] [CrossRef]
  36. Santos-Hernández, M.; Vivanco-Maroto, S.M.; Miralles, B.; Recio, I. Food Peptides as Inducers of CCK and GLP-1 Secretion and GPCRs Involved in Enteroendocrine Cell Signalling. Food Chem. 2023, 402, 134225. [Google Scholar] [CrossRef]
  37. Theysgeur, S.; Cudennec, B.; Deracinois, B.; Perrin, C.; Guiller, I.; Lepoudère, A.; Flahaut, C.; Ravallec, R. New Bioactive Peptides Identified from a Tilapia Byproduct Hydrolysate Exerting Effects on DPP-IV Activity and Intestinal Hormones Regulation after Canine Gastrointestinal Simulated Digestion. Molecules 2020, 26, 136. [Google Scholar] [CrossRef]
  38. Tulipano, G.; Faggi, L.; Cacciamali, A.; Caroli, A.M. Whey Protein-Derived Peptide Sensing by Enteroendocrine Cells Compared with Osteoblast-like Cells: Role of Peptide Length and Peptide Composition, Focussing on Products of β-Lactoglobulin Hydrolysis. Int. Dairy J. 2017, 72, 55–62. [Google Scholar] [CrossRef]
  39. Osborne, S.; Chen, W.; Addepalli, R.; Colgrave, M.; Singh, T.; Tran, C.; Day, L. In Vitro Transport and Satiety of a Beta-Lactoglobulin Dipeptide and Beta-Casomorphin-7 and Its Metabolites. Food Funct. 2014, 5, 2706–2718. [Google Scholar] [CrossRef]
Figure 1. SMILES enumeration enables data augmentation. One SMILES is ten-fold augmented to obtain ten different SMILESs.
Figure 1. SMILES enumeration enables data augmentation. One SMILES is ten-fold augmented to obtain ten different SMILESs.
Biomolecules 13 01372 g001
Figure 2. BiLSTM structure diagram. Xn represents the input, Hn represents the output, hn is the hidden layer vector, and Cn is the cell state at the moment n.
Figure 2. BiLSTM structure diagram. Xn represents the input, Hn represents the output, hn is the hidden layer vector, and Cn is the cell state at the moment n.
Biomolecules 13 01372 g002
Figure 3. Process diagram of transfer learning.
Figure 3. Process diagram of transfer learning.
Biomolecules 13 01372 g003
Figure 4. Flowchart of our proposed method. HAN: hierarchical attention network; ML: machine learning; PT: pre-training; TL: transfer learning.
Figure 4. Flowchart of our proposed method. HAN: hierarchical attention network; ML: machine learning; PT: pre-training; TL: transfer learning.
Biomolecules 13 01372 g004
Figure 5. Bar plot showing the AAC percentage of CCK-secretory peptides.
Figure 5. Bar plot showing the AAC percentage of CCK-secretory peptides.
Biomolecules 13 01372 g005
Figure 6. Trend of AUC with SMILES augmentation times.
Figure 6. Trend of AUC with SMILES augmentation times.
Biomolecules 13 01372 g006
Table 1. Details of the source dataset and target dataset used in this study.
Table 1. Details of the source dataset and target dataset used in this study.
DatasetNameTotalPositiveNegative
Source datasetDILI467235232
Pgb-sub1185586599
SR-MMP1840892948
Target datasetCCK995445
Table 2. Performance comparison with different machine learning models developed based on AAC.
Table 2. Performance comparison with different machine learning models developed based on AAC.
ModelSen (%)Spe (%)Acc (%)MCC (%)AUC (%)
SVM61.5471.4366.6733.1566.48
RF76.9264.2970.3741.4470.60
MLP69.2357.1462.9619.5659.62
KNN61.5450.0055.5611.6055.77
GBDT84.6285.7185.1970.3385.16
XGBoost76.9271.4374.0748.3574.18
Sen: sensitivity, Spe: specificity, Acc: accuracy, MCC: Matthews correlation coefficient, AUC: the area under the receiver operating characteristic curve.
Table 3. Performance comparison with different machine learning models developed based on DPC.
Table 3. Performance comparison with different machine learning models developed based on DPC.
ModelSen (%)Spe (%)Acc (%)MCC (%)AUC (%)
SVM61.5418.1857.1430.9359.09
RF80.0027.2752.388.5353.64
MLP60.0036.3647.623.7448.18
KNN60.0027.2742.8613.4843.64
GBDT90.009.0947.621.5549.55
XGBoost40.0072.7357.1413.4856.36
Sen: sensitivity, Spe: specificity, Acc: accuracy, MCC: Matthews correlation coefficient, AUC: the area under the receiver operating characteristic curve.
Table 4. Performance of pre-training models on different source datasets.
Table 4. Performance of pre-training models on different source datasets.
DatasetModelSen (%)Spe (%)Acc (%)MCC (%)AUC (%)
DILIHL77.3384.8080.4660.4785.84
AHL78.2886.5882.4165.1287.88
HB78.0985.9181.5961.8986.06
AHB85.6681.1683.3467.1990.05
pgb-subHL79.3770.0173.9248.8083.17
AHL79.9981.2080.6560.0288.16
HB75.5077.4375.7652.0985.42
AHB85.5480.0882.2864.5989.90
SR-MMPHL80.1082.2081.0461.9688.60
AHL78.0688.9883.2766.9390.86
HB79.3883.4981.7063.1889.47
AHB79.7787.9783.8167.6691.13
A: using data augmentation; H: hierarchical attention network (HAN). L: LSTM. B: BiLSTM. Sen: sensitivity, Spe: specificity, Acc: accuracy, MCC: Matthews correlation coefficient, AUC: the area under the receiver operating characteristic curve.
Table 5. Performance of different transfer learning models on the target dataset.
Table 5. Performance of different transfer learning models on the target dataset.
ModelSen (%)Spe (%)Acc (%)MCC (%)AUC (%)
TL-HL74.3682.9380.1755.9483.58
TL-AHL70.7795.7387.6971.1291.88
TL-HB66.6788.1682.1459.5887.87
TL-AHB90.7198.9295.9991.2798.07
TL: transfer learning. A: using data augmentation; H: hierarchical attention network (HAN). L: LSTM. B: BiLSTM. Sen: sensitivity, Spe: specificity, Acc: accuracy, MCC: Matthews correlation coefficient, AUC: the area under the receiver operating characteristic curve.
Table 6. Prediction performance of similar sequences using the model.
Table 6. Prediction performance of similar sequences using the model.
CCK-secretory Peptide SequenceClass LabelAccuracy (%)
RYLGY [36]196
SRYPS [36]190
IRGCRL [36]194
RYLG [36]086
RYPS [36]082
RGCRL [36]094
LEL [37]096
LLP [16]096
LL [38]096
IPI [38]092
GPI [39]094
Class label 1: effective peptides, Class label 0: ineffective peptides.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liu, J.; Chen, P.; Song, H.; Zhang, P.; Wang, M.; Sun, Z.; Guan, X. Prediction of Cholecystokinin-Secretory Peptides Using Bidirectional Long Short-term Memory Model Based on Transfer Learning and Hierarchical Attention Network Mechanism. Biomolecules 2023, 13, 1372. https://doi.org/10.3390/biom13091372

AMA Style

Liu J, Chen P, Song H, Zhang P, Wang M, Sun Z, Guan X. Prediction of Cholecystokinin-Secretory Peptides Using Bidirectional Long Short-term Memory Model Based on Transfer Learning and Hierarchical Attention Network Mechanism. Biomolecules. 2023; 13(9):1372. https://doi.org/10.3390/biom13091372

Chicago/Turabian Style

Liu, Jing, Pu Chen, Hongdong Song, Pengxiao Zhang, Man Wang, Zhenliang Sun, and Xiao Guan. 2023. "Prediction of Cholecystokinin-Secretory Peptides Using Bidirectional Long Short-term Memory Model Based on Transfer Learning and Hierarchical Attention Network Mechanism" Biomolecules 13, no. 9: 1372. https://doi.org/10.3390/biom13091372

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop