Next Article in Journal
Discovery of Novel 3,4-Dihydro-2(1H)-Quinolinone Sulfonamide Derivatives as New Tubulin Polymerization Inhibitors with Anti-Cancer Activity
Previous Article in Journal
Perspectives and Prospective on Solid Lipid Nanoparticles as Drug Delivery Systems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

ACPNet: A Deep Learning Network to Identify Anticancer Peptides by Hybrid Sequence Information

1
Key Laboratory of Symbol Computation and Knowledge Engineering of Ministry of Education, College of Computer Science and Technology, Jilin University, Changchun 130012, China
2
College of Computer Science and Technology, Tonghua Normal University, Tonghua 134000, China
3
School of Computer Science and Artificial Intelligence Aliyun School of Big Data School of Software, Changzhou University, Changzhou 213164, China
4
College of Software, Jilin University, Changchun 130012, China
*
Author to whom correspondence should be addressed.
Molecules 2022, 27(5), 1544; https://doi.org/10.3390/molecules27051544
Submission received: 3 January 2022 / Revised: 20 February 2022 / Accepted: 23 February 2022 / Published: 24 February 2022

Abstract

:
Cancer is one of the most dangerous threats to human health. One of the issues is drug resistance action, which leads to side effects after drug treatment. Numerous therapies have endeavored to relieve the drug resistance action. Recently, anticancer peptides could be a novel and promising anticancer candidate, which can inhibit tumor cell proliferation, migration, and suppress the formation of tumor blood vessels, with fewer side effects. However, it is costly, laborious and time consuming to identify anticancer peptides by biological experiments with a high throughput. Therefore, accurately identifying anti-cancer peptides becomes a key and indispensable step for anticancer peptides therapy. Although some existing computer methods have been developed to predict anticancer peptides, the accuracy still needs to be improved. Thus, in this study, we propose a deep learning-based model, called ACPNet, to distinguish anticancer peptides from non-anticancer peptides (non-ACPs). ACPNet employs three different types of peptide sequence information, peptide physicochemical properties and auto-encoding features linking the training process. ACPNet is a hybrid deep learning network, which fuses fully connected networks and recurrent neural networks. The comparison with other existing methods on ACPs82 datasets shows that ACPNet not only achieves the improvement of 1.2% Accuracy, 2.0% F1-score, and 7.2% Recall, but also gets balanced performance on the Matthews correlation coefficient. Meanwhile, ACPNet is verified on an independent dataset, with 20 proven anticancer peptides, and only one anticancer peptide is predicted as non-ACPs. The comparison and independent validation experiment indicate that ACPNet can accurately distinguish anticancer peptides from non-ACPs.

1. Introduction

Currently, cancer is an enormous threat to human health, reported by the International Agency for Research on Cancer (IARC) [1], leading to rapidly increasing mortality every year, reaching 18.1 million new cases and 9.6 million cancer deaths in 2018 alone. Moreover, the type of common cancer is increasing, including lung, breast, prostate, colorectal and so on. However, the diagnosis and therapy of cancer are challenging, which hinders the therapy process, such as surgery, radiotherapy, chemotherapy, and targeted therapy. Furthermore, traditional methods for the treatment of cancer always meet costly and high-risk problems, accompanied by the risk of tissue death, drug resistance and other serious side effects [2]. Fortunately, increasing evidence shows that some novel anticancer agents, for some peptides for instance, become novel and safe for therapy [3]. For example, peptide p28 is a post-translational, multi-target anticancer agent that preferentially enters a wide variety of solid tumor cells and binds to both wild-type and mutant p53 protein, inhibiting constitutional morphogenic protein 1 (Cop1)-mediated ubiquitination and proteasomal degradation of p53, which results in increased levels of p53 and induces cell-cycle arrest at G2/M and eventual apoptosis, which results in tumor cell shrinkage and death [4]. LL-37, a novel anticancer peptide, has a net positive charge and is amphiphilic, and can eliminate pathogenic microbes directly, via electrostatic attraction towards negatively charged bacterial membranes. Several studies have shown that LL-37 participates in various host immune systems, such as inflammatory responses and tissue repair [5]. Peptide-based therapeutics are more affordable, tolerable and safe, which are considered to be advanced therapy strategies [6]. Therefore, it is a significant step to rapidly and effectively find useful anticancer peptides (ACPs). Although there are many experimental methods to identify ACPs, they are usually laborious, expensive, time consuming and hard to achieve in a high-throughput manner. Furthermore, with the rapid development of data science, it is very desirable to design machine learning-based methods to identify ACPs [7].
Over the past few years, a dozen computational methods have been proposed to identify ACPs, including k-nearest neighbor (KNN), support vector machine (SVMs), random forest (RF) and so on. In 2019, Boopathi, V. et al. proposed a machine learning model to predict ACPs, called mACPpred [7], which uses seven types of encoding features, including amino acid composition (AAC), dipeptide composition (DPC), composition-transition-distribution (CTD), quasi-sequence-order (QSO), amino acid index (AAIF), binary profile (NC5) and conjoint triad (CTF) to represent a peptide sequence and cooperate with an SVM model to predict ACPs. In 2020, Li Qingwen, et al. employed five types of peptide sequence features, including amino acid composition (AAC), conjoint triad (CT), pseudo-amino acid composition (PAAC), grouped amino acid composition (GAAC) and C/T/D, and then fused multiple machine learning methods, containing SVM, RF, and LibD3C, to identify ACPs [8]. In 2020, Ge Ruiquan et al. proposed a machine model called EnACP, which introduces sequence composition, sequence-order, physicochemical properties, etc. to encode a peptide sequence and input the important feature selected by multiple ensemble classifiers to an SVM model to predict ACPs [9]. These methods try to find effective and useful features to represent a peptide and combine a high-performance machine model to identify ACPs. Another way to identify peptides is by introducing a deep learning model to identify ACPs from the raw sequence of peptides. In 2019,Yi Haichent et al. proposed a deep learning long short-term memory (LSTM) neural network model, called ACP-DL [10], which developed an efficient feature representation approach by integrating binary profile features, k-mer sparse matrix of the reduced amino acid alphabet and then implemented a deep LSTM model to identify ACPs. In 2021, Chen Xiangan et al. proposed an ACP prediction model, called ACP-DA [11], which uses data augmentation for insufficient samples and trains a multilayer perception model to improve the prediction performance. In 2020, Yu Lezheng et al. found that the recurrent neural network with bidirectional long short-term memory cells is a superior architecture to identify ACPs and implement a sequence-based deep learning tool, called DeepACP [12], to accurately predict ACPs.
Although these methods can predict ACPs accurately, the accuracy performance of existing methods still needs to be improved. It is also a challenge to represent peptide sequences to numerical vectors and further improve the prediction accuracy of ACPs. Therefore, in this paper, we propose a hybrid deep learning-based model, called ACPNet, which employs the raw peptide sequence and carefully selected sequence features as input, to fit recurrent neural networks, and fully connected network to further improve the predicting performance.

2. Materials and Methods

2.1. Materials

The sequences of ACP and non-ACP are downloaded from this research used in [12]. Three datasets are introduced including ACPs250, ACPs82, ACPs20. ACPs250 contains 250 ACPs and 250 non-ACPs sequence samples, ACPs82 is made up of 82 ACPs and 82 non-ACPs sequence samples, ACPs20 contains 10 ACP samples and 10 non-ACP samples. ACPs250 dataset is split into training dataset and validation dataset with 80% and 20% respectively. To further validate the performance of ACPNet, we conducted experiments on an independent test dataset ACPs82. Furthermore, ACPs20, another independent dataset, is introduced to further prove the performance of ACPNet. These datasets are listed in Table 1.

2.2. Methods

2.2.1. Features Construction

To further improve the prediction accuracy of ACPs, in this work, we employed three hybrid kinds of features to encode a peptide sequence to a numerical vector, including peptide sequence features, peptide physicochemical properties and automatic embedding features, which are listed in Table 2.

Sequence Features

The concept of PAAC (pseudo amino acid composition) [13] was introduced to avoid completely losing the sequence-order information [1]. In contrast with the conventional amino acid composition (AAC) that contains 20 components, the PAAC contains a set of greater than 20 discrete factors, where the first 20 represent the components of its conventional amino acid composition while the additional factors incorporate some sequence-order information via various pseudo components.
PAAC can be represented by P = [ p 1 ,   p 2 ,   ,   p 20 ,   p 20 + 1 ,   ,   p 20 + λ ] T ,   ( λ   <   L ) , where λ is an integrated parameter set by a user with recommended value 10, p u is calculated by Equation (1)
p u = { f u i = 1 20 f i + w k = 1 λ τ k ,   ( i u 20 )                               w τ u - 20 i = 1 20 f i + w k = 1 λ τ k ,   ( 21 u 20 + λ )
where f i is the frequency of each amino acid in a protein, w is the weight factor set by a user with default value 0.05, τ k is the k-th tier correlation factor that reflects the sequence order correlation between all the k-th most contiguous residues as formulated by Equation (2).
τ k = 1 L - k i = 1 L - k J i , i + k               ( k < L )
where J i , i + k can be calculated by Equation (3).
J i , i + k = 1 Γ q = 1 Γ [ Φ q ( R i + k ) - Φ q ( R i ) ] 2
where Φ q ( R i ) is the q-th function of the amino acid R i and Γ the total number of functions considered, such as hydrophobicity value, hydrophilicity value, and side-chain mass of amino acid.
Another two features are peptide sequence length and Shannon entropy of peptide sequences [14]. The Shannon entropy of peptide sequences can be obtained by Equation (4).
Shannon   entropy ( seq ) = i = 1 20 f i log ( f i )

Peptide Physicochemical Properties

Peptide, a short chain of amino acids, exhibits many similar properties to proteins and the physicochemical properties have close relations with protein functions. Therefore, three physicochemical properties of peptides are introduced to represent peptide sequences including Gravy [15], Molecular_weight [16] and Charge_at_pH [17]. Gravy feature is used to describe peptide gravy according to Kyte and Doolittle, Molecular_weight is employed to represent peptide molecular weight, Charge_at_pH is introduced to calculate the charge number of a protein when given pH set 10.

Embedding Features

A peptide can be seen as a sentence containing ‘word’ as ‘A’, ‘C’, ‘D’, ‘E’, ‘F’, ‘G’, ‘H’, ‘I’, ‘K’, ‘L’, ‘M’, ‘N’, ‘P’, ‘Q’, ‘R’, ‘S’, ‘T’, ‘V’, ‘W’, ‘Y’, which represent 20 kinds of amino acids. To encode a peptide sequence to a numerical vector, a peptide is converted to a numerical vector by amino acid index with the same length of peptide length such as [ 1 , 2 , 4 , , 20 , 5 ] . The corresponding map is A 1 ,   C 2 ,   ,   Y 20 . After a peptide is converted to a numerical vector, the following step is to map each index to a user-defined dimension vector. The embedding process is to turn positive integers (indexes) into dense vectors. For a peptide sequence “ADGF” is an example with a user-defined dimension as three, the representation process is shown in Figure 1.

2.2.2. Model Structure

Overall Workflow

Deep learning technology has obtained numerous achievements in many bioinformatics applications [18,19,20]. Therefore, in this paper, we propose a hybrid deep learning-based model, named ACPNet, for predicting ACPs. The overall workflow is shown in Figure 2. The detailed structure of ACPNet is in Supplementary Figure S1.
The manually selected features and auto-embedding features are fed to a fully connected neural network (Dense Network) and recurrent neural network (RNN) [21] respectively, and then the result of the previous process is merged for the final prediction. Overall, ACPNet combines Dense Network and RNN to construct a hybrid deep learning-based model to identify ACPs, which not only consider the importance of manually selected features but also automatically learn the potential features from raw peptide sequences as well.

Prediction Model Constructed by RNN and Dense Networks

RNN and Dense network, as the two most significant deep learning models, are widely applied on time serial and feature independent issues respectively. LSTM [22], a kind of implementation of RNN, is usually employed to further automatically extract potential features from the time serial vectors. The kernel process of LSTM can be illustrated by Equation (5).
f t = σ g ( W f x t + U f h t - 1 + b f ) i t = σ g ( W i x t + U i h t - 1 + b i ) ο t = σ g ( W ο x t + U ο h t - 1 + b ο ) c t = f t c t - 1 + i t σ c ( W c x t + U c h t - 1 + b c ) h t = ο t σ h ( c t )
where, x t is the input vector, h t is output vector, c t is the cell state vector, W , U and b are the learning parameters, f t is a forget gate vector to remember old information, i t is input gate vector to acquire new information, ο t is output gate vector as output candidate, σ g , σ c and σ h are three activate functions. LSTM is employed in ACPNet for the belief that the peptide sequence can be seen as time series data. The dense network processes 1-dimension data with independent features expertly, therefore a Dense network is employed to process manual-selected features and the final prediction.

Implementation of ACPNet

ACPNet is implemented by Tensorflow 2.5.0 and all scripts are written by Python 3.8. ACPNet is running on a personal computer with 4.3 GHz, 8 core CPU, and 64 GB RAM under an open Linux operating system.

2.2.3. Performance Evaluation of ACPNet

ACPNet is evaluated by the widely used standard performance metrics, which are Accuracy, F1-score, Recall, Precision and Matthews correlation coefficient (MCC). These evaluation metrics are defined as follows:
Accuracy = TP + TN TP + TN + FP + FN F 1 - score = 2 TP 2 TP + FP + FN Recall = TP TP + FN Precision = TP TP + FP Matthews Correlation coefficient ( MCC ) = TP TN - FP FN ( TP + FN ) ( TP + FP ) ( TN + FP ) ( TN + FN )
where TP, FP, TN and FN represent the true positives, false positives, true negatives and false negatives, respectively. We also plot the receiver operating characteristic curves (ROC) and computed Area Under the Curve (AUC) to show the performance of ACPNet.

3. Results

3.1. The Effects of Feature Combination

To explore the effect of the combination of manually selected features and automatic learning features, the performances conducted by three types of combination are compared on ACPs250 (as training dataset), and ACPs82 (as test dataset). The compared results are listed in Table 3, and we found that the hybrid-fused features model shows better performance on multiple metrics. Note, in terms of MCC, the feature-fused model surpasses the manually selected feature model by more than 16%, and on other metrics, the feature-fused model shows advanced performance as well. The results indicate that the combination of two types of features play a positive and reinforced role in distinguishing between ACPs and non-ACPs.

3.2. Manually Selected Features Importance Rank

To further show the importance of each manually selected feature, CatBoost [23], an ensemble machine learning framework, was introduced to calculate the importance score of each feature. Figure 3 shows the importance score of each manually selected feature. Length, Gravy, MW (Molecular_weight), and SH (Shannon entropy) obtained a relatively high score and surpassed the majority of PAAC’s features. The PAAC features also show a positive contribution. The PAAC features get the best score at the 14th feature and show not particularly important contributions at the 18th feature. Besides, each PAAC feature provides different contributions to identify ACP. Overall, the manually selected features contribute to the classification of ACPs and non-ACPs.

3.3. Feature Visualization

We use Uniform Manifold Approximation and Projection (UMAP) [24] to visualize the distribution of ACPs and non-ACPs by media vector, generated on the ACPNet inner layer into two-dimensional space. Figure 4 illustrates that ACPs and non-ACPs in the training and test dataset can be easily classified by these features, which reconfirms that the constructed features contribute to the identification of ACPs and non-ACPs.

3.4. Performance Comparison of Models on Independent Datasets

To show the advance of ACPNet, traditional machine learning- and deep learning-based methods are employed as comparisons. For a fair comparison, the same training dataset, ACP250s, and independent test dataset, ACP82s, were used to train and test all methods. For traditional machine learning, SVM, RF, and CatBoost are introduced for the comparisons. Note, the auto embedding features are replaced by the index encoding for peptide sequence because the auto embedding links the training process. Therefore, the index encoding method and manually selected features are employed to encode a peptide for traditional machine learning. The comparison results of traditional machine learning are listed in Table 4.
For the comparison with the deep learning-based model, seven existing models are employed as the comparisons, which include AntiCP [25], Hajisharifi [26], iACP [27] and ACPred-FL [28], CNN (Convolutional Neural Network) [29], CNN+RNN, DeepACP [12]. AntiCP contains two models, which are AntiCP_ACC and Anticp_DPC, both based on SVM. AntiCP_ACC is built by amino acid composition features, while Anticp_DPC is constructed by dipeptide composition features. Hajisharifi combines two integrative SVM-based classification models for the prediction of anticancer peptides on the base of local alignment kernel and PAAC parameters. Further, iACP employs g-gap dipeptide components and SVM to predict ACPs. CNN is fed one-hot encoding matrixes to identify ACPs. For the CNN+RNN model, the input is the same as the CNN model, but after the CNN part, an RNN model links the following process. DeepACP, an end-to-end method, tries to fuse the peptide sequence encoding and training process to identify the ACPs. In the same way as the comparison with traditional machine learning, the ACP250s dataset and independent test dataset, ACP82s, are used to train and test all methods. The comparison results of deep learning-based methods are listed in Table 5. From Table 5, we can find that ACPNet shows better performance compared with the other seven methods in multiple metrics. In terms of precision, ACPNet is slightly lower than ACPred-FL, but shows nearly 7% improvement in Recall and achieves a more balanced performance. For other methods, except ACPred-FL, ACPNet obtained more than 6%, 6%, 10%, 3.5%, and 12% improvement, in terms of accuracy, F1-score, recall, precision and MCC. Overall, ACPNet shows higher performance compared with existing methods.
Furthermore, we also plot the receiver operating characteristic curves (ROC) to further show the performance of ACPNet, shown in Figure 5, with AUC at 0.945 and PRAUC (Area Under the Precision–Recall Curve) [30] at 0.947, respectively. Figure 5 indicates that ACPNet shows better performance both on AUC and PRAUC.

3.5. Independent Validation

Furthermore, ACPNet is verified on an independent dataset, with 10 ACPs and 10 non-ACPs, and only one proven ACP is predicted as non-ACP, while all non-ACPs are predicted as non-ACPs. The independent validation results are listed in Table 6. Most ACPs obtained a higher score, large than 0.7. If the predicting score is larger than 0.5, the corresponding peptide will be treated as an ACP. The independent validation results indicate that ACPNet can accurately predict ACPs.

4. Discussion

Cancer is one of the most dangerous threats to human health. Anticancer peptides could be novel agents for the therapy of cancers [33]. Therefore, accurately identifying anticancer peptides is a key step for the therapy. Nowadays, although deep neural network models have been developed to predict ACP, the accuracy still needs to improve. Thus, in this study, we proposed a hybrid deep learning-based model, called ACPNet, to distinguish ACPs from non-ACPs. For the feature construction, three types of features were introduced. The first type of features are manually selected features, which include PAAC, Length, and Shannon entropy, calculated from peptide sequence information. The second type of features are Molecular_weight, Charge_at_pH, Gravy, which source from peptide physicochemical properties. The third type of features are autoencoding features, which link the training process by encoding each amino acid index to a vector. In part 3.3, the performance of the combination of three types of features is compared. The comparison results show that the combination of three types of features play a positive role in identifying ACPs. Each manually selected feature, including sequence information and peptide physicochemical properties features, is relatively independent, so a fully connected neural network is employed to train the manually selected features. Autoencoding features of peptides can be seen as time serial data, which fit the learning pattern of BiLSTM. Therefore, Sequence information and peptide physicochemical properties are merged to feed fully connected networks. Autoencoding features are input into a BiLSTM network [34]. After passing the two types of networks, the media vectors are merged to feed a fully connected network for the final prediction. ACPNet is a hybrid deep learning network, which fuses the advance of two types of networks to build the neural network structure and full use of three kinds of information features as inputs, to improve the prediction accuracy of ACPs. For a comparison with other existing methods, ACPNet not only shows higher performance in multiple metrics, including Accuracy, F1-score, Recall, Precision and MCC, but also shows balanced performance metrics. This means that ACPNet may show better robustness for future identification. Furthermore, media vectors generated on the ACPNet inner layer are compressed into two dimensions, to further show the entire performance directly. The visualization results indicate that three different types of features of peptides and hybrid deep learning-based models can accurately distinguish ACPs from non-ACPs. Furthermore, ACPNet is verified on an independent dataset, with 10 ACPs and 10 non-ACPs. Only one proven ACP was predicted as a non-ACP, and all the non-ACPs were predicted as non-ACPs. The independent validation results indicate that ACPNet can accurately distinguish ACPs from non-ACPs. The learning pattern of ACPNet also fits other peptide-related works, which may provide a useful clue in solving these problems. ACPNet also meets some limitations. For example, ACPNet does not provide a user-friendly web server, which may make it difficult for some people who don’t know how to program. Furthermore, ACPNet does not consider the different types of cancer, which may generate bias. In the future, we will try to build a new ACP prediction model, based on the different types of cancer, and provide a user-friendly web server to the public.

5. Conclusions

In this work, we proposed a deep learning-based method, ACPNet, to identify ACPs, by combining a hybrid deep learning-based model with manually selected features and automatic encoding features as the input. For the feature construction, three types of features were introduced, namely, peptide sequence component information features, peptide physicochemical properties and auto-encoding features. Three different types of features play a positive role in distinguishing ACPs and non-ACPs. A fully connected network and recurrent neural network were introduced to process the constructed features. Compared with existing methods, ACPNet shows better and more balanced performance, with 1.2% accuracy, 2.0% F1-score, and 7.2% Recall improvement. On a three-part dataset, with 10 ACPs and 10 non-ACPs, ACPNet accurately predicts nine in ten ACPs. The series experiments show that ACPNet can accurately distinguish ACPs from non-ACPs.

Supplementary Materials

The following supporting information can be downloaded online. Figure S1: detailed structure of ACPNet.

Author Contributions

M.S. conceived the algorithm, developed the program, and wrote the manuscript. M.S., X.H., S.Y. and Y.Z. helped with manuscript editing and design. X.H. prepared the datasets. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (Grants nos. 61772227, 61972174, 78361972175), Science and Technology Development Foundation of Jilin Province (nos. 20180201045GX, 20200201300JC, 20200401083GX, 20200201163JC), and the Paul K. and Diane Shumaker Endowment Fund to DX.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Publicly available datasets were analyzed in this study. Codes and data are available here: https://github.com/abcair/ACPNet accessed on 2 January 2022.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. WHO. International Agency for Research on Cancer; WHO: Geneva, Switzerland, 2019. [Google Scholar]
  2. Chiangjong, W.; Chutipongtanate, S.; Hongeng, S. Anticancer peptide: Physicochemical property, functional aspect and trend in clinical application (Review). Int. J. Oncol. 2020, 57, 678–696. [Google Scholar] [CrossRef] [PubMed]
  3. Xie, M.; Liu, D.; Yang, Y. Anti-cancer peptides: Classification, mechanism of action, reconstruction and modification. Open Biol. 2020, 10, 200004. [Google Scholar] [CrossRef] [PubMed]
  4. Yaghoubi, A.; Khazaei, M.; Avan, A.; Hasanian, S.M.; Cho, W.C.; Soleimanpour, S. P28 Bacterial Peptide, as an Anticancer Agent. Front. Oncol. 2020, 10, 1303. [Google Scholar] [CrossRef]
  5. Kuroda, K.; Okumura, K.; Isogai, H.; Isogai, E. The Human Cathelicidin Antimicrobial Peptide LL-37 and Mimics are Potential Anticancer Drugs. Front. Oncol. 2015, 5, 144. [Google Scholar] [CrossRef] [Green Version]
  6. Tyagi, A.; Kapoor, P.; Kumar, R.; Chaudhary, K.; Gautam, A.; Raghava, G.P.S. In Silico Models for Designing and Discovering Novel Anticancer Peptides. Sci. Rep. 2013, 3, srep02984. [Google Scholar] [CrossRef]
  7. Boopathi, V.; Subramaniyam, S.; Malik, A.; Lee, G.; Manavalan, B.; Yang, D.-C. mACPpred: A Support Vector Machine-Based Meta-Predictor for Identification of Anticancer Peptides. Int. J. Mol. Sci. 2019, 20, 1964. [Google Scholar] [CrossRef] [Green Version]
  8. Li, Q.; Zhou, W.; Wang, D.; Wang, S.; Li, Q. Prediction of Anticancer Peptides Using a Low-Dimensional Feature Model. Front. Bioeng. Biotechnol. 2020, 8, 892. [Google Scholar] [CrossRef]
  9. Ge, R.; Feng, G.; Jing, X.; Zhang, R.; Wang, P.; Wu, Q. EnACP: An Ensemble Learning Model for Identification of Anticancer Peptides. Front. Genet. 2020, 11, 760. [Google Scholar] [CrossRef]
  10. Yi, H.-C.; You, Z.-H.; Zhou, X.; Cheng, L.; Li, X.; Jiang, T.-H.; Chen, Z.-H. ACP-DL: A Deep Learning Long Short-Term Memory Model to Predict Anticancer Peptides Using High-Efficiency Feature Representation. Mol. Ther. Nucleic Acids 2019, 17, 1–9. [Google Scholar] [CrossRef] [Green Version]
  11. Chen, X.-G.; Zhang, W.; Yang, X.; Li, C.; Chen, H. ACP-DA: Improving the Prediction of Anticancer Peptides Using Data Augmentation. Front. Genet. 2021, 12, 1131. [Google Scholar] [CrossRef]
  12. Yu, L.; Jing, R.; Liu, F.; Luo, J.; Li, Y. DeepACP: A Novel Computational Approach for Accurate Identification of Anticancer Peptides by Deep Learning Algorithm. Mol. Ther.-Nucleic Acids 2020, 22, 862–870. Mol. Ther. Nucleic Acids 2020, 22, 862–870. [Google Scholar] [CrossRef] [PubMed]
  13. Shen, H.-B.; Chou, K.-C. PseAAC: A flexible web server for generating various kinds of protein pseudo amino acid composition. Anal. Biochem. 2008, 373, 386–388. [Google Scholar] [CrossRef] [PubMed]
  14. Strait, B.; Dewey, T. The Shannon information entropy of protein sequences. Biophys. J. 1996, 71, 148–155. [Google Scholar] [CrossRef] [Green Version]
  15. Sharma, R.; Dill, B.D.; Chourey, K.; Shah, M.; VerBerkmoes, N.C.; Hettich, R.L. Coupling a Detergent Lysis/Cleanup Methodology with Intact Protein Fractionation for Enhanced Proteome Characterization. J. Proteome Res. 2012, 11, 6008–6018. [Google Scholar] [CrossRef] [PubMed]
  16. Taguchi, T.; Kodera, Y.; Oba, K.; Saito, T.; Nakagawa, Y.; Kawashima, Y.; Shichiri, M. Suprabasin-derived bioactive peptides identified by plasma peptidomics. Sci. Rep. 2021, 11, 1–17. [Google Scholar] [CrossRef] [PubMed]
  17. García-Moreno, P.J.; Yang, J.; Gregersen, S.; Jones, N.C.; Berton-Carabin, C.C.; Sagis, L.M.; Hoffmann, S.V.; Marcatili, P.; Overgaard, M.T.; Hansen, E.B.; et al. The structure, viscoelasticity and charge of potato peptides adsorbed at the oil-water interface determine the physicochemical stability of fish oil-in-water emulsions. Food Hydrocoll. 2021, 115, 106605. [Google Scholar] [CrossRef]
  18. Min, S.; Lee, B.; Yoon, S. Deep learning in bioinformatics. Brief. Bioinform. 2016, 18, 851–869. [Google Scholar] [CrossRef] [Green Version]
  19. Li, Y.; Huang, C.; Ding, L.; Li, Z.; Pan, Y.; Gao, X. Deep learning in bioinformatics: Introduction, application, and perspective in the big data era. Methods 2019, 166, 4–21. [Google Scholar] [CrossRef] [Green Version]
  20. Li, H.; Tian, S.; Li, Y.; Fang, Q.; Tan, R.; Pan, Y.; Huang, C.; Xu, Y.; Gao, X. Modern deep learning in bioinformatics. J. Mol. Cell Biol. 2020, 12, 823–827. [Google Scholar] [CrossRef]
  21. Sherstinsky, A. Fundamentals of Recurrent Neural Network (RNN) and Long Short-Term Memory (LSTM) Network. Phys. D Nonlinear Phenom. 2020, 404, 132306. [Google Scholar] [CrossRef] [Green Version]
  22. Karim, F.; Majumdar, S.; Darabi, H.; Chen, S. LSTM Fully Convolutional Networks for Time Series Classification. IEEE Access 2017, 6, 1662–1669. [Google Scholar] [CrossRef]
  23. Huang, G.; Wu, L.; Ma, X.; Zhang, W.; Fan, J.; Yu, X.; Zeng, W.; Zhou, H. Evaluation of CatBoost method for prediction of reference evapotranspiration in humid regions. J. Hydrol. 2019, 574, 1029–1041. [Google Scholar] [CrossRef]
  24. McInnes, L.; Healy, J.; Melville, J. Umap: Uniform Manifold Approximation and Projection for Dimension Reduction. arXiv 2018, arXiv:1802.03426. [Google Scholar]
  25. Agrawal, P.; Bhagat, D.; Mahalwal, M.; Sharma, N.; Raghava, G.P.S. AntiCP 2.0: An updated model for predicting anticancer peptides. Brief. Bioinform. 2020, 22, bbaa153. [Google Scholar] [CrossRef]
  26. Hajisharifi, Z.; Piryaiee, M.; Beigi, M.M.; Behbahani, M.; Mohabatkar, H. Predicting anticancer peptides with Chou′s pseudo amino acid composition and investigating their mutagenicity via Ames test. J. Theor. Biol. 2014, 341, 34–40. [Google Scholar] [CrossRef]
  27. Akbar, S.; Hayat, M.; Iqbal, M.; Jan, M.A. iACP-GAEnsC: Evolutionary genetic algorithm based ensemble classification of anticancer peptides by utilizing hybrid feature space. Artif. Intell. Med. 2017, 79, 62–70. [Google Scholar] [CrossRef]
  28. ACPred-FL: A Sequence-Based Predictor Using Effective Feature Representation to Improve the Prediction of Anti-Cancer Peptides|Bioinformatics|Oxford Academic. Available online: https://academic.oup.com/bioinformatics/article/34/23/4007/5026665?login=true (accessed on 31 October 2021).
  29. Lawrence, S.; Giles, C.L.; Tsoi, A.C.; Back, A.D. Face recognition: A convolutional neural-network approach. IEEE Trans. Neural Netw. 1997, 8, 98–113. [Google Scholar] [CrossRef] [Green Version]
  30. Boyd, K.; Eng, K.H.; Page, C.D. Area under the Precision-Recall Curve: Point Estimates and Confidence Intervals. In Machine Learning and Knowledge Discovery in Databases; Blockeel, H., Kersting, K., Nijssen, S., Železný, F., Eds.; Springer: Berlin/Heidelberg, Germany, 2013; pp. 451–466. [Google Scholar]
  31. Grisoni, F.; Neuhaus, C.S.; Gabernet, G.; Müller, A.T.; Hiss, J.A.; Schneider, G. Designing Anticancer Peptides by Constructive Machine Learning. ChemMedChem 2018, 13, 1300–1302. [Google Scholar] [CrossRef]
  32. Rao, B.; Zhou, C.; Zhang, G.; Su, R.; Wei, L. ACPred-Fuse: Fusing multi-view information improves the prediction of anticancer peptides. Brief. Bioinform. 2019, 21, 1846–1855. [Google Scholar] [CrossRef]
  33. Triclosan Offers Protection against Blood Stages of Malaria by Inhibiting Enoyl-ACP Reductase of Plasmodium Falciparum|Nature Medicine. Available online: https://www.nature.com/articles/nm0201_167 (accessed on 31 October 2021).
  34. Sentiment Analysis of Comment Texts Based on BiLSTM|IEEE Journals & Magazine|IEEE Xplore. Available online: https://ieeexplore.ieee.org/abstract/document/8684825 (accessed on 31 October 2021).
Figure 1. The process to auto-embed a peptide sequence to a matrix.
Figure 1. The process to auto-embed a peptide sequence to a matrix.
Molecules 27 01544 g001
Figure 2. The overall workflow of ACPNet.
Figure 2. The overall workflow of ACPNet.
Molecules 27 01544 g002
Figure 3. The importance rank of manually selected features.
Figure 3. The importance rank of manually selected features.
Molecules 27 01544 g003
Figure 4. The importance rank of manually selected features.
Figure 4. The importance rank of manually selected features.
Molecules 27 01544 g004
Figure 5. AURoc (a) and PRRoc (b) on test dataset.
Figure 5. AURoc (a) and PRRoc (b) on test dataset.
Molecules 27 01544 g005
Table 1. Datasets for train, test and independent validation of ACPNet.
Table 1. Datasets for train, test and independent validation of ACPNet.
ACPs NumberNon-ACPs NumberAverage LengthMax LengthMin Length
ACPs250250250279711
ACPs8282822720711
ACPs201010244713
Table 2. Three kinds of hybrid features to encode peptide sequences.
Table 2. Three kinds of hybrid features to encode peptide sequences.
Feature TypesFeature NameDimensions
Sequence featuresPAAC30
Length1
Shannon entropy1
Peptide physicochemical propertiesGravy1
Molecular_weight1
Charge_at_pH(10)1
Embedding featuresPosition embedding50
Table 3. The performance results compared by three types of combinations on ACPs82.
Table 3. The performance results compared by three types of combinations on ACPs82.
TPTNFPFNAccuracyF1-ScoreRecallPreciseMCCAUCPRAUC
MS6568141781.080.779.282.20.6220.8410.832
AE677391585.384.881.788.10.7090.8670.878
MS + AE727571089.689.487.890.10.7930.9450.947
MS means manually selected features, AE means automatic learning features.
Table 4. Comparing results with traditional machine learning.
Table 4. Comparing results with traditional machine learning.
TPTNFPFNAccuracyF1-ScoreRecallPrecisionMCCAUCPRAUC
SVM6063221975.075.573.175.90.5000.7750.763
RF812815466.474.698.760.00.4310.7040.697
CatBoost647718585.984.778.092.70.7280.8830.891
ACPNet727571089.689.487.891.10.7930.9450.947
Table 5. Performance comparisons of ACPNet with the existing methods.
Table 5. Performance comparisons of ACPNet with the existing methods.
TPTNFPFNAccuracyF1-ScoreRecallPrecisionMCCAUCPRAUC
AntiCP_ACC5671261177.475.268.383.60.5580.8050.793
Anticp_DPC6169211379.378.274.482.40.5880.8160.883
Hajisharifi5571271176.874.367.183.30.5470.7790.752
IACP5666261674.472.768.378.80.4910.7520.723
ACPred-FL667916388.487.480.595.70.7780.9060.886
CNN-RNN5967231576.875.672.079.70.5390.7850.758
CNN6465181778.678.578.079.00.5730.8230.845
DeepACP6472181082.982.078.086.50.6620.8590.800
ACPNet727571089.689.487.891.130.7930.9450.947
Table 6. The independent validation results of ACPNet.
Table 6. The independent validation results of ACPNet.
IdSequenceScoreLabel
1KLWKKIEKLIKKLLTSIR0.9999ACP
2YIWARAERVWLWWGKFLSL0.9994ACP
3DLFKQLQRLFLGILYCLYKIW0.8732ACP
4AIKKFGPLAKIVAKV0.7043ACP
5RWNGRIIKGFYNLVKIWKDLKG0.9620ACP
6KVWKIKKNIRRLLHGIKRGWKG0.9993ACP
7GFWARIGKVFAAVKNL0.9988ACP
8AFLYRLTRQIRPWWRWLYKW0.4979Non-ACP
9RIWGKHSRYIKIVKRLIQ0.9993ACP
10QIWHKIRKLWQIIKDGF0.9997ACP
11CGESCVWIPCVTSIFNCKCKENKVCYHDKIP0.0001Non-ACP
12SDEKASPDKHHRFSLSRYAKLANRLANPKLLETFLSKWIGDRGNRSV0.2383Non-ACP
13DVKGMKKAIKGILDCVIEKGYDKLAAKLKKVIQQLWE0.4986Non-ACP
14AGWGSIFKHIFKAGKFIHGAIQAHND0.011Non-ACP
15ATCDLASGFGVGSSLCAAHCIARRYRGGYCNSKAVCVCRN0.0032Non-ACP
16GWKIGKKLEHHGQNIRDGLISAGPAVFAVGQAATIYAAAK0.0015Non-ACP
17FLGALIKGAIHGGRFIHGMIQNHH0.4750Non-ACP
18FLPAIAGILSQLF0.1818Non-ACP
19ALWMTLLKKVLKAAAKALNAVLVGANA0.0052Non-ACP
20EGGGPQWAVGHFM0.1243Non-ACP
Note: the independent validation peptide sequences (1–10, ACP) and (11–20, non-ACP) source from [31,32] respectively.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Sun, M.; Yang, S.; Hu, X.; Zhou, Y. ACPNet: A Deep Learning Network to Identify Anticancer Peptides by Hybrid Sequence Information. Molecules 2022, 27, 1544. https://doi.org/10.3390/molecules27051544

AMA Style

Sun M, Yang S, Hu X, Zhou Y. ACPNet: A Deep Learning Network to Identify Anticancer Peptides by Hybrid Sequence Information. Molecules. 2022; 27(5):1544. https://doi.org/10.3390/molecules27051544

Chicago/Turabian Style

Sun, Mingwei, Sen Yang, Xuemei Hu, and You Zhou. 2022. "ACPNet: A Deep Learning Network to Identify Anticancer Peptides by Hybrid Sequence Information" Molecules 27, no. 5: 1544. https://doi.org/10.3390/molecules27051544

APA Style

Sun, M., Yang, S., Hu, X., & Zhou, Y. (2022). ACPNet: A Deep Learning Network to Identify Anticancer Peptides by Hybrid Sequence Information. Molecules, 27(5), 1544. https://doi.org/10.3390/molecules27051544

Article Metrics

Back to TopTop