Next Article in Journal
Effects of Fenofibrate and Gemfibrozil on Kynurenic Acid Production in Rat Kidneys In Vitro: Old Drugs, New Properties
Previous Article in Journal
Virtual Insights into Natural Compounds as Potential 5α-Reductase Type II Inhibitors: A Structure-Based Screening and Molecular Dynamics Simulation Study
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Enhancing Sumoylation Site Prediction: A Deep Neural Network with Discriminative Features

1
Department of Computer Science, Abdul Wali Khan University, Mardan 23200, Pakistan
2
Department of Information Technology, The University of Haripur, Haripur 22620, Pakistan
3
Department of Convergence Engineering for Intelligent Drone, Sejong University, Seoul 05006, Republic of Korea
4
Department of Information Systems, College of Computer and Information Sciences, Jouf University, Sakaka 72388, Saudi Arabia
5
Department of Computer Sciences, College of Computer and Information Sciences, Princess Nourah bint Abdulrahman University (PNU), P.O. Box 84428, Riyadh 11671, Saudi Arabia
*
Author to whom correspondence should be addressed.
Life 2023, 13(11), 2153; https://doi.org/10.3390/life13112153
Submission received: 7 September 2023 / Revised: 18 October 2023 / Accepted: 25 October 2023 / Published: 2 November 2023

Abstract

:
Sumoylation is a post-translation modification (PTM) mechanism that involves many critical biological processes, such as gene expression, localizing and stabilizing proteins, and replicating the genome. Moreover, sumoylation sites are associated with different diseases, including Parkinson’s and Alzheimer’s. Due to its vital role in the biological process, identifying sumoylation sites in proteins is significant for monitoring protein functions and discovering multiple diseases. Therefore, in the literature, several computational models utilizing conventional ML methods have been introduced to classify sumoylation sites. However, these models cannot accurately classify the sumoylation sites due to intrinsic limitations associated with the conventional learning methods. This paper proposes a robust computational model (called Deep-Sumo) for predicting sumoylation sites based on a deep-learning algorithm with efficient feature representation methods. The proposed model employs a half-sphere exposure method to represent protein sequences in a feature vector. Principal Component Analysis is applied to extract discriminative features by eliminating noisy and redundant features. The discriminant features are given to a multilayer Deep Neural Network (DNN) model to predict sumoylation sites accurately. The performance of the proposed model is extensively evaluated using a 10-fold cross-validation test by considering various statistical-based performance measurement metrics. Initially, the proposed DNN is compared with the traditional learning algorithm, and subsequently, the performance of the Deep-Sumo is compared with the existing models. The validation results show that the proposed model reports an average accuracy of 96.47%, with improvement compared with the existing models. It is anticipated that the proposed model can be used as an effective tool for drug discovery and the diagnosis of multiple diseases.

1. Introduction

Sumoylation is a post-translation modification (PTM) mechanism involved in many critical biological processes, including gene expression, protein localization and stabilization, and genome replication. Sumoylation is an enzyme-mediated covalent modification of a protein sequence in which a chemical group is added to a specific location [1]. It modifies the activity of most eukaryote proteins. These modifications include nitrosylation, phosphorylation, lipidation, proteolysis, ubiquitination, glycosylation, methylation, and acetylation [2,3]. These modifications significantly increase the variety of possible protein final forms generated using the same genetic sequence [4]. Moreover, the analysis of these modifications results in mass changes in proteins; however, their identification generates vital insight into biological function.
Recently, 450 new distinctive protein variations have been recognized, i.e., ubiquitination [5], acetylation [6], phosphorylation [7], and sumoylation [8]. These distinctive protein modifications can modify protein interactions, intracellular distribution, activity of target proteins, and protein longevity through PTM [9]. Among all these protein modifications, the sumoylation sites have received great attention from the research community because the sumoylation sites are widely used for post-translational protein modification. The sumoylation sites play an important role in the functional proteomic [10]. For example, it expands genetic code, underlies protein recycling and degradation, localizes proteins within the cell, regulates the activity of cellular physiology, activates or deactivates transporters and enzymes, and interacts with other cellular molecules. In addition, recent studies have shown that numerous diseases and disorders, i.e., Parkinson’s diseases and Alzheimer’s, have been found to be closely related to the sumoylation site. To extract the structure information, a series of bioinformatics tools have been developed [11,12,13]. Given the importance of sequential and structural bioinformatics in modeling drug discovery, computational biology has played a significant role [14]. Therefore, The discovery of sumoylation sites in proteins is crucial not only for gaining a thorough understanding of many critical biological processes but also for the creation of effective medications that can be utilized as a possible therapeutic target for cancer and other disorders [15].
Given the significance of sumoylation sites in the genome, detecting and identifying them has become a crucial research area in computational biology and bioinformatics. Several models have been proposed in the literature for the identification of sumoylation sites. These models include SUMOsp [16], SUMOsp2.0 [17], and GPS-SUMO [18]. However, the mentioned models are developed based on the group-based phosphorylation scoring algorithm (GPS). The GPS is a simple and intuitive approach that calculates peptide similarities and decides which peptides are closest to the given peptides after clustering known peptide groups [19]. Similarly, Xu et al. [20] presented another technique, SUMO_LDA, developed by merging three distinct types of sequence characteristics into the general pseudo-amino acid composition (PseAAC). The proposed model generated quite promising results; however, the SUMO_LDA is based on linear discriminant analysis (LDA), which suffers from several problems. First, the LDA method fails to find the lower-dimensional space if the number of features is higher than the number of samples in the feature vector, which leads to the problem of small sample size (SSS) [21]. Second, the LDA algorithm faces the linearity problem if the classes are non-linearly separable. In this case, the LDA algorithm will be unable to discriminate the classes accurately [22].
Recently, Xu et al. [23] and Chen et al. [24] proposed computational models known as SUMOPre and SUMOhydro, respectively. The proposed predictors used SVM as a learning algorithm for the classification of SUMO sites. Similarly, Jia et al. [25] developed the pSumo-CD predictor. The proposed predictor employed the general PseAAC method and SVM and RF algorithms for the classification and identification of sumoylation sites. Most recently, Sharma et al. [26] proposed HseSUMO for the prediction of sumoylation sites. The HseSUMO employed the half-sphere exposure method and decision tree algorithm as the learning method. The above-mentioned models have produced promising results; however, conventional learning models require significant human expertise to extract prominent features automatically [27]. Additionally, the prediction performance of these models is limited in the case of nonlinearity in the dataset.
Hence, a more influential and vigorous computational predictor is required in the field of bioinformatics for the identification of sumoylation sites. This paper suggests a predictor for the accurate identification of sumoylation sites using an optimized deep-learning algorithm with efficient feature-extraction methods known as Deep-Sumo. First, the proposed predictor applies half-sphere exposure (HSE) measures as a feature-extraction method. Second, to select efficient features by eliminating irrelevant and noisy features, an unsupervised PCA method is applied. Finally, a multilayer DNN method is functional as a classifier. The proposed model performance is rigorously examined based on diverse performance measurement metrics using a 10-fold cross-validation method. The experiments showed that the proposed predictor attained an average prediction accuracy of 96.47%. Furthermore, a comparison with recently published predictors revealed that the proposed Deep-Sumo outperformed the existing predictors regarding accuracy and other performance measurement metrics. The major contributions of this paper are as follows:
  • An intelligent and robust computational model built on multiple layers.
  • A deep model with automatic weight optimization using standard learning procedure for the accurate prediction of sumoylation sites.
  • The employment of the half-sphere exposure method that efficiently transforms peptide sequences into a feature vector.
  • Efficient features extraction by removing noisy and irrelevant features using an unsupervised PCA algorithm.
  • Mathew’s Correlation Coefficient (MCC), accuracy, sensitivity, specificity, and area under the ROC Curve (AUC) are used to evaluate the performance of the proposed model.

2. Proposed Model Design

This section presents the prototype and design of the proposed model. Its architecture is depicted in Figure 1. The different components of the model are explained in detail below:

2.1. Benchmark Dataset

A legitimate and trustworthy dataset is always required to generate an effective and dependable computational model [28,29]. Therefore, we decided to use a reliable benchmark dataset from [26] for training and verification of the results of the suggested model. The chosen benchmark dataset may be visualized as shown below:
S = S p   U   S N
where S   represents sumoylation and non-sumoylation site sequences. S p and S N represent a subgroup of sumoylation samples and non-sumoylation site samples, respectively. The dataset used in this study was taken from the database Compendium of Protein Lysine Modification (CPLM) [30], a resource that manually collects details of lysine PTMs of 12 different forms. The original benchmark dataset contained 780 sumoylation site sequences (i.e., positive samples) and 21,353 non-sumoylation site sequences (i.e., negative samples). The benchmark dataset needs to be more balanced. A classification problem occurs when the categorization is biased towards the majority class in an unbalanced dataset. The literature has studied several data and algorithm-level strategies for addressing the class imbalance problem and the problem of the unbalanced dataset. However, a data-level approach, which includes oversampling and undersampling approaches, is extensively used [31]. The undersampling strategy can often provide a modest answer, but the oversampling method may raise the danger of model overfitting owing to data sample repetition. Therefore, to balance the benchmark dataset in this work, we used the undersampling approach with the NearMiss algorithm [32] developed in the Python programming language. After employing the undersampling method, we obtained a balanced benchmark dataset, which contained 780 negative samples and 780 positive samples. Additionally, we divided the data randomly into 10 groups: 9/10 (i.e., 702 positives and 702 negatives) as the cross-validation dataset and 1/10 (i.e., 78 positives and 78 negatives) as the independent test dataset. The data frame sample function in the pandas Python library was used for random number generation [33].

2.2. Feature Formulation Technique

The most fundamental and difficult step in computational biology is a discrete model or vector representation of a given sequence [34]. The discrete model should preserve the structure information and keep the key potential characteristics of the given sequence.
This paper used HSE as a feature-extraction method—a two-dimensional solvent exposure measure developed by Hamelryck [35]. Further solvent exposure measures, i.e., RD [36], rASA [37], ASA [38], and CN [39], have drawbacks compared to the HSE method. Moreover, the HSE method reported superior performances than the existing exposure measures in terms of conservation among different folds, computational complexity, protein stability, sensitivity, predictability, and computational speed.
The HSE method divides the sphere of a residue into two half-spheres, i.e., the first half-sphere is HSEβ or HSE-up, and the second half-sphere is referred to as HSEα or HSE-down, as shown in Figure 2. The HSE-up and HSE-down represent the upper and lower spheres concerning the residue’s chain side. These two half-spheres possess distinct geometries and energy properties, exhibiting interesting characteristics for characterizing the residue’s spatial neighborhood. We employed a sphere radius of rd = 13 Å for all residues to calculate the measurements. This resulted in a feature vector of 124 features, representing the protein sequence in a discrete model.

2.3. Feature Selection

The feature vector’s presence of noisy, duplicated, or irrelevant information may severely impact the classifier performance. Therefore, we use Principal Component Analysis (PCA) for feature selection. The PCA algorithm is a multivariate data-processing technique that computes eigenvectors and covariance matrices to decrease the dimensionality of feature vectors.
Consider an input feature vector S , i.e., 124 × 1560, and “k” characterizes the number of desired features.
s 1 , s 2 , s 3 , s n
Using the PCA algorithm,
Step 1: Feature vector mean value using Equation (3)
S _ = 1 i n = 1 i s n
Step 2: Subtract the mean value using Equation (4)
S n = s n s _
Step 3: Covariance matrix:
C n = S n . ( S n ) T B B T
where B = S 1 , S 2 , , S p ( i j ) and transposition of a matrix, T.
Step 4: The Eigenvalues of a matrix are computed in the fourth step. The eigenvalue values γ , of the covariance matrix can be solved using Equation (6).
det ( ( γ I C n ) = 0
where d e t represents the determinant of the matrix, and I represents the identity matrix. The first eigenvalue must be higher than the second eigenvalue.
γ 1 > γ 2 > γ 3 γ n
Step 5: Eigenvector:
c n : σ 1 , σ 2 , σ 3 σ n
Last step: we choose k values with max eigenvalues and select the eigenvector with the maximum eigenvalue. To attain a high level of classification accuracy in our study, our empirical findings demonstrate that utilizing 80 components effectively captured the essential information within the dataset for feature representation. Interestingly, we observe that increasing the number of features beyond 80 has a detrimental effect on the overall performance of the classification model. Moreover, we carefully chose a subset of 80 features that not only sufficed for accurate classification but also helped prevent overfitting and noise, ultimately leading to better model performance.

2.4. Deep Architecture

DNN is an ML algorithm inspired by the human brain [40,41] that includes input, output, and multiple hidden layers in its network topology, as shown in Figure 3.
The DNN’s hidden layers are crucial for learning, but increasing their number can lead to computational costs, overfitting, and model complexity [42,43]. DNNs automatically extract relevant features from unstructured or unlabeled data, making them superior to traditional learning methods. DNNs have been applied successfully in numerous fields, i.e., natural language processing, bio-engineering, and image and speech recognition [44].
As shown in Figure 3, this work predicts sumoylation sites using a benchmark dataset and a DNN model with four hidden layers. The DNN design employs several neurons to assess the input feature vector and create output using Equation (8). The weight matrix for each neuron is established using the Xavier function [45], which may keep the variance constant through each layer. In addition, the weight matrix is updated using a back-propagation approach to eliminate errors between the output and target classes. The input layer and buried layers both use the Tanh nonlinear activation function. A dataset’s nonlinearity and complicated patterns can be learned by a model with the aid of the activation function. Additionally, it decides whether a neuron may be fired or ignored based on the output that a specific neuron produces [46]. A SoftMax activation function is also used at the output layer to provide a value between [0, 1] that reflects the probability that a certain class of data points would be included in the output.
y a = f ( B a + b = 1 m x b w b a )
f ( i ) = e i 1 + e i
The DNN has confines, including high computational complexity/cost and a need for many computational resources to configure a complex model. Additionally, many training samples are required for effective model training.

3. Performance Evaluation

Statistical machine-learning models must be evaluated using performance metrics before deployment in a real production environment. Accuracy is an important metric, but it alone is insufficient. Various performance metrics have been proposed, and the choice of metrics depends on the specific application and problem domain [47,48]. Performance measures include sensitivity (SN), specificity (SP), the area under the curve (AUC), characteristic (ROC), and MCC. As previously stated, this study uses performance measurements to evaluate how well the proposed Deep-Sumo performs because these metrics have been widely utilized in various publications. The performance measurement metrics can be calculated as follows:
A C C = 1 S + + S + S + + S
S P = 1 S + S +
S N = 1 S + S
M C C = 1 ( S +   +   S + S +   +   S ) ( 1 + S +   +   S + S + ) ( 1 + S +   +   S + S )

4. Results and Analysis

4.1. Hyper-Parameter Analysis

The motivation of this section is to obtain optimum configuration values for the hyperparameters used in DNN topology. The important hyperparameters in a DNN include the number of layers and neurons, seed, regularization techniques l1 and l2, activation function, weight initialization, momentum, dropout, updater, iteration, learning rate, and optimizer, as shown in Table 1. These hyperparameters play a crucial role in determining the performance and behavior of the neural network. For example, the number of layers and neurons per layer refers to the architecture of a neural network and can affect the network’s capacity to learn as well as its ability to fit the training data. A seed is a predetermined starting point used to initialize or control random processes, ensuring reproducibility in experiments or computations. Regularization techniques like L1 and L2 regularization help prevent overfitting by adding penalty terms to the loss function. Activation functions are used in each neuron to introduce nonlinearity into the model. The initialization of weights refers to the process of setting the initial values for the parameters (weights and biases) of the network’s neurons before training, and momentum enhances the optimization process by incorporating past gradient information to accelerate convergence and improve stability during training. Dropout is a regularization technique where a random fraction of neurons is dropped during each training iteration. Finally, the “updater” is the mechanism responsible for adjusting model parameters, “iteration” represents a single cycle through the training data, “learning rate” controls the size of weight updates, and the “optimizer” is the overarching algorithm that guides the optimization process by determining how weights are updated in each iteration. To analyze the performance of the DNN on different hyperparameters, we used a grid-search technique that applies different combinations of parameters. For performance analysis, we considered only those hyperparameters that largely affect the performance of the DNN model. These parameters include activation function, learning rate, and number of iterations.
We conducted experiments to examine the effects of activation functions and LR. The results of the experiments are given in Table 2. From the table, it can be observed that the highest accuracy, i.e., 96.47%, is obtained by the DNN classifier at a learning rate value of 0.1 using Tanh as an activation function. Furthermore, the DNN model is continuously improved by decreasing the learning rate; however, after reducing the LR (i.e., 0.09 and 0.08), the DNN model accuracy could not be significantly improved. Hence, the DNN model presented a high accuracy at a learning rate of 0.1 with the Tanh activation function.
Second, we performed several experiments to analyze the DNN performance from the aspect of various iteration numbers at the model training stage. From the experimental results, at 600 iterations, the error losses have become stable using the Tanh, ReLU, and Sigmoid activation functions.

4.2. Cross-Validation Scheme Performance Analysis

Statistical learning models in computational biology and bioinformatics can be assessed by validation tests, i.e., jackknife, k-fold, and sub-sampling tests. Among these, k-fold cross-validation is broadly implemented and provides unbiased outcomes. This study analyzed the proposed method’s performance using 10-fold and 5-fold CV tests. The results in Table 3 indicate that the Deep-Sumo model achieved higher accuracy (91.99%) with a 10-fold CV compared to a 5-fold CV (91.03%).
The feature vector obtained through the HSE method contained 124 features that may comprehend inappropriate, redundant, and deafening features. To obtain efficient features and reduce the dimensionality, we have employed the feature-selection method, as discussed in Section 2.3. We reduced the feature vector dimension from 124 × 1560 to 80 × 1560. The evaluation of the proposed model includes the assessment of its performance using both comprehensive and optimized feature sets, ensuring a thorough analysis of its capabilities. The experimental results of this evaluation are shown in Table 3, which shows that the proposed model performance has better quality using an efficient feature set in comparison to the entire feature set. It can be seen from Table 3 that the accuracy of the proposed model is pointedly enhanced in terms of yielded accuracy—91.99% in the case of the entire feature set—and it obtained an average accuracy of 96.47% in the case of the efficient feature set. Similar improvement is reported in the case of other performance measurement metrics, i.e., specificity (96.25%), sensitivity (96.71%), and MCC (0.929), utilizing efficient feature sets in comparison to the entire feature set.
The Deep-Sumo model performance was evaluated using the AUC metric, which measures the accuracy of binary classifiers. Higher AUC values indicate better performance. Figure 4 shows that the Deep-Sumo model achieved the highest AUC values of 0.984 with 10-fold cross-validation and 0.972 with 5-fold cross-validation when using the efficient feature set. These results confirm the superior prediction of the proposed model performance with a 10-fold CV and selected features.

4.3. Performance Comparison of Different Classifiers

In this section, a comprehensive comparison is conducted between the performance of the Deep Neural Network (DNN) model and other commonly employed machine-learning algorithms, specifically, K-Nearest Neighbor (KNN), Random Forest (RF), and Support Vector Machine (SVM). The KNN algorithm, a non-parametric instance-based learning method, is utilized in image processing to classify instances based on their Euclidean distances [40,41,43,48,49]. Contrarily, RF is a supervised learning method frequently used for regression and classification problems. The bootstrap approach produces numerous decision trees based on random sample selections. This analysis sheds light on each approach’s advantages and disadvantages. The RF algorithm is a supervised learning algorithm widely used for classification and regression problems, which employs a bootstrap algorithm on training samples to create several decision trees based on random selection. The SVM algorithm is a powerful classification algorithm used in bioinformatics that computes an optimal hyperplane to distinguish among groups for linear and nonlinear classification. To achieve a fair comparison across all learning algorithms, we used the same efficient feature set, performance measurement standards, and validation technique.
Table 4 offers a performance comparison of several algorithms. The table shows that the DNN model performed better than the competing models. For instance, the DNN model achieved an average accuracy of 96.47%, while the RF model achieved the next highest accuracy of 94.10%. Similarly, in terms of MCC, which reflects the stability of a model, the DNN model achieved the highest value of 0.929, surpassing the next highest value of 0.883 obtained by the RF algorithm. The KNN model exhibited the worst performance in the case of all performance measurement metrics.
The analysis of the results suggests two foremost causes for the superior performance of the proposed model compared to conventional learning algorithms. First, traditional learning methods typically employ a single-stack processing layer, which may not effectively handle complex datasets with high nonlinearity. In contrast, to process complex datasets effectively, the DNN model utilizes multi-stack processing layers and automatically extracts optimal features. Furthermore, the performance of the DNN model was compared to conventional learning algorithms in terms of AUC (area under the ROC curve), as shown in Figure 5. The figure illustrates that the proposed DNN model achieved the highest AUC value compared to the conventional learning algorithms. For instance, the DNN model achieved an AUC value of 0.984, while the RF, SVM, and KNN algorithms achieved AUC values of 0.963, 0.920, and 0.902, respectively.

4.4. Existing Model Performance Comparisons

Deep-Sumo’s performance was compared to current predictors described by the authors in [26], and the comparative results are shown in Table 5. Deep-Sumo clearly outperformed the previously released predictors, according to the data. For example, the suggested DNN model obtained the best accuracy of 96.47%, whereas the current predictor (HseSUMO) had the second-highest success rate of 89.50%. Similarly, Deep-Sumo had the greatest MCC value of 0.929, whereas HseSUMO obtained an MCC score of 0.790. These findings demonstrate that Deep-Sumo outperformed the current models, with an average success rate improvement of 7.59%.

4.5. Performance Comparison on Independent Dataset

To ensure the stability and reliability of the proposed Deep-Sumo, we perform an independent dataset test. The outcome of the Deep-Sumo on an independent dataset in comparison with other learning models is given in Table 6. As shown in Table 6, Deep-Sumo performed remarkably well and reported the highest accuracies, i.e., 92.23%. Similarly, Deep-Sumo reported better sensitivity, specificity, and MCC in comparison with other learning models, as shown in Table 6. Furthermore, in the case of conventional classifiers, RF reported the highest accuracies, i.e., 91.43%, in comparison with SVM and KNN.

5. Conclusions

This paper proposed Deep-Sumo, an intelligent computational model for the prediction of sumoylation sites. The proposed model applied the half-sphere exposure method to depict protein samples into a feature vector, and it employed the DNN model as a prediction engine. Additionally, an unsupervised PCA algorithm was employed for optimum feature selection, and a data-level (i.e., undersampling) method was employed to balance the benchmark dataset. The performance of the proposed model was extensively evaluated, and experimental results showed that the Deep-Sumo model accurately predicts the sumoylation sites. Moreover, the performance of the proposed Deep-Sumo was compared with the existing model, and the comparison analysis yielded that Deep-Sumo outperformed the existing models.
It is proposed that the suggested method can be an accessible prediction method for the identification of sumoylation sites. Moreover, it could be utilized for drug discovery and basic research. In future work, as mentioned in a series of recent publications [50,51], we plan to develop a publicly accessible web server for Deep-Sumo, allowing experimental scientists to use it to identify sumoylation sites easily. In addition, a huge amount of genome data is generated due to advancements in next-generation sequencing technology, which poses computational challenges for sequential computing approaches. We also plan to apply parallel programming techniques to parallelize computations on the number of processing nodes using big-data analytics platforms [52]. For the convenience of researchers, the source code of the proposed predictor, along with the dataset, is available at https://github.com/salman-khan-mrd/Deep-Sumo (accessed on 8 September 2023).

Author Contributions

Methodology, validation, writing—original draft preparation, S.K.; supervision, M.K.; supervision, N.I.; writing—review and editing, N.D.; formal analysis, investigation, M.F.A.; data curation, supervision, project administration, funding acquisition, N.A. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Princess Nourah bint Abdulrahman University Researchers Supporting Project number (PNURSP2023R321), Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Dataset will be provided on request.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Mann, M.; Jensen, O.N. Proteomic Analysis of Post-Translational Modifications. Nat. Biotechnol. 2003, 21, 255–261. [Google Scholar] [CrossRef] [PubMed]
  2. Kessler, B.M.; Edelmann, M.J. PTMs in Conversation: Activity and Function of Deubiquitinating Enzymes Regulated via Post-Translational Modifications. Cell Biochem. Biophys. 2011, 60, 21–38. [Google Scholar] [CrossRef] [PubMed]
  3. Huber, S.C.; Hardin, S.C. Numerous Posttranslational Modifications Provide Opportunities for the Intricate Regulation of Metabolic Enzymes at Multiple Levels. Curr. Opin. Plant Biol. 2004, 7, 318–322. [Google Scholar] [CrossRef] [PubMed]
  4. Jensen, O.N. Interpreting the Protein Language Using Proteomics. Nat. Rev. Mol. Cell Biol. 2006, 7, 391–403. [Google Scholar] [CrossRef]
  5. Kim, W.; Bennett, E.J.; Huttlin, E.L.; Guo, A.; Li, J.; Possemato, A.; Sowa, M.E.; Rad, R.; Rush, J.; Comb, M.J.; et al. Systematic and Quantitative Assessment of the Ubiquitin-Modified Proteome. Mol. Cell 2011, 44, 325–340. [Google Scholar] [CrossRef]
  6. Drazic, A.; Myklebust, L.M.; Ree, R.; Arnesen, T. The World of Protein Acetylation. Biochim. Biophys. Acta-Proteins Proteomics 2016, 1864, 1372–1401. [Google Scholar] [CrossRef]
  7. Guo, M.; Huang, B.X. Integration of Phosphoproteomic, Chemical, and Biological Strategies for the Functional Analysis of Targeted Protein Phosphorylation. Proteomics 2013, 13, 424–437. [Google Scholar] [CrossRef]
  8. Venne, A.S.; Kollipara, L.; Zahedi, R.P. The next Level of Complexity: Crosstalk of Posttranslational Modifications. Proteomics 2014, 14, 513–524. [Google Scholar] [CrossRef]
  9. Verdin, E.; Ott, M. 50 Years of Protein Acetylation: From Gene Regulation to Epigenetics, Metabolism and Beyond. Nat. Rev. Mol. Cell Biol. 2015, 16, 258–264. [Google Scholar] [CrossRef]
  10. Warden, S.M.; Richardson, C.; O’Donnell, J., Jr.; Stapleton, D.; Kemp, B.E.; Witters, L.A. Post-Translational Modifications of the β-1 Subunit of AMP-Activated Protein Kinase Affect Enzyme Activity and Cellular Localization. Biochem. J. 2001, 354, 275. [Google Scholar] [CrossRef]
  11. Lee, H.; Iqbal, N.; Chang, W.; Lee, S.-Y. A Calibration Method for Eye-Gaze Estimation Systems Based on 3D Geometrical Optics. IEEE Sens. J. 2013, 13, 3219–3225. [Google Scholar] [CrossRef]
  12. OuYang, B.; Xie, S.; Berardi, M.J.; Zhao, X.; Dev, J.; Yu, W.; Sun, B.; Chou, J.J. Unusual Architecture of the P7 Channel from Hepatitis C Virus. Nature 2013, 498, 521–525. [Google Scholar] [CrossRef]
  13. Oxenoid, K.; Dong, Y.; Cao, C.; Cui, T.; Sancak, Y.; Markhard, A.L.; Grabarek, Z.; Kong, L.; Liu, Z.; Ouyang, B.; et al. Architecture of the Mitochondrial Calcium Uniporter. Nature 2016, 533, 269–273. [Google Scholar] [CrossRef]
  14. Liu, B.; Wu, H.; Chou, K.-C. Pse-in-One 2.0: An Improved Package of Web Servers for Generating Various Modes of Pseudo Components of DNA, RNA, and Protein Sequences. Nat. Sci. 2017, 9, 67–91. [Google Scholar] [CrossRef]
  15. Bettermann, K.; Benesch, M.; Weis, S.; Haybaeck, J. SUMOylation in Carcinogenesis. Cancer Lett. 2012, 316, 113–125. [Google Scholar] [CrossRef]
  16. Xue, Y.; Zhou, F.; Fu, C.; Xu, Y.; Yao, X. SUMOsp: A Web Server for Sumoylation Site Prediction. Nucleic Acids Res. 2006, 34, W254–W257. [Google Scholar] [CrossRef]
  17. Ren, J.; Gao, X.; Jin, C.; Zhu, M.; Wang, X.; Shaw, A.; Wen, L.; Yao, X.; Xue, Y. Systematic Study of Protein Sumoylation: Development of a Site-Specific Predictor of SUMOsp 2.0. Proteomics 2009, 9, 3409–3412. [Google Scholar] [CrossRef]
  18. Zhao, Q.; Xie, Y.; Zheng, Y.; Jiang, S.; Liu, W.; Mu, W.; Liu, Z.; Zhao, Y.; Xue, Y.; Ren, J. GPS-SUMO: A Tool for the Prediction of Sumoylation Sites and SUMO-Interaction Motifs. Nucleic Acids Res. 2014, 42, W325–W330. [Google Scholar] [CrossRef]
  19. Zhou, F.-F.; Xue, Y.; Chen, G.-L.; Yao, X. GPS: A Novel Group-Based Phosphorylation Predicting and Scoring Method. Biochem. Biophys. Res. Commun. 2004, 325, 1443–1448. [Google Scholar] [CrossRef]
  20. Xu, Y.; Ding, Y.-X.; Deng, N.-Y.; Liu, L.-M. Prediction of Sumoylation Sites in Proteins Using Linear Discriminant Analysis. Gene 2016, 576, 99–104. [Google Scholar] [CrossRef]
  21. Yang, W. Regularized Complete Linear Discriminant Analysis for Small Sample Size Problems. In Communications in Computer and Information Science; Springer: Berlin/Heidelberg, Germany, 2012; pp. 67–73. ISBN 9783642318368. [Google Scholar]
  22. Tharwat, A.; Gaber, T.; Ibrahim, A.; Hassanien, A.E. Linear Discriminant Analysis: A Detailed Tutorial. AI Commun. 2017, 30, 169–190. [Google Scholar] [CrossRef]
  23. Xu, J.; He, Y.; Qiang, B.; Yuan, J.; Peng, X.; Pan, X.-M. A Novel Method for High Accuracy Sumoylation Site Prediction from Protein Sequences. BMC Bioinform. 2008, 9, 8. [Google Scholar] [CrossRef] [PubMed]
  24. Chen, Y.-Z.; Chen, Z.; Gong, Y.-A.; Ying, G. SUMOhydro: A Novel Method for the Prediction of Sumoylation Sites Based on Hydrophobic Properties. PLoS ONE 2012, 7, e39195. [Google Scholar] [CrossRef] [PubMed]
  25. Jia, J.; Zhang, L.; Liu, Z.; Xiao, X.; Chou, K.-C. PSumo-CD: Predicting Sumoylation Sites in Proteins with Covariance Discriminant Algorithm by Incorporating Sequence-Coupled Effects into General PseAAC. Bioinformatics 2016, 32, 3133–3141. [Google Scholar] [CrossRef] [PubMed]
  26. Sharma, A.; Lysenko, A.; López, Y.; Dehzangi, A.; Sharma, R.; Reddy, H.; Sattar, A.; Tsunoda, T. HseSUMO: Sumoylation Site Prediction Using Half-Sphere Exposures of Amino Acids Residues. BMC Genom. 2019, 19, 982. [Google Scholar] [CrossRef]
  27. Lecun, Y.; Bengio, Y.; Hinton, G. Deep Learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  28. Chou, K.-C.; Shen, H.-B. REVIEW: Recent Advances in Developing Web-Servers for Predicting Protein Attributes. Nat. Sci. 2009, 1, 63–92. [Google Scholar] [CrossRef]
  29. Chou, K.-C.C. Some Remarks on Protein Attribute Prediction and Pseudo Amino Acid Composition. J. Theor. Biol. 2011, 273, 236–247. [Google Scholar] [CrossRef]
  30. Liu, Z.; Wang, Y.; Gao, T.; Pan, Z.; Cheng, H.; Yang, Q.; Cheng, Z.; Guo, A.; Ren, J.; Xue, Y. CPLM: A Database of Protein Lysine Modifications. Nucleic Acids Res. 2014, 42, D531–D536. [Google Scholar] [CrossRef]
  31. Kaur, P.; Gosain, A. Comparing the Behavior of Oversampling and Undersampling Approach of Class Imbalance Learning by Combining Class Imbalance Problem with Noise. In Advances in Intelligent Systems and Computing; Springer: Singapore, 2018; pp. 23–30. ISBN 9789811066016. [Google Scholar]
  32. Yen, S.-J.; Lee, Y.-S. Under-Sampling Approaches for Improving Prediction of the Minority Class in an Imbalanced Dataset. In Intelligent Control and Automation; Springer: Berlin/Heidelberg, Germany, 2006; pp. 731–740. ISBN 3540372555. [Google Scholar]
  33. Zhu, Y.; Liu, Y.; Chen, Y.; Li, L. ResSUMO: A Deep Learning Architecture Based on Residual Structure for Prediction of Lysine SUMOylation Sites. Cells 2022, 11, 2646. [Google Scholar] [CrossRef]
  34. Chou, K.-C. Pseudo Amino Acid Composition and Its Applications in Bioinformatics, Proteomics and System Biology. Curr. Proteom. 2009, 6, 262–274. [Google Scholar] [CrossRef]
  35. Hamelryck, T. An Amino Acid Has Two Sides: A New 2D Measure Provides a Different View of Solvent Exposure. Proteins Struct. Funct. Bioinforma. 2005, 59, 38–48. [Google Scholar] [CrossRef] [PubMed]
  36. Chakravarty, S.; Varadarajan, R. Residue Depth: A Novel Parameter for the Analysis of Protein Structure and Stability. Structure 1999, 7, 723–732. [Google Scholar] [CrossRef]
  37. Rost, B.; Sander, C. Conservation and Prediction of Solvent Accessibility in Protein Families. Proteins Struct. Funct. Genet. 1994, 20, 216–226. [Google Scholar] [CrossRef]
  38. Miller, S.; Janin, J.; Lesk, A.M.; Chothia, C. Interior and Surface of Monomeric Proteins. J. Mol. Biol. 1987, 196, 641–656. [Google Scholar] [CrossRef]
  39. Nishikawa, K.; Ooi, T. Prediction of the Surface-Interior Diagram of Globular Proteins by an Empirical Method. Int. J. Pept. Protein Res. 1980, 16, 19–32. [Google Scholar] [CrossRef]
  40. Khan, S.; Khan, M.; Iqbal, N.; Amiruddin Abd Rahman, M.; Khalis Abdul Karim, M. Deep-PiRNA: Bi-Layered Prediction Model for PIWI-Interacting RNA Using Discriminative Features. Comput. Mater. Contin. 2022, 72, 2243–2258. [Google Scholar] [CrossRef]
  41. Khan, S.; Khan, M.; Iqbal, N.; Khan, S.A.; Chou, K.-C. Prediction of PiRNAs and Their Function Based on Discriminative Intelligent Model Using Hybrid Features into Chou’s PseKNC. Chemom. Intell. Lab. Syst. 2020, 203, 104056. [Google Scholar] [CrossRef]
  42. Wu, Y.; Tan, H.; Qin, L.; Ran, B.; Jiang, Z. A Hybrid Deep Learning Based Traffic Flow Prediction Method and Its Understanding. Transp. Res. Part C Emerg. Technol. 2018, 90, 166–180. [Google Scholar] [CrossRef]
  43. Khan, S.; Khan, M.; Iqbal, N.; Hussain, T.; Khan, S.A.; Chou, K.-C. A Two-Level Computation Model Based on Deep Learning Algorithm for Identification of PiRNA and Their Functions via Chou’s 5-Steps Rule. Int. J. Pept. Res. Ther. 2020, 26, 795–809. [Google Scholar] [CrossRef]
  44. Al-Jumaili, M.H.A.; Siddique, F.; Abul Qais, F.; Hashem, H.E.; Chtita, S.; Rani, A.; Uzair, M.; Almzaien, K.A. Analysis and Prediction Pathways of Natural Products and Their Cytotoxicity against HeLa Cell Line Protein Using Docking, Molecular Dynamics and ADMET. J. Biomol. Struct. Dyn. 2023, 41, 765–777. [Google Scholar] [CrossRef] [PubMed]
  45. Glorot, X.; Bengio, Y. Understanding the Difficulty of Training Deep Feedforward Neural Networks. In Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics, Sardinia, Italy, 13–15 May 2010; Volume 9, pp. 249–256. [Google Scholar]
  46. Voisin, T.; Rouet-Benzineb, P.; Reuter, N.; Laburthe, M. Orexins and Their Receptors: Structural Aspects and Role in Peripheral Tissues. Cell. Mol. Life Sci. 2003, 60, 72–87. [Google Scholar] [CrossRef] [PubMed]
  47. Baratloo, A.; Hosseini, M.; Negida, A.; El Ashal, G. Part 1: Simple Definition and Calculation of Accuracy, Sensitivity and Specificity. Emergency 2015, 3, 48–49. [Google Scholar] [CrossRef] [PubMed]
  48. Khan, S.; Naeem, M.; Qiyas, M. Deep Intelligent Predictive Model for the Identification of Diabetes. AIMS Math. 2023, 8, 16446–16462. [Google Scholar] [CrossRef]
  49. Khan, S.; Khan, M.; Iqbal, N.; Li, M.; Khan, D.M. Spark-Based Parallel Deep Neural Network Model for Classification of Large Scale RNAs into PiRNAs and Non-PiRNAs. IEEE Access 2020, 8, 136978–136991. [Google Scholar] [CrossRef]
  50. Chen, W.; Tang, H.; Ye, J.; Lin, H.; Chou, K.C. IRNA-PseU: Identifying RNA Pseudouridine Sites. Mol. Ther.-Nucleic Acids 2016, 5, e332. [Google Scholar] [CrossRef]
  51. Khan, F.; Khan, M.; Iqbal, N.; Khan, S.; Muhammad Khan, D.; Khan, A.; Wei, D.-Q. Prediction of Recombination Spots Using Novel Hybrid Feature Extraction Method via Deep Learning Approach. Front. Genet. 2020, 11, 1052. [Google Scholar] [CrossRef]
  52. Khan, S.; Khan, M.A.; Khan, M.; Iqbal, N.; AlQahtani, S.A.; Al-Rakhami, M.S.; Khan, D.M. Optimized Feature Learning for Anti-Inflammatory Peptide Prediction Using Parallel Distributed Computing. Appl. Sci. 2023, 13, 7059. [Google Scholar] [CrossRef]
Figure 1. Proposed computational model framework.
Figure 1. Proposed computational model framework.
Life 13 02153 g001
Figure 2. Overall structure of half-sphere exposure [26].
Figure 2. Overall structure of half-sphere exposure [26].
Life 13 02153 g002
Figure 3. DNN configuration topology.
Figure 3. DNN configuration topology.
Life 13 02153 g003
Figure 4. AUC of different cross-validation schemes.
Figure 4. AUC of different cross-validation schemes.
Life 13 02153 g004
Figure 5. AUC performance comparison of ML algorithms.
Figure 5. AUC performance comparison of ML algorithms.
Life 13 02153 g005
Table 1. DNN model optimum hyperparameter values.
Table 1. DNN model optimum hyperparameter values.
List of ParametersOptimal Values
Neurons at hidden layers80-72-56-23-16-2
Seed12345L
Regularization l20.001
Activation FunctionsTanh and SoftMax
Weight initialization functionXAVIER function
Momentum0.9
Dropout0.25
Number of hidden layers4
UpdaterADAGRAD function
Training iterations600
OptimizerSGD Method
LR0.1
Table 2. Grid search of the DNN model.
Table 2. Grid search of the DNN model.
LRTanhSigmoidReLU
0.0896.4789.2392.01
0.0996.4789.2392.01
0.196.4789.2392.01
0.293.4187.7890.71
0.391.7285.8089.21
0.490.3584.1787.84
0.587.9782.4686.44
0.685.9780.7485.04
0.783.9679.0383.64
0.881.9577.3182.24
0.979.9575.6080.84
Table 3. Performance evaluation of the Deep-Sumo model using the entire feature set and efficient features set.
Table 3. Performance evaluation of the Deep-Sumo model using the entire feature set and efficient features set.
MethodACC (%)SP (%)SN (%)MCC
Deep-Sumo (5-fold)91.0385.2195.880.821
Deep-Sumo (10-fold)91.9989.4494.700.841
After Feature Selection
Deep-Sumo (5-fold)95.1993.0497.400.905
Deep-Sumo (10-fold)96.4796.2596.710.929
Table 4. Proposed model comparison with other ML.
Table 4. Proposed model comparison with other ML.
MLSP (%)MCCACC (%)SN (%)
Deep-Sumo96.250.92996.4796.71
SVM 93.080.80290.0086.92
RF95.900.88394.1092.31
KNN90.010.80189.1078.33
Table 5. Existing model performance comparisons.
Table 5. Existing model performance comparisons.
MethodACC (%)SP (%)SN (%)MCC
pSumo-CD72.8092.1053.600.494
HseSUMO89.5089.5089.500.790
Deep-Sumo96.4796.2596.710.929
Table 6. Proposed model comparison with other ML on an independent dataset.
Table 6. Proposed model comparison with other ML on an independent dataset.
MLACC (%)SP (%)SN (%)MCC
Deep-Sumo92.2393.5390.930.892
RF91.4392.1290.740.854
SVM88.3291.7684.880.782
KNN87.6588.1287.180.781
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Khan, S.; Khan, M.; Iqbal, N.; Dilshad, N.; Almufareh, M.F.; Alsubaie, N. Enhancing Sumoylation Site Prediction: A Deep Neural Network with Discriminative Features. Life 2023, 13, 2153. https://doi.org/10.3390/life13112153

AMA Style

Khan S, Khan M, Iqbal N, Dilshad N, Almufareh MF, Alsubaie N. Enhancing Sumoylation Site Prediction: A Deep Neural Network with Discriminative Features. Life. 2023; 13(11):2153. https://doi.org/10.3390/life13112153

Chicago/Turabian Style

Khan, Salman, Mukhtaj Khan, Nadeem Iqbal, Naqqash Dilshad, Maram Fahaad Almufareh, and Najah Alsubaie. 2023. "Enhancing Sumoylation Site Prediction: A Deep Neural Network with Discriminative Features" Life 13, no. 11: 2153. https://doi.org/10.3390/life13112153

APA Style

Khan, S., Khan, M., Iqbal, N., Dilshad, N., Almufareh, M. F., & Alsubaie, N. (2023). Enhancing Sumoylation Site Prediction: A Deep Neural Network with Discriminative Features. Life, 13(11), 2153. https://doi.org/10.3390/life13112153

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop