Next Article in Journal
NOTCH3 and Pulmonary Arterial Hypertension
Previous Article in Journal
Regulation of Neuronal Chloride Homeostasis by Pro- and Mature Brain-Derived Neurotrophic Factor (BDNF) via KCC2 Cation–Chloride Cotransporters in Rat Cortical Neurons
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Enhancing Machine-Learning Prediction of Enzyme Catalytic Temperature Optima through Amino Acid Conservation Analysis

1
College of Biotechnology, Tianjin University of Science and Technology, Tianjin 300457, China
2
Tianjin Institute of Industrial Biotechnology, Chinese Academy of Sciences, Tianjin 300308, China
3
Department of Life Sciences and Medicine, University of Science and Technology of China, Hefei 230022, China
4
University of Chinese Academy of Sciences, Beijing 100049, China
5
National Center of Technology Innovation for Synthetic Biology, Tianjin 300308, China
*
Authors to whom correspondence should be addressed.
Int. J. Mol. Sci. 2024, 25(11), 6252; https://doi.org/10.3390/ijms25116252
Submission received: 2 April 2024 / Revised: 22 May 2024 / Accepted: 30 May 2024 / Published: 6 June 2024
(This article belongs to the Section Molecular Informatics)

Abstract

:
Enzymes play a crucial role in various industrial production and pharmaceutical developments, serving as catalysts for numerous biochemical reactions. Determining the optimal catalytic temperature (Topt) of enzymes is crucial for optimizing reaction conditions, enhancing catalytic efficiency, and accelerating the industrial processes. However, due to the limited availability of experimentally determined Topt data and the insufficient accuracy of existing computational methods in predicting Topt, there is an urgent need for a computational approach to predict the Topt values of enzymes accurately. In this study, using phosphatase (EC 3.1.3.X) as an example, we constructed a machine learning model utilizing amino acid frequency and protein molecular weight information as features and employing the K-nearest neighbors regression algorithm to predict the Topt of enzymes. Usually, when conducting engineering for enzyme thermostability, researchers tend not to modify conserved amino acids. Therefore, we utilized this machine learning model to predict the Topt of phosphatase sequences after removing conserved amino acids. We found that the predictive model’s mean coefficient of determination (R2) value increased from 0.599 to 0.755 compared to the model based on the complete sequences. Subsequently, experimental validation on 10 phosphatase enzymes with undetermined optimal catalytic temperatures shows that the predicted values of most phosphatase enzymes based on the sequence without conservative amino acids are closer to the experimental optimal catalytic temperature values. This study lays the foundation for the rapid selection of enzymes suitable for industrial conditions.

1. Introduction

Enzymes, as catalysts for biochemical reactions, play an indispensable role in various biotechnologies, industrial production, and drug development. Accurately predicting the optimal reaction temperature of enzymes helps improve system stability and reaction rates, reduce production costs, and provide critical support for industrial manufacturing. Furthermore, accurately predicting the optimal reaction temperature of enzymes can enable researchers to rapidly screen enzymes suitable for specific processes or experimental conditions, providing robust support for accelerated product development [1].
In the laboratory, accurately determining proteins’ optimal catalytic temperature (Topt) requires a series of complex procedures. First, substrates related to the specific catalytic activity of the protein must be selected. Then, protein expression and purification are conducted, followed by reactions at different temperatures. The protein’s optimal catalytic temperature is determined by comparing the changes in reaction rates. Due to high costs and long experimental durations, determining the Topt of a vast amount of enzymes due to significant advancements in sequencing technology through experimental methods is impractical. Directly predicting the Topt of enzymes from sequences would undoubtedly accelerate enzyme screening. With the development of bioinformatics and artificial intelligence [2], researchers have designed various machine-learning models to predict the optimal catalytic temperature of proteins [3]. Among them, the TOME machine learning model developed by Li et al. [4] combines the optimal growth temperature (OGT) of organisms with the sequence information of individual enzymes to predict the enzyme Topt. This model employs amino acid frequency and OGT data as features and the random forest regression as the model algorithm, achieving a maximum R2 of 0.51. However, this study only provides the codes for predicting OGT on the GitHub code repository and does not offer a specific approach for predicting the optimal enzyme catalytic temperature. Subsequently, Gao et al. proposed TOMER [5], which uses the training set of TOME followed by a normal distribution and employs ensemble learning and resampling strategies to increase the R2 value to 0.632. Although the model has made progress in predicting the thermal stability of enzymes, it still relies on OGT as supplementary information and cannot directly predict the optimal catalytic temperature of enzymes based solely on sequences. More recently, Engqvist et al. developed a deep-learning-based model, DeepET (https://github.com/EngqvistLab/DeepET_reps) (accessed on 28 March 2024) [6], using optimal growth temperatures (OGT) from 21,498 microorganisms through transfer learning with an R2 of 0.57 on the test dataset. Due to the model’s utilization of bacterial growth temperature as the dataset, the predictive performance may not be ideal for certain specific families. However, it is a feasible and reasonable strategy to determine the optimal catalytic temperature of enzymes based on the known Topt values of other members within the same enzyme family and then extrapolate to predict the optimal catalytic temperature of enzymes from different families [7]. In this study, we selected phosphatases as our research subject. Phosphatase (EC 3.1.3.X), which catalyzes irreversible dephosphorylation reaction, is often designed as the last step of many in vitro synthetic enzymatic biosystems (ivSEBs) for the production of various chemicals, such as inositol [8,9], glucosamine [10], and allulose, with high product yields [11]. Furthermore, ivSEB often requires thermally stable enzymes because of system stability and easy enzyme purification by heat treatment [9]. Therefore, quickly identifying the Topt of phosphatases that meet production conditions becomes crucial in accelerating the industrialization of ivSEB. Therefore, in this study, this enzyme family was used as an example to develop a strategy for the quick prediction of enzyme Topt. Furthermore, the phosphatase family (EC 3.1.3.X) is a family recorded in the BRENDA database with relatively large records of Topt values (Supplementary Data, Table S1).
Conserved amino acids remain unchanged during the long evolutionary process of protein sequences, playing a crucial role in maintaining the basic structure and catalytic function of enzymes [12]. In previous thermal stability studies, researchers have typically utilized non-conserved amino acids to modify enzymes: for instance, Li et al. significantly improved the thermal stability and catalytic efficiency of β-glucuronidase from Aspergillus terreus Li-20 by truncation of non-conservative sequences [13]; Xu et al. improved the thermal stability and activity of the thermophilic Bacillus subtilis enzyme by substituting non-conservative residues of thermophilic B. subtilis enzyme with non-conservative residues of thermophilic B. subtilis enzyme [14]; Guo et al. improved the thermal stability of Bs APA aspartic protease via the targeted mutagenesis of non-conserved amino acids near the autocatalytic site region [15]. Therefore, we speculate that conserved and non-conserved amino acids may have varying degrees of influence on enzyme thermostability.
Here, we develop a new machine learning model for predicting the Topt of phosphatases (EC 3.1.3.X) based only on the amino acid sequence, with the potential to be generalized to predict the Topt of other enzymes from different families. We obtained phosphatase information from the BRENDA and UniProt databases and employed multiple sequence alignments to remove the conserved amino acids from the sequences. Subsequently, we computed multiple features for each phosphatase sequence and attempted to establish a machine-learning prediction model for the Topt of phosphatase based on both wild-type and non-conserved amino acid sequence information. Compared to the wild-type sequences, the predicted Topt values obtained from the sequences without conserved amino acids are generally closer to the experimentally determined values. To further validate the reliability of the model, we collected 10 phosphatases with undetermined optimal catalytic temperatures for experimental validation. The results also indicate that the predicted values of nine phosphatases based on the sequence without conserved amino acids are closer to the experimental values, further confirming the reliability of our model. The strategy developed in this study may provide a generalized protocol for accelerating the screening process of suitable enzymes that will be utilized in harsh environments.

2. Results

2.1. Phosphatase Data Collection and Processing

We obtained a dataset containing 249 phosphatases with experimentally validated Topt values from the BRENDA and UniProt databases. This dataset is referred to as the Topt249 dataset (Supplementary Data, Table S2). We grouped these phosphatases based on the fourth digit of their EC numbers, resulting in 37 groups. After removing sequences with lengths less than 50 and groups containing only one enzyme, 234 data points remained (23 groups). Due to the small number of conserved amino acids in the groups with EC numbers 3.1.3.2, 3.1.3.8, 3.1.3.11, 3.1.3.16, 3.1.3.26, and 3.1.3.48, we constructed sequence similarity networks (SSNs) (Supplementary Material, Figure S1) for each group and removed 14 clusters that contained only a single node in the similarity network. The dataset was then reduced to 220 data points. This process allowed us to determine the conserved amino acids for each group (Supplementary Data, Table S3) and remove these conserved amino acids from the original sequences to generate a new set of sequences. Due to the sequence length being less than 50 after removing conserved amino acids for the group with the EC number 3.1.3.B9, we excluded two data points from this group. In the end, we obtained 22 groups containing 218 phosphatases without conserved amino acids, which formed the Topt_rcaa218 dataset (Supplementary Data, Table S4). To compare prediction accuracy, these 218 phosphatases with the original sequence formed another dataset named Topt218 (Supplementary Data, Table S5). We conducted a statistical analysis of the phosphatase counts under different EC groups in the datasets of Topt249, Topt_rcaa218 and Topt218. Among them, the group with EC number 3.1.3.8 has the highest number of phosphatases, with 61 in the Topt249 dataset and 58 in the Topt218 and Topt_rcaa218 datasets. Secondly, the group with EC number 3.1.3.26 has 36 phosphatases in the Topt249 dataset and 35 in the Topt218 and Topt_rcaa218 datasets (Figure 1a).
We divided the Topt218 and Topt_rcaa218 datasets into five temperature ranges based on their measured Topt values: [0–30 °C, 30–50 °C, 50–65 °C, 65–85 °C, 85–100 °C] (Figure 1b). We found that about 50% of enzymes were in the 30–50 °C range, while less than 3% were in the 85–100 °C range. To ensure that data for each temperature band is included in both the training and test sets, half of the data in the smallest set (85–100 °C) is used for training. In the end, we selected two data points from each of the five temperature ranges, totaling 10 data points, to form the test set. The remaining data was used for the training set. (See Section 3.3)
In this study, the Topt of phosphatases exhibit a normal distribution. This imbalanced distribution may lead to insufficient sampling in temperature regions with fewer data points, thereby affecting prediction outcomes and reducing the overall model performance [16,17,18]. To address the issue of data imbalance, we referenced the study by Gado et al. [5], who resolved the problem of uneven distribution in the training set of the TOME model through resampling strategies, thus improving the predictive accuracy of the TOME model. Therefore, before fitting regression variables each time, we applied a regression resampling strategy to the training dataset, ensuring that the quantity of data contained in the rare domain is relatively balanced with the normal domain. By applying the resampling strategy to the training set of Topt249 (Supplementary Data, Table S6), a significant decrease in the number of temperature ranges with rich data distribution was observed: reduced to one-fourth of the original. The number of temperature ranges with less data distribution significantly increased, approximately tripled compared to the original (Figure 1c). Additionally, applying the resampling strategy to the training set of Topt218 and Topt_rcaa218 resulted in similar effects (Figure 1d), thereby achieving a relatively balanced distribution of data across all temperature ranges within the entire training set.

2.2. Prediction of Optimal Catalytic Temperature for Phosphatases by Previous Tools

We first used the deep neural network model DeepET developed by Engqvist et al. [6] to predict the above Topt249 and Topt218 datasets of phosphatase enzymes. On the Topt249 phosphatase dataset, the model estimated an R2 of 0.372 (Supplementary Material, Figure S2a), and on the Topt218 phosphatase dataset, the model estimated an R2 of 0.358 (Supplementary Material, Figure S2b). It is evident that the DeepET model does not achieve satisfactory accuracy in predicting the optimal catalytic temperature for the phosphatase enzyme family.
Later, we employed the TOMER model developed by Gado et al. [5] to predict the Topt249 and Topt218 datasets of phosphatase enzymes. As this model requires the growth temperature of bacteria as auxiliary information, we obtained the bacterial growth temperature information from DSMZ (https://webshop.dsmz.de/index.php) (accessed on 28 March 2024). For the Topt249 dataset, only 116 data points had corresponding bacterial growth temperature information, and we used the TOMER model to predict these 116 data points (Topt116) (Supplementary Data, Table S7), resulting in an R² value of 0.522 (Supplementary Material, Figure S3a). For the Topt218 dataset, only 100 data points had corresponding bacterial growth temperature information. Similarly, we predicted these 100 data points (Topt100) (Supplementary Data, Table S8), resulting in an R² value of 0.47 (Supplementary Material, Figure S3b).
Overall, DeepET’s performance in predicting phosphatases is not ideal. Although TOMER shows some improvement in accuracy compared to DeepET, it still cannot predict enzymes without bacterial growth temperature information. This result indicates that it is necessary to develop an alternative model with better accuracy in predicting the Topt of enzymes.

2.3. Feature Extraction and Model Evaluation Based on Topt249

We extracted four sets of features from the phosphatase sequences, including amino acid frequency, dipeptide frequency, protein molecular weight [19], and protein descriptors (conjoint triad features [20] and distribution descriptors [21]). We trained and tested every feature as well as combinations of these features (total 12 sets of feature and feature combinations, Table 1) using five different machine learning regression models to predict the Topt of phosphatase. First, we computed the multiple sets of features for the phosphatase data in the Topt249 dataset and evaluated the predictive performance based on the mean coefficient of determination (R2) from ten-fold Monte Carlo cross-validation After comparing all combinations of features and algorithms, we found that using the combination of amino acid frequency and protein molecular weight as features, along with the AdaBoost Regression machine learning algorithm for prediction, resulted in the highest prediction accuracy, albeit only 0.495 (Figure 2a).

2.4. Removing the Conserved Amino Acid in the Protein Sequence Facilitates the Prediction of Phosphatase Catalytic Temperature Optima by Machine-Learning

To validate our hypothesis regarding the potentially different impacts of conservative and non-conservative amino acids on enzyme thermal stability, subsequently, we also computed the 12 sets of features for the Topt218 and Topt_rcaa218 phosphatase datasets. We then fitted the same five machine learning regression models mentioned above, adjusting parameters to optimize the models. The performance of the different regression algorithms was evaluated using the same evaluation methodology described above, and in the case of negative values when fitting the algorithms for some features, we also set breakpoints to better represent all the R2 values.
We initially employed the AdaBoost regression machine learning algorithm, combined with the 12 different feature sets from Table 1, to predict the optimal catalytic temperature of phosphatases. We observed that when considering the five feature sets (1, 4, 5, 7, 10) listed in Table 1, the R2 values based on complete sequence were higher than those based on the sequence without conserved amino acids, while R2 values were higher based on the sequence without conserved amino acids for the other seven feature sets. However, the highest predictive accuracy (R2 = 0.55) was achieved based on the complete sequence and employed the feature combination from the 10th group in Table 1 (Figure 2b).
Using the linear regression machine learning algorithm for prediction, it was found, only using the 1st group in Table 1 as a feature, that the R2 value based on the complete sequence was higher than that based on the sequence without conserved amino acids, while R2 values based on the sequence without conserved amino acids were all higher than those based on the complete sequence. In addition, most of the R2 values based on the complete sequence were negative. When employing the feature of the 7th group in Table 1 to predict the Topt based on sequence without conserved amino acids, it achieved the highest predictive accuracy with an R2 value of 0.55 on the test set (Figure 2c).
Using the K-nearest neighbors regression machine learning algorithm, we observed that only when considering the eight feature sets listed in Table 1 (2, 3,4, 5, 8, 9, 10, 11), the prediction based on the complete sequence showed higher accuracy compared to the scenario where conserved amino acids were excluded. However, the best-performing prediction occurred based on the sequence without conserved amino acids, with the 7th group of features from Table 1 achieving the highest R2 prediction accuracy of 0.755 on the test set (Figure 2d).
When using the random forest regression machine learning algorithm, we found that most of the predictions based on sequences without conserved amino acids are more accurate than those based on the complete sequence. The best-performing model utilized only non-conservative amino acid sequences with the feature combination from the 9th group of Table 1, achieving the highest R2 prediction accuracy of 0.49 on the test set (Figure 2e).
Using the least angle regression machine learning algorithm, when considering the 6th, 7th, and 12th feature sets in Table 1, the model performs better based on the complete sequence than the sequence without conserved amino acids. The optimal model utilized the complete sequence and employed the 7th feature set from Table 1. It achieved the highest predictive accuracy with an R2 of 0.566 on the test set (Figure 2f).
Throughout the entire training process, we observed that for both the complete sequence and the sequence with conserved amino acids removed, most models demonstrated higher predictive accuracy with the sequence excluding conserved amino acids. The best of these prediction models were all trained by the K-nearest neighbor regression algorithm using a combination of amino acid frequencies and protein molecular weight features. Compared to the best-predicting accuracy based on a complete sequence (R2 = 0.599), this optimal model improves the prediction accuracy to 0.755 based on the sequence without conserved amino acids (Wilcoxon test, p-value = 0.006), which is significantly higher than the accuracy of the complete sequence.
Subsequently, we utilized the combination of amino acid frequency and protein molecular weight to fit the K-nearest neighbors’ regression algorithm, predicting the Topt of unmeasured phosphatases by removing conserved amino acids.

2.5. Prediction of Topt of Unmeasured Phosphatases by Optimized Machine Learning Model and Experimental Validation

We randomly selected 10 phosphatases with UniProt IDs of O58216, Q5SLK1, E8UU74, Q97C22, Q72I84, Q5SKD4, A0LR15, Q8Z989, O59374, and Q5SI65 (Table 2). The dataset comprising the complete sequences of these 10 phosphatases was named Topt10 (Supplementary Data, Table S9). These phosphatases have not been experimentally determined for Topt. The specific process for handling the above phosphatase data is illustrated in Figure 3a. Firstly, we need to determine the EC numbers of these phosphatases. We constructed 10 phylogenetic trees by aligning the sequences of these phosphatases with the sequences in the model training set (Supplementary Material, Figure S4). Based on the EC numbers of phosphatase sequences with relatively short evolutionary distances to the target phosphatase, we determined the EC number of the target phosphatase (Table 2). Subsequently, the target sequence was aligned with sequences from the same EC number, and conservative amino acids were removed from the target sequence (Supplementary Data, Table S10). Thus, the other dataset (Topt_rcaa10) was obtained, consisting of the 10 sequences with conservative amino acids removed (Supplementary Data, Table S11). Next, we extracted the amino acid frequency and protein molecular weight information from the phosphatase sequences and then utilized the K-nearest neighbors regression algorithm to predict the Topt (Table 2) of phosphatase sequences in the two datasets (Topt10 and Topt_rcaa10). Meanwhile, through experiments, the experimental Topt values for the 10 phosphatases were determined to be 78 °C, 80 °C, 75 °C, 65 °C, 83 °C, 78 °C, 67 °C, 37 °C, 78 °C, and 70 °C (Supplementary Material, Figure S5), respectively. By comparing the predicted values with the experimental values, we calculated the root mean square error (RMSE) and R-square (R2). We observed that the prediction results based on the sequence without conserved amino acids (R2 = 0.389, RMSE = 13.2) were significantly better than those based on the complete sequence of phosphatases (R2 = 0.03, RMSE = 30.7) (Figure 3).
Interestingly, we observed that the 10 randomly selected test samples predominantly originated from EC numbers with fewer training samples (e.g., EC 3.3.3.102, 3.3.3.104, 3.3.3.70, 3.3.3.3, 3.3.3.18), posing a challenge to our model. This task was more difficult compared to selecting samples from EC numbers with a larger sample size. The test results of the final 10 samples indicated that our approach still maintained a reasonable level of prediction accuracy, demonstrating the model’s good generalization capability.

3. Materials and Methods

3.1. Prediction Model for Optimal Catalytic Temperature of Phosphatases

The basic process of this research work includes the following six steps. (1) Data collection and processing: Download phosphatase data from the BRENDA and UniProt databases, group by the fourth digit of the EC number for multiple sequence alignment, and remove conserved amino acids. The issue of too few conserved amino acids in the group was addressed by constructing a sequence similarity network (SSN). Finally, datasets containing sequences with conserved amino acids were removed, and complete sequences were obtained. (2) Training set and test set division: The dataset was divided into five temperature ranges. From the temperature range containing the least data, 1/4 of the data was taken, and an equal amount was taken from each of the other temperature ranges to form the test set. The remaining data was used as the training set. Before each feature extraction, the training set was resampled. (3) Feature extraction and selection: Multiple features including amino acid frequency, dipeptide frequency, protein molecular weight, and protein descriptors (conjoint triad features and distribution descriptors) were computed for each phosphatase sequence. (4) Model selection and training: Various machine learning algorithms were fitted with different feature combinations, and parameter tuning was conducted. (5) Model evaluation: The performance of the prediction models was assessed using ten-fold Monte Carlo cross-validation. (6) Experimental validation: Predictions were made for the unmeasured experimental values of phosphatase sequences using the optimal model, followed by experimental validation. The framework flowchart is illustrated in Figure 4.

3.2. Dataset Construction

3.2.1. Construction of the Dataset for Wild-type Sequences of Phosphatases

In this study, we obtained phosphatase data from the BRENDA (https://www.brenda-enzymes.org) (accessed on 15 March 2022) [22] and UniProt (https://www.UniProt.org/) (accessed on 15 March 2022) [23] databases. Firstly, we developed a custom Python script [24] to extract the EC numbers (EC 3.1.3.X) of 113 phosphatases from the webpage of the BRENDA database (https://www.brenda-enzymes.org/all_enzymes.php) (accessed on 15 March 2022). Subsequently, we constructed a URL (https://www.brenda-enzymes.org/enzyme.php?ecno=) (accessed on 17 March 2022) based on these EC numbers and retrieved and saved the information web pages of the phosphatases using custom scripts for the application programming interface. We parsed the web pages of the phosphatases using custom scripts, analyzed the page content, and extracted the UniProt ID and temperature information of the phosphatases from the ‘TEMPERATURE OPTIMUM’ table. Phosphatases without temperature records were directly removed. Finally, we constructed a URL (https://www.uniprot.org/uniprotkb) (accessed on 20 March 2022) fiusing the UniProt ID of the phosphatases to retrieve the corresponding amino acid sequences in FASTA format from the UniProt database. After removing duplicate phosphatases, we obtained a total of 249 phosphatases from Brenda and UniProt.

3.2.2. Construction of Phosphatase Dataset with Conserved Amino Acids Removed

Divide the obtained phosphatase data into groups based on the fourth digit of the EC number. By calling the ClustalOmegaCommandlinemethod from the Bio.Align.Applications.ClustalOmegaCommandline module and setting the similarity threshold to threshold = 1, we conducted a multiple sequence alignment of the phosphatases within the same group [25,26]. This method inserts gaps in shorter sequences during the alignment process to ensure that the number of amino acids in the protein sequences are equal, generating intermediate FASTA files for sequence alignment. Subsequently, we used the AlignIO.readmethod from Biopython to read this FASTA file, thereby obtaining the UniProt ID corresponding to the aligned sequences, the position indices of conserved amino acids in the original sequence, and the respective information of these conserved amino acids. Finally, we removed the corresponding conserved amino acids based on the UniProt ID of the phosphatase and the position indices of conserved amino acids in the original sequence.
Based on the grouping, we found that some groups contained only one phosphatase. Since multiple sequence alignment could not be performed in such cases, we removed these entries from the dataset. At the same time, some groups contained two phosphatase data, where the number of remaining amino acids after removing the conserved amino acids from the original sequence was less than 50. We removed these sequences from the dataset to ensure the training set’s quality and accuracy. In the end, a total of 17 data points were removed.
On the other hand, within the groups with a larger number of enzymes, as the number of aligned sequences increases, the number of conserved amino acids also decreases significantly. In these groups with a large number of enzymes, we utilized the ENZYME FUNCTION INITIATIVE TOOLS (EFI) online network tool (http://efi.igb.illinois.edu/efi-est/) (accessed on 7 October 2022) [27] to construct a sequence similarity network (SSN) [28]. We further grouped and aligned the sequences based on gene clusters. First, we compiled the phosphatase enzyme sequences from the same group into a FASTA file and uploaded it to EFI. We then conducted a protein sequence similarity analysis using the default parameters. Based on the results returned by the website, we adjusted the similarity threshold (Supplementary Material, Table S1) to obtain a sequence similarity network with the fewest isolated nodes and generated the corresponding XGMML file.
Subsequently, we visualized the generated sequence similarity network XGMML file using Cytoscape (V 3.10.0). In the network, each node represents a protein, and the corresponding UniProt ID is displayed on the node. We then used each cluster of enzymes within the network as a new sub-group, followed by sequence alignment and removal of conserved amino acids. Clusters with only one node in the sequence similarity network were removed, resulting in a total of 14 data points being eliminated.
Due to the removal of some phosphatase enzymes during the sequence alignment process, to ensure that the training and test sets are the same before and after removing conserved amino acids, we constructed a dataset containing the complete sequences identical to the dataset without conserved amino acids. Each dataset contains 218 phosphatase enzyme entries.

3.3. Training and Testing Set Splitting

The data of optimal catalytic temperatures for phosphatases follow a normal distribution, albeit with an imbalance, and the majority of phosphatases exhibit optimal catalytic temperatures around 37 °C. We analyzed the temperature information in the dataset and divided the entire dataset into five sets based on the target value (Topt): [0–30 °C, 30–50 °C, 50–65 °C, 65–85 °C, 85–100 °C] [5]. Then, we counted the number of data contained in these five sets. From the set with the fewest data points (85–100 °C), we selected 1/4 of the samples. The same number of samples was selected from each of the other sets to form the test set, and the remaining data was used for the training set.
To mitigate the issue of phosphatase data imbalance, we applied the regression resampling strategy proposed by Gado et al. [5] to the training set. By utilizing the random_oversamplemethod from the resreg package, with parameters cl = 21 and ch = 64, we divided the data with temperatures less than 21 °C and greater than 64 °C into the rare domain, while the remaining data constituted the normal domain. Then, by setting the ‘relevance_threshold’ parameter to its default value of 0.5, we randomly oversampled half of the samples in the rare domain, ensuring that after sampling, the rare domain contained an equal number of phosphatases as the normal domain.

3.4. Feature Extraction and Selection

We computed multiple sets of features for each protein. Firstly, we computed a set of 1 × 20-dimensional amino acid frequencies (AA) and a set of 20 × 20-dimensional dipeptide frequencies (Dipeptide).
The formula for calculating amino acid frequency is as follows:
f ( t ) = N ( t ) N , t A , C , D , , Y
N is the number of amino acids in the protein sequence, and t is some natural amino acid in the sequence.
The formula for calculating dipeptide frequency is as follows:
F r s = N r s N 1 , r , s A , C , D , , Y
where Nrs represents the number of occurrences of dipeptide rs in the protein sequence, and N represents the total length of the protein sequence.
In addition, we analyzed the phosphatase sequences using the ProteinAnalysis method from the Bio.SeqUtils.ProtParam module in the Biopython library [29] and obtained a set of 1 × 1-dimensional protein molecular weights by invoking the molecular_weightmethod. We then computed two sets of protein descriptors using the PyDPI (V1.0) Python toolkit [30,31]. One set is a 1 × 343-dimensional conjoint triad feature (CTF) obtained by analyzing the phosphatase enzyme sequence using the GetProDesmethod and then invoking the GetTriadmethod. The other set is a 1 × 105-dimensional distribution descriptor (CTDD) obtained by invoking the CalculateD method. We combined single and multiple feature sets and utilized the StandardScalerclass from the scikit-learn (V 0.24.1) package to generate a StandardScalerobject. We then standardized the feature combinations by calling the transformmethod. Then, we fitted five different machine learning algorithms for training and testing [32].

3.5. Construction and Training of Machine Learning Models

In this study, we utilized various regression algorithms available in the Python machine learning package Scikit-learn (V 0.24.1), including linear regression, random forest regression, K-nearest neighbors regression, AdaBoost regression, and least angle regression (Lars CV), to construct suitable regression models for predicting the optimal catalytic temperature of phosphatases.
We employed the ordinary least squares linear regression method in multiple linear regression. For the random forest regression algorithm, we adjusted the number of trees grown and tested different values of n_estimators (ranging from 10 to 100), while keeping other parameters at their default values. In the K-nearest neighbors regression, we considered both uniform weighting and Euclidean distance and tested different values of K (ranging from 1 to 15). We ensemble 50 regression trees as base regression models for the AdaBoost regression algorithm, with each tree considering at least two instances per leaf. As for the Lars CV regression algorithm, we set it up using default parameters.

3.6. Evaluation of Machine Learning Models

Because the data follows a normal distribution, random splitting of the data would still result in an imbalanced distribution between the training and testing sets, leading to performance metrics (R2) being overly weighted towards frequently occurring data and failing to adequately capture performance in the tails of the distribution. Therefore, when evaluating the performance of the regressors, we did not employ traditional k-fold cross-validation techniques. Instead, we utilized 10 iterations of Monte Carlo cross-validation (MCCV) [33] on a nearly uniformly distributed test set. The dataset was divided into training and testing sets ten times. Then, by computing the coefficient of determination (R2) between the predicted and actual values of the test set over ten iterations, the average R2 was calculated to evaluate the model’s predictive performance.

3.7. Enzyme Expression and Purification

To further validate the reliability of our model, based on the growth temperatures of the bacteria, we randomly selected nine phosphatases from bacteria with relatively high growth temperatures and one phosphatase from bacteria with a lower growth temperature (The UniProt IDs are O58216, Q5SLK1, E8UU74, Q97C22, Q72I84, Q5SKD4, A0LR15, Q8Z989, O59374, and Q5SI65) from the study by Huang et al. [34]. These enzymes all lack experimental measurements for Topt. The genes of the aforementioned 10 phosphatases were synthesized and subcloned into the pET28a vector by Suzhou Geneweave Biotech Co., Ltd. (Suzhou, China). Subsequently, the pET28a plasmids containing the target genes were transformed into Escherichia coli TOP10 for plasmid amplification. Subsequently, the plasmids were extracted and transformed into E. coli BL21(DE3) for the expression of phosphatases. The recombinant strains containing the plasmids were cultured in a Luria-Bertani medium at 37 °C. When the OD600 of the culture reached 0.8–1.2, gene expression was induced by adding isopropyl β-D-1-thiogalactopyranoside (IPTG) to a final concentration of 100 μM, followed by further incubation at 16 °C for 18 h. The culture was then centrifuged at 4000× g for 30 min to collect the cells. The enzymes were purified using the Ni-NTA method, and the purity of the proteins was assessed via sodium dodecyl sulfate-polyacrylamide gel electrophoresis (SDS-PAGE).

3.8. Determination of Enzyme Activity and Topt of Phosphatases

The enzyme at a certain concentration was incubated with 1 mM para-nitrophenyl phosphate in a 100 mM Tris-HCl buffer (pH 7.4) containing 5 mM MgCl2 at different temperatures (37–80 °C) for 10 min. The activity of the phosphatase was determined by measuring the amount of para-nitrophenol produced. One unit of phosphatase activity is defined as the amount of enzyme required to release 1 μmol of para-nitrophenol per minute [35]. The absorbance of the product para-nitrophenol at 405 nm was monitored in real-time using a UV-visible spectrophotometer (model: Evolution one). We used bovine serum albumin as the standard and determined the protein concentration using the Bradford method [36]. The temperature at which the enzyme activity was 100% is defined as Topt, with the maximum activity measured for a single enzyme considered as 100%. The enzyme activity at other temperatures can then be expressed as relative enzyme activity relative to the maximum activity.

3.9. Determining the EC Number of Phosphatase Enzymes

We utilized the ClustalW online tool (https://www.genome.jp/tools-bin/clustalw) (accessed on 5 July 2022) [37,38,39] to perform sequence similarity alignment between the sequences of the 10 phosphatase enzymes without annotated optimal catalytic temperatures and the sequences in the model training dataset, obtaining the alignment results in DND format. We employed the MEGA11 software (V 11.0.13) to visualize the DND files containing the alignment results, constructing 10 evolutionary trees. We determined the position of the target phosphatase in the evolutionary tree topology using its UniProt ID and selected phosphatases with relatively shorter evolutionary distances from the target phosphatase. Based on the EC numbers of the related phosphatases, we determined the EC number of the target phosphatase.

4. Discussion

In this study, we proposed a tool that solely relies on protein sequences, without additional experimental conditions, to directly predict the Topt of phosphatases. This tool can assist researchers in rapidly selecting phosphatases suitable for specific processes or experimental conditions.
In model development, we discovered that a basic feature set with the least number of features combined with the KNN algorithm provided superior prediction outcomes. This success is partly due to the significance of amino acid frequency, supported by Li et al.s’ [4] research, which identified amino acid frequency and bacterial growth temperature as key features with the former accounting for half of the model’s performance. In this study, we found that, in addition to amino acid ratio, molecular weight is a potent predictor for KNN, substantially enhancing accuracy (R² rose from 0.46 to 0.755). This may be because KNN excels in identifying similar neighborhoods in low-dimensional spaces with simple features, leading to more precise predictions.
Additionally, it is interesting to note that by removing conserved amino acids from the protein sequences, we achieved better accuracy in predicting the optimal catalytic temperature of phosphatases. Removing conserved amino acids improves the model’s predictive performance (R2) by approximately 15%. Finally, we used the optimal model to predict the Topt of some unknown phosphatases. The results show that removing conserved amino acids reduced RMSE by approximately 57.04% from 30.7 to 13.2. Therefore, we speculate that conserved and non-conserved amino acids may have varying degrees of influence on enzyme thermostability. This study also lays the groundwork for utilizing machine learning to predict the optimal catalytic temperature of other enzyme families by removing conserved amino acids from their sequences.
Due to the limited number of reported optimal catalytic temperatures for phosphatases in the literature, this study faced the challenge of having a relatively small dataset for training the model. Furthermore, there is currently a lack of research findings regarding highly correlated features with the optimal catalytic temperature of enzymes. If more experimental data and richer feature information were available, such as from aspects like secondary structure content [40], three-dimensional structures [41], and the physicochemical properties of amino acids [42], we hypothesize that the predictive accuracy of Topt could be further improved. In this study, we aimed to predict the optimal catalytic temperature solely based on sequence information as it could help us screen the desired enzyme quickly. In future work, we will consider predicting Topt based on the 3-D structure of enzymes using AI tools, such as AlphaFold, as well as other features related to the optimal catalytic temperature of the enzyme. On the other hand, we will actively gather more optimal catalytic temperature data from other enzyme families. Applying our validated approach of removing conserved amino acids from protein sequences significantly improves the accuracy of machine-learning predictions for the optimal catalytic temperature of enzymes. This conclusion guides us in developing prediction tools for the Topt of other enzyme families.

5. Availability of Final Model, Data, and Code

All phosphatase sequence files and related datasets can be found in the GitHub repository at https://github.com/cyinyin/Tpho/tree/main/data (accessed on 28 March 2024). This includes the Topt249, Topt218, and Topt_rcaa218 datasets. Additional Topt10 and Topt_rcaa10 datasets for evaluating model performance can be found in the GitHub repository at https://github.com/cyinyin/Tpho/tree/main/data/Experimentalvalidationdata (accessed on 28 March 2024).

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/ijms25116252/s1.

Author Contributions

Conceptualization, X.L. and C.Y. (Chun You); methodology, Y.C., D.Y. and C.Y. (Chunhe Yang); data curation, Y.C., B.Q., X.N., L.F. and Y.Q.; writing—original draft preparation, Y.C. and B.Q.; writing—review and editing, H.M., X.L. and C.Y. (Chun You); project administration, C.Y. (Chun You). All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Key R&D Program of China (grant numbers 2023YFA0914000 and 2021YFA0910601, to C.Y.), the National Natural Science Foundation of China (grant numbers 32271473, to C.Y., 12326611, to L.X.), Tianjin Synthetic Biotechnology Innovation Capacity Improvement Projects (TSBICIP-PTJJ-007 and TSBICIP-PTJJ-012, to L.X.).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The standalone version and related data are provided at https://github.com/cyinyin/Tpho (accessed on 28 March 2024).

Acknowledgments

The numerical calculations in this study were carried out on an ORISE Supercomputer.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Bornscheuer, U.T.; Huisman, G.W.; Kazlauskas, R.J.; Lutz, S.; Moore, J.C.; Robins, K. Engineering the third wave of biocatalysis. Nature 2012, 485, 185–194. [Google Scholar] [CrossRef] [PubMed]
  2. Nicolas, J. Artificial intelligence and bioinformatics. In A Guided Tour of Artificial Intelligence Research: Volume III: Interfaces and Applications of Artificial Intelligence; Springer: Berlin/Heidelberg, Germany, 2020; pp. 209–264. [Google Scholar]
  3. Dou, Z.; Sun, Y.; Jiang, X.; Wu, X.; Li, Y.; Gong, B.; Wang, L. Data-driven strategies for the computational design of enzyme thermal stability: Trends, perspectives, and prospects: Data-driven strategies for enzyme thermostability design. Acta Biochim. Biophys. Sin. 2023, 55, 343. [Google Scholar] [CrossRef] [PubMed]
  4. Li, G.; Rabe, K.S.; Nielsen, J.; Engqvist, M.K. Machine learning applied to predicting microorganism growth temperatures and enzyme catalytic optima. ACS Synth. Biol. 2019, 8, 1411–1420. [Google Scholar] [CrossRef] [PubMed]
  5. Gado, J.E.; Beckham, G.T.; Payne, C.M. Improving enzyme optimum temperature prediction with resampling strategies and ensemble learning. J. Chem. Inf. Model. 2020, 60, 4098–4107. [Google Scholar] [CrossRef] [PubMed]
  6. Li, G.; Buric, F.; Zrimec, J.; Viknander, S.; Nielsen, J.; Zelezniak, A.; Engqvist, M.K. Learning deep representations of enzyme thermal adaptation. Protein Sci. 2022, 31, e4480. [Google Scholar] [CrossRef] [PubMed]
  7. Ariaeenejad, S.; Mousivand, M.; Moradi Dezfouli, P.; Hashemi, M.; Kavousi, K.; Hosseini Salekdeh, G. A computational method for prediction of xylanase enzymes activity in strains of Bacillus subtilis based on pseudo amino acid composition features. PLoS ONE 2018, 13, e0205796. [Google Scholar] [CrossRef] [PubMed]
  8. Meng, D.; Wei, X.; Zhang, Y.H.P.J.; Zhu, Z.; You, C.; Ma, Y. Stoichiometric conversion of cellulosic biomass by in vitro synthetic enzymatic biosystems for biomanufacturing. ACS Catal. 2018, 8, 9550–9559. [Google Scholar] [CrossRef]
  9. You, C.; Shi, T.; Li, Y.; Han, P.; Zhou, X.; Zhang, Y.H.P. An in vitro synthetic biology platform for the industrial biomanufacturing of myo-inositol from starch. Biotechnol. Bioeng. 2017, 114, 1855–1864. [Google Scholar] [CrossRef] [PubMed]
  10. Meng, D.; Wei, X.; Bai, X.; Zhou, W.; You, C. Artificial in vitro synthetic enzymatic biosystem for the one-pot sustainable biomanufacturing of glucosamine from starch and inorganic ammonia. ACS Catal. 2020, 10, 13809–13819. [Google Scholar] [CrossRef]
  11. Li, Y.; Shi, T.; Han, P.; You, C. Thermodynamics-driven production of value-added D-allulose from inexpensive starch by an in vitro enzymatic synthetic biosystem. ACS Catal. 2021, 11, 5088–5099. [Google Scholar] [CrossRef]
  12. Lynch, M.; Force, A. The probability of duplicate gene preservation by subfunctionalization. Genetics 2000, 154, 459–473. [Google Scholar] [CrossRef] [PubMed]
  13. Liu, Y.; Huangfu, J.; Qi, F.; Kaleem, I.; E, W.; Li, C. Effects of a non-conservative sequence on the properties of β-glucuronidase from Aspergillus terreus Li-20. PLoS ONE 2012, 7, e30998. [Google Scholar] [CrossRef] [PubMed]
  14. Xu, B.L.; Dai, M.; Chen, Y.; Meng, D.; Wang, Y.; Fang, N.; Tang, X.F.; Tang, B. Improving the thermostability and activity of a thermophilic subtilase by incorporating structural elements of its psychrophilic counterpart. Appl. Environ. Microbiol. 2015, 81, 6302–6313. [Google Scholar] [CrossRef] [PubMed]
  15. Guo, Y.; Tu, T.; Zheng, J.; Bai, Y.; Huang, H.; Su, X.; Wang, Y.; Wang, Y.; Yao, B.; Luo, H. Improvement of Bs APA aspartic protease thermostability via autocatalysis-resistant mutation. J. Agric. Food Chem. 2019, 67, 10505–10512. [Google Scholar] [CrossRef] [PubMed]
  16. He, H.; Garcia, E.A. Learning from imbalanced data. IEEE Trans. Knowl. Data Eng. 2009, 21, 1263–1284. [Google Scholar]
  17. Krawczyk, B. Learning from imbalanced data: Open challenges and future directions. Prog. Artif. Intell. 2016, 5, 221–232. [Google Scholar] [CrossRef]
  18. López, V.; Fernández, A.; García, S.; Palade, V.; Herrera, F. An insight into classification with imbalanced data: Empirical results and current trends on using data intrinsic characteristics. Inf. Sci. 2013, 250, 113–141. [Google Scholar] [CrossRef]
  19. Treetharnmathurot, B.; Ovartlarnporn, C.; Wungsintaweekul, J.; Duncan, R.; Wiwattanapatapee, R. Effect of PEG molecular weight and linking chemistry on the biological activity and thermal stability of PEGylated trypsin. Int. J. Pharm. 2008, 357, 252–259. [Google Scholar] [CrossRef] [PubMed]
  20. Mckenna, A.; Dubey, S. Machine learning based predictive model for the analysis of sequence activity relationships using protein spectra and protein descriptors. J. Biomed. Inform. 2022, 128, 104016. [Google Scholar] [CrossRef]
  21. Zeldovich, K.B.; Berezovsky, I.N.; Shakhnovich, E.I. Protein and DNA sequence determinants of thermophilic adaptation. PLoS Comput. Biol. 2007, 3, e5. [Google Scholar] [CrossRef]
  22. Schomburg, I.; Jeske, L.; Ulbrich, M.; Placzek, S.; Chang, A.; Schomburg, D. The BRENDA enzyme information system–From a database to an expert system. J. Biotechnol. 2017, 261, 194–206. [Google Scholar] [CrossRef]
  23. UniProt Consortium, T. UniProt: The universal protein knowledgebase. Nucleic Acids Res. 2017, 45, D158–D169. [Google Scholar] [CrossRef] [PubMed]
  24. Li, F.; Yuan, L.; Lu, H.; Li, G.; Chen, Y.; Engqvist, M.K.; Kerkhoven, E.J.; Nielsen, J. Deep learning-based k cat prediction enables improved enzyme-constrained model reconstruction. Nat. Catal. 2022, 5, 662–672. [Google Scholar] [CrossRef]
  25. Sievers, F.; Higgins, D.G. Clustal Omega for making accurate alignments of many protein sequences. Protein Sci. 2018, 27, 135–145. [Google Scholar] [CrossRef] [PubMed]
  26. Sievers, F.; Wilm, A.; Dineen, D.; Gibson, T.J.; Karplus, K.; Li, W.; Lopez, R.; McWilliam, H.; Remmert, M.; Söding, J.; et al. Fast, scalable generation of high-quality protein multiple sequence alignments using Clustal Omega. Mol. Syst. Biol. 2011, 7, 539. [Google Scholar] [CrossRef] [PubMed]
  27. Gerlt, J.A.; Bouvier, J.T.; Davidson, D.B.; Imker, H.J.; Sadkhin, B.; Slater, D.R.; Whalen, K.L. Enzyme function initiative-enzyme similarity tool (EFI-EST): A web tool for generating protein sequence similarity networks. Biochim. Biophys. Acta (BBA)-Proteins Proteom. 2015, 1854, 1019–1037. [Google Scholar] [CrossRef] [PubMed]
  28. Atkinson, H.J.; Morris, J.H.; Ferrin, T.E.; Babbitt, P.C. Using sequence similarity networks for visualization of relationships across diverse protein superfamilies. PLoS ONE 2009, 4, e4345. [Google Scholar] [CrossRef] [PubMed]
  29. Cock, P.J.; Antao, T.; Chang, J.T.; Chapman, B.A.; Cox, C.J.; Dalke, A.; Friedberg, I.; Hamelryck, T.; Kauff, F.; Wilczynski, B.; et al. Biopython: Freely available Python tools for computational molecular biology and bioinformatics. Bioinformatics 2009, 25, 1422. [Google Scholar] [CrossRef] [PubMed]
  30. Cao, D.S.; Liang, Y.Z.; Yan, J.; Tan, G.S.; Xu, Q.S.; Liu, S. PyDPI: Freely Available Python Package for Chemoinformatics, Bioinformatics, and Chemogenomics Studies; ACS Publications: Washington, DC, USA, 2013. [Google Scholar]
  31. Shen, J.; Zhang, J.; Luo, X.; Zhu, W.; Yu, K.; Chen, K.; Li, Y.; Jiang, H. Predicting protein–protein interactions based only on sequences information. Proc. Natl. Acad. Sci. USA 2007, 104, 4337–4341. [Google Scholar] [CrossRef] [PubMed]
  32. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R. Scikit-learn: Machine learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
  33. Molinaro, A.M.; Simon, R.; Pfeiffer, R.M. Prediction error estimation: A comparison of resampling methods. Bioinformatics 2005, 21, 3301–3307. [Google Scholar] [CrossRef] [PubMed]
  34. Huang, H.; Pandya, C.; Liu, C.; Al-Obaidi, N.F.; Wang, M.; Zheng, L.; Toews Keating, S.; Aono, M.; Love, J.D.; Evans, B.; et al. Panoramic view of a superfamily of phosphatases through substrate profiling. Proc. Natl. Acad. Sci. USA 2015, 112, E1974–E1983. [Google Scholar] [CrossRef] [PubMed]
  35. Meng, D.; Liang, A.; Wei, X.; You, C. Enzymatic characterization of a thermostable phosphatase from Thermomicrobium roseum and its application for biosynthesis of fructose from maltodextrin. Appl. Microbiol. Biotechnol. 2019, 103, 6129–6139. [Google Scholar] [CrossRef] [PubMed]
  36. Bradford, M.M. A rapid and sensitive method for the quantitation of microgram quantities of protein utilizing the principle of protein-dye binding. Anal. Biochem. 1976, 72, 248–254. [Google Scholar] [CrossRef] [PubMed]
  37. Thompson, J.D.; Higgins, D.G.; Gibson, T.J. CLUSTAL W: Improving the sensitivity of progressive multiple sequence alignment through sequence weighting, position-specific gap penalties and weight matrix choice. Nucleic Acids Res. 1994, 22, 4673–4680. [Google Scholar] [CrossRef] [PubMed]
  38. Chenna, R.; Sugawara, H.; Koike, T.; Lopez, R.; Gibson, T.J.; Higgins, D.G.; Thompson, J.D. Multiple sequence alignment with the Clustal series of programs. Nucleic Acids Res. 2003, 31, 3497–3500. [Google Scholar] [CrossRef] [PubMed]
  39. Larkin, M.A.; Blackshields, G.; Brown, N.P.; Chenna, R.; McGettigan, P.A.; McWilliam, H.; Valentin, F.; Wallace, I.M.; Wilm, A.; Lopez, R.; et al. Clustal W and Clustal X version 2.0. Bioinformatics 2007, 23, 2947–2948. [Google Scholar] [CrossRef] [PubMed]
  40. Pucci, F.; Rooman, M. Physical and molecular bases of protein thermal stability and cold adaptation. Curr. Opin. Struct. Biol. 2017, 42, 117–128. [Google Scholar] [CrossRef] [PubMed]
  41. Kawashima, S.; Kanehisa, M. AAindex: Amino acid index database. Nucleic Acids Res. 2000, 28, 374. [Google Scholar] [CrossRef] [PubMed]
  42. Magnan, C.N.; Baldi, P. SSpro/ACCpro 5: Almost perfect prediction of protein secondary structure and relative solvent accessibility using profiles, machine learning and structural similarity. Bioinformatics 2014, 30, 2592–2597. [Google Scholar] [CrossRef]
Figure 1. (a). Distribution of phosphatase under different EC groups in our defined datasets. (b). Distribution of phosphatase datasets across different temperature ranges. (c). Distribution of the Topt249 phosphatase training dataset. (d). Distribution of the Topt218 and Topt_rcaa218 phosphatase training datasets.
Figure 1. (a). Distribution of phosphatase under different EC groups in our defined datasets. (b). Distribution of phosphatase datasets across different temperature ranges. (c). Distribution of the Topt249 phosphatase training dataset. (d). Distribution of the Topt218 and Topt_rcaa218 phosphatase training datasets.
Ijms 25 06252 g001
Figure 2. Ten Monte Carlo cross-validation coefficient of determination (R2) results for five regression models fitted with different combinations of features in different datasets. The error bars represent the standard deviation of the R2 scores. R2 values for five regression models fitted with different combinations of features in Topt249 dataset (a), R2 values for AdaBoost regression machine learning algorithm (b), linear regression machine learning algorithm (c), K-nearest neighbors regression machine learning algorithm (d), random forest regression machine learning algorithm (e), and least angle regression machine learning algorithm (f) fitted with different combinations of features in Topt218 and Topt_rcaa218 datasets. Features 1–12 are listed in Table 1 for details.
Figure 2. Ten Monte Carlo cross-validation coefficient of determination (R2) results for five regression models fitted with different combinations of features in different datasets. The error bars represent the standard deviation of the R2 scores. R2 values for five regression models fitted with different combinations of features in Topt249 dataset (a), R2 values for AdaBoost regression machine learning algorithm (b), linear regression machine learning algorithm (c), K-nearest neighbors regression machine learning algorithm (d), random forest regression machine learning algorithm (e), and least angle regression machine learning algorithm (f) fitted with different combinations of features in Topt218 and Topt_rcaa218 datasets. Features 1–12 are listed in Table 1 for details.
Ijms 25 06252 g002
Figure 3. Phosphatase data processing and comparison of predicted and experimental values for Topt10 and Topt_rcaa10 Datasets. RMSE: root mean square error. (a). Processing workflow for predicting the Topt of unknown phosphatase. (b). Comparison between predicted Topt values and experimental Topt values for the Topt10 phosphatase dataset. (c). Comparison between predicted Topt values and experimental Topt values for the Topt_rcaa10 phosphatase dataset.
Figure 3. Phosphatase data processing and comparison of predicted and experimental values for Topt10 and Topt_rcaa10 Datasets. RMSE: root mean square error. (a). Processing workflow for predicting the Topt of unknown phosphatase. (b). Comparison between predicted Topt values and experimental Topt values for the Topt10 phosphatase dataset. (c). Comparison between predicted Topt values and experimental Topt values for the Topt_rcaa10 phosphatase dataset.
Ijms 25 06252 g003
Figure 4. Schematic overview of the machine learning model for predicting the optimal catalytic temperature of phosphatases. AA: amino acid frequencies. Dipeptide: dimer frequencies. Molecular Weight: protein molecular weight. Triad: conjoint triad features. CTDD: distribution descriptor.
Figure 4. Schematic overview of the machine learning model for predicting the optimal catalytic temperature of phosphatases. AA: amino acid frequencies. Dipeptide: dimer frequencies. Molecular Weight: protein molecular weight. Triad: conjoint triad features. CTDD: distribution descriptor.
Ijms 25 06252 g004
Table 1. A summary of different feature combinations and dimensions generated from enzyme sequences used in this study.
Table 1. A summary of different feature combinations and dimensions generated from enzyme sequences used in this study.
No.Feature GroupNumber of Descriptors
1amino acid frequencies20
2dipeptide frequencies400
3amino acid frequencies and dipeptide frequencies420
4amino acid frequencies, protein descriptors (conjoint triad features and distribution)468
5dipeptide frequencies, protein descriptors (conjoint triad features and distribution)848
6amino acid frequencies, dipeptide frequencies, protein descriptors (conjoint triad features and distribution)868
7amino acid frequencies, protein molecular weights21
8dipeptide frequencies, protein molecular weights401
9amino acid frequencies, dipeptide frequencies, protein molecular weights421
10amino acid frequencies, protein molecular weights, protein descriptors (conjoint triad features and distribution)469
11dipeptide frequencies, protein molecular weights, protein descriptors (conjoint triad features and distribution)849
12amino acid frequencies, dipeptide frequencies, protein molecular weights, protein descriptors (conjoint triad features and distribution)869
Table 2. Predicted and experimental Topt of undetermined phosphatases.
Table 2. Predicted and experimental Topt of undetermined phosphatases.
No.UniProt IDEC NumberTopt10
Predicted Topt (°C)
Topt_rcaa10 Predicted Topt (°C)Experimental Topt (°C)
1O582163.1.3.102808078
2Q5SLK13.1.3.104787880
3E8UU743.1.3.3404075
4Q97C223.1.3.70158065
5Q72I843.1.3.102787883
6Q5SKD43.1.3.70786578
7A0LR153.1.3.12306567
8Q8Z9893.1.3.18653737
9O593743.1.3.104227078
10Q5SI653.1.3.102506570
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Cao, Y.; Qiu, B.; Ning, X.; Fan, L.; Qin, Y.; Yu, D.; Yang, C.; Ma, H.; Liao, X.; You, C. Enhancing Machine-Learning Prediction of Enzyme Catalytic Temperature Optima through Amino Acid Conservation Analysis. Int. J. Mol. Sci. 2024, 25, 6252. https://doi.org/10.3390/ijms25116252

AMA Style

Cao Y, Qiu B, Ning X, Fan L, Qin Y, Yu D, Yang C, Ma H, Liao X, You C. Enhancing Machine-Learning Prediction of Enzyme Catalytic Temperature Optima through Amino Acid Conservation Analysis. International Journal of Molecular Sciences. 2024; 25(11):6252. https://doi.org/10.3390/ijms25116252

Chicago/Turabian Style

Cao, Yinyin, Boyu Qiu, Xiao Ning, Lin Fan, Yanmei Qin, Dong Yu, Chunhe Yang, Hongwu Ma, Xiaoping Liao, and Chun You. 2024. "Enhancing Machine-Learning Prediction of Enzyme Catalytic Temperature Optima through Amino Acid Conservation Analysis" International Journal of Molecular Sciences 25, no. 11: 6252. https://doi.org/10.3390/ijms25116252

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop