Next Article in Journal
HK2 Mediated Glycolytic Metabolism in Mouse Photoreceptors Is Not Required to Cause Late Stage Age-Related Macular Degeneration-Like Pathologies
Next Article in Special Issue
Ligands of Adrenergic Receptors: A Structural Point of View
Previous Article in Journal
4,6′-Anhydrooxysporidinone from Fusarium lateritium SSF2 Induces Autophagic and Apoptosis Cell Death in MCF-7 Breast Cancer Cells
Previous Article in Special Issue
DEER Analysis of GPCR Conformational Heterogeneity
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Integrated Multi-Class Classification and Prediction of GPCR Allosteric Modulators by Machine Learning Intelligence

1
Department of Pharmaceutical Sciences, Computational Chemical Genomics Screen (CCGS) Center and Pharmacometrics System Pharmacology Program, School of Pharmacy, University of Pittsburgh, Pittsburgh, PA 15261, USA
2
NIH National Center of Excellence for Computational Drug Abuse Research (CDAR), University of Pittsburgh, Pittsburgh, PA 15261, USA
3
Drug Discovery Institute, Departments of Computational Biology and of Structural Biology, University of Pittsburgh, Pittsburgh, PA 15261, USA
*
Author to whom correspondence should be addressed.
These authors contributed equally.
Biomolecules 2021, 11(6), 870; https://doi.org/10.3390/biom11060870
Submission received: 23 April 2021 / Revised: 30 May 2021 / Accepted: 8 June 2021 / Published: 11 June 2021
(This article belongs to the Special Issue GPCRs: Structure, Biology and Potential Applications)

Abstract

:
G-protein-coupled receptors (GPCRs) are the largest and most diverse group of cell surface receptors that respond to various extracellular signals. The allosteric modulation of GPCRs has emerged in recent years as a promising approach for developing target-selective therapies. Moreover, the discovery of new GPCR allosteric modulators can greatly benefit the further understanding of GPCR cell signaling mechanisms. It is critical but also challenging to make an accurate distinction of modulators for different GPCR groups in an efficient and effective manner. In this study, we focus on an 11-class classification task with 10 GPCR subtype classes and a random compounds class. We used a dataset containing 34,434 compounds with allosteric modulators collected from classical GPCR families A, B, and C, as well as random drug-like compounds. Six types of machine learning models, including support vector machine, naïve Bayes, decision tree, random forest, logistic regression, and multilayer perceptron, were trained using different combinations of features including molecular descriptors, Atom-pair fingerprints, MACCS fingerprints, and ECFP6 fingerprints. The performances of trained machine learning models with different feature combinations were closely investigated and discussed. To the best of our knowledge, this is the first work on the multi-class classification of GPCR allosteric modulators. We believe that the classification models developed in this study can be used as simple and accurate tools for the discovery and development of GPCR allosteric modulators.

1. Introduction

G-protein-coupled receptors (GPCRs) are the largest family of membrane proteins in the human genome and regulate a variety of extracellular signal transduction pathways, including photons, ions, hormones, neurotransmitters, odorants, and other stimuli [1,2]. Based on similarity and diversity of amino acid sequences and functions, GPCRs can be categorized into three main subfamilies, termed A, B, and C. As GPCRs are therapeutic targets for a broad spectrum of diseases, they have been of long-standing interest as therapeutic targets, and account for ~34% of the global market share of therapeutic drugs [3,4]. Most of the GPCR-targeted drugs are functionally active by binding to the orthosteric site of the receptor, which is the pocket bound by endogenous activating ligand [5]. However, an increasing number of drugs targeting the orthosteric sites have been withdrawn from the market due to low efficacy and undesired side effects [6,7]. The major issue that is associated with the orthosteric ligand is that their binding sites are often highly conserved across a single GPCR subfamily, making it difficult to achieve high selectivity for specific GPCR subtypes [8]. In recent years, great attention has been devoted to the discovery of drugs targeting GPCRs as allosteric modulators [9,10,11]. These small molecules bind to a site (allosteric site) that is topographically distinct from the orthosteric site of the GPCR protein and thus do not compete with orthosteric ligands [12]. Compared to the highly conserved orthosteric sites, allosteric binding pockets are more diverse across the same subfamily of GPCRs. This mechanism allows allosteric modulators to confer subtype selectivity. Meanwhile, the allosteric modulators also show a preferable safety profile due to the ‘ceiling’ effect [5,13,14]. In addition, Yiran Wu et al. added that the distinct pathways and allosteric pockets may enable allosteric modulators’ cooperativities among different protein subtypes [15]. Therefore, it is important to develop allosteric modulators as both therapeutic agents and research tools to bring new opportunities to drug discovery towards GPCRs and have an in-depth understanding of receptor modulation mechanisms.
Drug discovery is expensive. The process requires molecule design, lead optimization, in vitro and in vivo data analysis [16]. High-throughput screening (HTS) is a modern technique that is commonly used to facilitate the discovery of the allosteric modulators of GPCRs [17]. However, it is costly and often has been plagued with problems of high false-positive rates [18]. Alternatively, computational approaches, including homology modeling, molecular docking, and molecular dynamics simulation have been applied to aid drug discovery of novel allosteric modulators [19,20,21,22]. Yet, developing in silico screening methods that attain high accuracy remains challenging. At present, there is still an urgent demand for computational tools that can identify allosteric drugs from inactive random compounds and increase the chances of success in the development of allosteric modulators as lead clinical compounds.
While there are many traditional cheminformatic tools to assist the development of allosteric modulators, there have been few examples using machine learning (ML). ML has emerged as a promising pillar to promote data-driven decision-making, facilitate the process, and reduce the failure rates in drug discovery and development [23,24,25,26,27,28,29,30]. Kumar et al. developed multiple in silico models (Support Vector Machines (SVM), k-nearest neighbor algorithms, partial least square (PLS), etc.) to predict human intestinal absorption of diverse chemicals [31]. Jacob and Vert [32] used tensor-product-based features and applied SVM to predict protein-ligand interactions. Remarkably, in 2020, Google’s DeepMind’s AlphaFold made a scientific breakthrough for its astonishing performance on predicting protein 3D structures by using deep learning approaches [33,34].
We have previously reported an application of developing ML-based classification models for the prediction of orthosteric and allosteric regulations on cannabinoid receptors [13]. To expand the application to a broader scope of GPCR families and to handle a more diversified chemical space, in this paper, we proposed an 11-class classification task to discriminate allosteric modulators among different subtypes of GPCRs A, B, C subfamilies, and inactive compounds simultaneously. Diverse types of molecular features and multiple machine learning algorithms were applied for model training. The combinations of different types of molecular features and ML algorithms were carefully investigated to search which set of features works best for a specific classifier. The performance of trained ML models was systematically evaluated by using different metrics. This research gives the first report on the multi-class classification of GPCR allosteric modulators. The study can be of value for facilitating in silico screening and providing guidance for future discovery and development of GPCR allosteric modulators.

2. Materials and Methods

2.1. Data Collection and Preparation

In this study, the allosteric database (ASD) [35] was used for collecting GPCR allosteric modulators. Classical GPCR subfamilies A, B, and C were selected as targets to collect allosteric modulators. So far, the number of discovered allosteric modulators across different GPCR subtypes is of high variance. To construct a dataset that mimics this nature, we selected some common subtypes that are of a distinct number of allosteric modulators from each GPCR subtype. The number of collected allosteric modulators from each subtype is shown in Table 1. We also collected drug-like random compounds from the ZINC database [36] to serve as inactive decoy compounds. More than eight thousand drug-like compounds were randomly collected and integrated into the GPCR allosteric modulator datasets. Allosteric modulators collected from 10 GPCR subtypes were combined with the inactive compounds to make up the final dataset containing 34,434 compounds (Table 1).

2.2. Molecular Fingerprint and Descriptor Calculation

Both molecular descriptors and molecular fingerprints were used as molecular representations for all compounds in the datasets. A total of 119 molecular descriptors (ExactMW, SlogP, TPSA, NumHBD, NumHBA, etc.,), which characterize the physicochemical properties of the studied compounds were calculated using RDKit (http://www.rdkit.org/, accessed on 29 October 2020). Three different types of molecular fingerprints, Atom-pair fingerprints [37], MACCS fingerprints [38], and ECFP6 fingerprints [39] were calculated with a CDK toolkit [40]. Atom-pair fingerprints were encoded as standard bit vectors of length 1024 based on the atomic environments and shortest path separations of every atom pair in the molecule. MACCS fingerprints consist of 166-bit fingerprints representing the presence or absence of 166 substructural keys. ECFP6 are circular topological fingerprints that represent circular topological atom neighborhoods with 1024 descriptors.

2.3. Model Building

Here, six ML algorithms were employed to develop the classification models to discriminate between different subtypes of GPCR allosteric modulators and inactive drugs, including support vector machine (SVM), neural network/multilayer perceptron (MLP), decision tree (DT), random forest (RF), naïve Bayes (NB), and logistic regression. The open-source scikit-learn (http://scikitlearn.org/, accessed on 14 November 2020) was used for model building, tuning, validation, and result interpretation.
Support vector machine (SVM) [41] is a kernel-based algorithm widely used for binary classification and regression tasks. Each chemical structure was described as a binary string and worked as an eigenvector for SVM. The eigenvector was trained using the SVM algorithm, which results in a decision function for classification. The svm.SVC() method with three kernel functions (linear, rbf, poly) from scikit-learn was applied. The penalty parameter C and parameter γ for rbf and poly kernels were tuned on the training set by five-fold cross-validation using the grid search strategy.
Multilayer perceptron (MLP) [42] is a type of fully connected, feed-forward artificial neural network (ANN), consisting of three types of layers: the input layer, hidden layer, and output layer. An arbitrary number of hidden layers between the input and output layer is the true computational engine of the MLP. For each hidden layer, different numbers of hidden neurons can be assigned. MLP is based on calculating the values of hidden neurons in a current layer as the activated summation of weighted outputs of hidden neurons from a previous layer. The weights of the neuron connections are initially random but then adjusted through the backward propagation learning algorithm. Similar to SVM, we used grid search to optimize the hyperparameter for the MLPClassifier() method in scikit-learn. We searched for the optimal number of layers, number of hidden units, activation function (identity, logistic, tanh, ReLu), regularization parameter (0.0001, 0.001, 0.01, 0.1), and learning rate (0.1, 0.01, 0.001, 0.0001).
Decision tree (DT) [43] is a simple classic supervised learning method used to solve classification and regression problems. At each node of the tree, the attribute that gives the greatest information gain is chosen to make the decision. The data are divided in the most homogeneous way. Then the process is repeated on each smaller subset in a recursive manner. DecisionTreeClassifier() was applied for generating models with tuning on max_depth.
Random forest (RF) [44] is an ensemble method that leverages the power of a multitude of decision trees. In classification, regression, or other tasks, the final output is obtained by averaging the results of classification and regression trees that are grown on bootstrap samples. RandomForestClassifier() was applied. The best model was saved after the tuning on n_estimators (10, 50, 100) and max_depth (2, 3, 4, 5).
Naïve Bayes (NB) [45] is a simple probabilistic classifier based on Bayes’ theorem for conditional probability. This algorithm assumes that the attributes in a dataset are independent of each other. In other words, the NB classifier ignores the possible dependencies among the inputs and reduces a multivariate problem to a group of univariate problems.
Logistic regression (LR) [46] is a classification algorithm used for the prediction of the outcome of a categorical dependent variable from a set of predictor or independent variables. It is mainly used for prediction also calculating the probability of success. LogisticRegression() was applied and tuned with penalty (l1, l2, elasiticnet).

2.4. Chemical Space Analysis

The classification tasks in this study are completely implemented by machine intelligence without using any chemical or pharmacy domain knowledge. To investigate the possible relationships between the molecular properties or fingerprints and classification tasks, we visualized the chemical space distribution of each class by using the t-Distributed Stochastic Neighbor Embedding (t-SNE) method. Atom-pair, ECFP6, MACCS fingerprints, and molecular descriptors were used to represent molecules and were decomposed into two dimensions by t-SNE. The scikit-learn was applied for t-SNE analysis. Matplotlib library was used for plotting.
t-Distributed stochastic neighbor embedding (t-SNE) [47] is a nonlinear embedding technique developed by van der Maaten, L. and G. Hinton in 2008. It was used as a dimension reduction method well-suited for embedding high-dimensional data into a low-dimensional space of two or three dimensions [48]. t-SNE calculates the similarity measure between pairs of instances in the high dimensional space and the low dimensional space. The pairwise similarities of points were converted to joint probabilities. In this process, the Kullback–Leibler divergence between the joint probabilities of the low-dimensional embedding and the high-dimensional data was minimized. This neighbor embedding property makes t-SNE effective for identifying local clusters in the data.

2.5. Model Evaluation

In this study, the dataset was randomly divided into a training set (80%) and test set (20%) using a stratified sampling method. Then 80% of the training sets were randomly selected and used for training. The remaining compounds were used as a validation set. Various evaluation metrics were calculated to evaluate the performance of different ML models, including accuracy (ACC), balanced accuracy (Bal_ACC), precision, recall, f1-score, area under the receiver operating characteristic (ROC) curve (AUC) score, Cohen’s κ (CK) [49], and Matthews correlation coefficient (MCC) [50].
Micro-average and macro-average are two strategies being used for multiclass classification tasks. Here, we applied the macro-average method to calculate the precision, recall, and f1-score. Macro-average calculates each metric for each label separately and returns their unweighted mean. An ML model with good performance on macro-averaging metrics means it can recognize each class perfectly, even on small classes, which is a suitable case for our study. As for micro-averaging, we are aggregating the contributions of all classes to compute the average metric, which only emphasizes the performance of majority classes. The macro-averaged metric for each class is defined as:
M ¯ = i = 1 C M i · 1 C
where M is the current metric, C is the number of total classes for the classification task, i denotes the ith class.
The following abbreviations are used for metric definitions: the number of true positives (TP), the number of false positives (FP), the number of true negatives (TN), and the number of false negatives (FN). The AUC score is one of the most widely used metrics that measures the overall performance of a classification model. It ranges between 0.5 and 1. A model with an AUC of 1 means perfect separation whereas an AUC of 0.5 means the model has no class separation capacity.
ACC is a measure of systematic error. It is the number of correctly predicted data points out of all data points.
ACC   = TP + FN TP + TN + FP + FN
Recall is also called the true positive rate or sensitivity, which measures the ability of a classifier to find all of the positive samples.
recall   =   TP     TP   +   FN  
Precision is known as the positive predictive value, which is the proportion of the predicted true label among all the retrieved instances.
precision   = TP TP + FP
The f1-score is a weighted average of the precision and recall and takes both false positives and false negatives into account.
f 1 score = 2 · precision · recall precision + recall
CK is used to estimate overall model performance by measuring the proximity of the predicted classes to the actual classes when compared to a random classification.
CK   = ACC     Pe 1   Pe
where:
Pe = ( TP + FP ) × ( TP + FN ) + ( TN + FP ) × ( TN + FN ) ( TP + TN + FP + FN ) 2
When the dataset is imbalanced, Bal_ACC can be used to evaluate the general performance of an algorithm. It avoids inflated performance estimates and is computed as the average between sensitivity and specificity.
Bal _ ACC   = [ TP TP + FN + TN TN + FP ] · 1 n
where n represents the total number of classes.
MCC is another useful metric when the dataset has varying classes and is imbalanced. It is a correlation coefficient between observed and predicted classes and has a range between −1 and 1. A value of −1 indicates a completely wrong prediction, while a coefficient of 1 indicates a perfect prediction.
MCC = TP × TN FP × FN ( TP + FP ) ( TP + FN ) ( TN + PP ) ( TN + FN )

3. Results and Discussion

3.1. Overall Workflow

In this study, four subtypes from class A GPCRs, three subtypes from class B GPCRs, and three subtypes from class C GPCRs were collected from the ASD database. Inactive compounds were randomly selected from the ZINC database to cover a large drug-like chemical space. The allosteric modulators from 10 GPCRs subtypes and one random drug-like compounds class (decoys class) were used to construct the dataset for our 11-class classification task (Table 1, Figure 1A). Atom-pair, ECFP6, MACCS fingerprints, and molecular descriptors were calculated from the constructed dataset to represent four types of features. Different types of features can be used to evaluate the properties of the compounds from diverse aspects, which may affect the performance of ML models. Therefore, besides using one type of feature at each time for model training, we also paired fingerprints with molecular descriptors, as well as combining all four feature types to cover the best possible combinations (Figure 1A). Eight new datasets were generated from the different feature combinations. A total of 11 classes were labeled for classification from 0 to 10, respectively. Six supervised ML algorithms were applied to build classifiers for each dataset. We conducted a five-fold cross-validation for each dataset to select the best-performing model (Figure 1B). The ML models’ performance was evaluated by ACC, precision, recall, f1-score, MCC, CK, Bal_ACC, and AUC values.

3.2. Data Set Analysis

To investigate the possible relationships between the physicochemical properties or fingerprints and classification tasks, the chemical space distribution of each class was analyzed. t-SNE was used to decompose the molecular descriptors and fingerprints into two dimensions for visualizing. Figure 2, Figure 3 and Figure 4 show the results of the chemical space distribution of compounds in the dataset, which are represented by three types of fingerprints and molecular descriptors, respectively. As shown in Figure 2, Figure 3 and Figure 4, the blue dots represent drug-like molecules that define the background of the overall chemical property space. All 10 allosteric modulator subtypes fell in the defined chemical spaces (based on molecular fingerprints and properties), indicating that both similar and distinctive drug-like molecules were involved in the classification tasks for target-specific allosteric modulators. In comparison with known allosteric modulators, both similar and distinctive drug-like molecules were presented in the model training and validation processes. Similar molecules challenge the robustness of the model training process while distinctive molecules exhibit the vast chemical space to classifiers. Each of the allosteric modulator subtypes also occupies several specific regions which are distinct from one another, which may indicate their subtype selectivity. The distinct chemical space distribution of different allosteric subtypes shows the feasibility of applying our machine intelligence models.

3.3. Performance Evaluation on Different Feature Types

The ACC, precision, recall, f1-score, MCC, CK, Bal_ACC, and AUC values of all machine learning (ML) models on validation sets are summarized in Tables S1–S3. Radar charts are also plotted to visualize all the above metrics for both training sets and test sets (Figures S1–S4). The results on test sets are shown in Table 2, Table 3 and Table 4.
According to the results, SVM trained on ECFP6 (SVM—ECFP6) outperformed other ML models with the highest scores on AUC, ACC, f1-score, CK, and MCC. It also showed satisfactory results on Bal_AUC (0.950), precision (0.978), and recall (0.950). The good result on Bal_AUC is exciting for an imbalanced dataset. The high f1-score of SVM—ECFP6 is also expected because it is a weighted average of precision and recall. MCC is a reliable and comprehensive assessment of the models’ performance. High MCC values mean that the ML model was able to correctly predict both the positive data instances as well as negative data instances, indicating an outstanding discriminant capability of the models’ performance. The confusion matrix is plotted to summarize the classification result of SVM—ECFP6 (Figure 5) on the test set. The results of all other models are shown in supplementary data (Figures S5–S16). As is shown in the confusion matrix, the separability of SVM—ECFP6 works well on all 11 classes, with most classes being correctly predicted. This model also shows a generalization capability even in small classes. Notably, a total of 17 test cases from the seventh class (PTHrP) was all correctly classified.
By comparing different ML algorithms, while SVM performed well when trained on one fingerprint type, its performance is significantly reduced when trained with molecular descriptors alone or in combination with other types of features. Compared to SVM, MLP shows a more stable performance on all datasets. When trained on datasets with two or more feature types, MLP showed superior overall performance, with AUC values above 0.95. It also showed the best performance on the molecular descriptor dataset, with an AUC value of 0.943. MLP can be considered a subset of deep neural networks (DNN), where its neural network architecture gives a competitive performance on high-dimensional datasets. This may result in the good performance of MLP on these datasets. RF has a balanced performance on most of the datasets. It generally shows good AUC scores but compared to SVM and MLP, the Bal_ACC and recall values are lower, indicating this classifier is more likely to be affected by the imbalanced dataset. LR also shows similar results as RF. NB and DT performed the worst among all ML models, which is also reasonable since they are more simple models and are often used as the baseline for comparison.
The selection of different features will also affect the ML models’ performance. As summarized in Table 5, models trained on the ECFP6 feature showed the best overall results; whereas, the molecular descriptors feature has poor performance compared with all other feature types or feature combinations. The results also showed that models trained on MACCS fingerprint underperformed the results of ECFP6 and Atom-pair fingerprint. The underperformance of MACCS could possibly be due to the fact that MACCS contains less information (166-bit) than ECFP6 (1024-bit) and Atom-pair (1024-bit), which would result in a less robust model. However, a fingerprint containing higher information density does not necessarily mean that it also should achieve a better outcome. Compared to ECFP6, which only contains structural information, Atom-pair is usually considered a hybrid type of fingerprint that contains both atomic and structural information. However, in our case, ECFP6 still shows the best outcome across all GPCR classes. Since there are 10 GPCR protein targets and one random decoy class involved in this study, the good result on ECFP6 means that ML models trained on this fingerprint can have a good impact on the majority of classes of selected GPCR protein targets. To obtain a deeper understanding of the predictive models, we further extracted the SlogP, molecular weight (M.W.), hydrogen bond acceptor (HBA), and hydrogen bond donor (HBD) from the molecular descriptors and conducted pair-wise distribution comparisons on each GPCR class with drug-like molecules, which are shown in Figures S17–S26 in the Supplementary Data. The comparison plots showed that most classes (all except the PTHrP) generally follow Lipinski’s rule of five since the distributions of drug-like molecules overlap with the target GPCR classes. However, the physicochemical properties of allosteric modulators from some of the GPCR classes do form distinguishable distributions from that of drug-like molecules. For example, the distribution of CB1 on M.W. is more concentrated than drug-like molecules with a mode around 500 whereas the mode of drug-like molecules is around 400. The distributions of GLP1-R and GCGR on SlogP are around 5, which is also different from the mode of SlogP for drug-like molecules that is around 4. It is worth noting that compared to the performance on datasets combining two feature types, MLP, LR, and RF all showed enhanced performance when trained on the dataset with all four feature types’ combinations, which is possibly due to the fact that the four feature types can give complementary information that is favored by these ML algorithms.

3.4. Performance Evaluation on Individual GPCR Classes

The f1-score takes the imbalanced data distribution into account and is the harmonic mean of precision and recall. Here we selected the f1-score as a representative metric to evaluate the ML models’ performance on one feature type (Table 6), two feature types (Table 7), and four feature types (Table 8) for each GPCR class. The f1-scores across different GPCR families were also compared and shown in supporting information (Table S4).
From Table 6, we can see that the best model SVM—ECFP6 has generally a high f1-score across all GPCR classes. It has an f1-score over 0.98 on CB1, mAchR M1, S1P3, PTHrP, mGlu2, and mGlu5. The f1-scores on FFA2, GLP1-R, GCGR of SVM—ECFP6 are all over 0.90 but below 0.95. The small sample sizes they have could impede the training process for robust models. PTHrP has a similar sample size (88 samples in total) to FFA2 (89 samples in total). Surprisingly, it has an f1-score of 1, indicating a precise and robust classification. The 88 samples from PTHrP are all polypeptides and share a core scaffold. They are very focused chemicals with small modifications to the core scaffolds. The highly distinguishable pattern may explain the good performance of the ML models in this class. While a large sample size is appreciated for building ML-based classification models, the current challenge remains on limited numbers of available GPCR allosteric modulators. In this study, the SVM-ECFP6 gives a satisfying performance across 11 classes (including small classes with around 90 samples), showing the feasibility of our method to be generalized to broader GPCR families.
In compliance with the previous observation, SVM and MLP have better performance on most classes than other ML models when trained with one feature type (except for molecular descriptors feature). When trained with two or more feature types, the SVM’s performance is significantly reduced while MLP, LR, and RF show better performance in each class (Table 6, Table 7 and Table 8). Moreover, similar to the result from SVM-ECFP6, many other models did not perform well on FFA2, GLP1-R, and GCGR, but most models have a high f1-score on PTHrP. In general, ML models trained on GPCR family A and C show a better f1-score than GPCR family B (Table S4) as all datasets for class B GPCRs have small numbers of instances.

4. Conclusions

In this study, four types of features, Atom-pair, ECFP6, MACCS fingerprints, and molecular descriptors were used to construct a series of datasets. The chemical space analysis of all four features demonstrated that the 10 allosteric subtypes form spatial patterns that are distinguishable from each other. Six ML models were built and trained on datasets with different feature combinations. SVM—ECFP6 shows the best results (AUC = 0.974, ACC = 0.976, Bal_ACC = 0.950, f1-score = 0.963, CK = 0.971, MCC = 0.971, precision = 0.978, recall = 0.950). MLP has the most stable performance across different feature combinations. In particular, it outperformed other ML models on datasets constructed with two or more features. By comparing the ML model’s performance on different features, we found that when only using one feature type for training, ECFP6 is the best choice for its good performance on most ML models. Mixed effects were seen on datasets with various feature combinations on different ML models. In the field of drug discovery, we need to frequently deal with imbalanced datasets [51]. The model developed in our study shows a good generalization capability on an imbalanced dataset. To the best of our knowledge, this study is the first work on the multi-class classification of GPCR allosteric modulators. The developed multi-class classifiers provide alternative options on virtual screening besides the conventional structure-based and ligand-based methods. Besides being of benefit to potential hit identification campaigns on GPCR allosteric modulators, this study can also be of value to demonstrate the possibility of adapting machine learning to the broad area of drug discovery.

Supplementary Materials

The following are available online at https://www.mdpi.com/article/10.3390/biom11060870/s1. Collected raw chemical datasets. Results of the validation set on each dataset with different feature types (Tables S1–S3); The f1-score among different GPCR families on each dataset (Table S4); Radar chart of validation sets on different feature types (Figures S1–S4); Confusion matrix of each ML model on different feature types (Figures S5–S16). Pair-wise distribution comparisons of each GPCR class with drug-like molecules on SlogP, molecular weight (M.W.), hydrogen bond acceptor (HBA), and hydrogen bond doner (HBD) (Figures S17–S26).

Author Contributions

X.-Q.X., T.H. and Y.B. conceived the project studies. T.H. and Y.B. designed and carried out experiments. T.H. wrote the manuscript. X.-Q.X., Y.B. and T.M. made the manuscript revisions. All authors have read and agreed to the published version of the manuscript.

Funding

This work was funded by NIH NIDA grants R01DA052329-01A1 (X.-Q.X.) and P30 DA035778A1 (X.-Q.X.)

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available in Supplementary Materials.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ritter, S.L.; Hall, R.A. Fine-tuning of GPCR activity by receptor-interacting proteins. Nat. Rev. Mol. Cell Biol. 2009, 10, 819–830. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Raschka, S.; Kaufman, B. Machine learning and AI-based approaches for bioactive ligand discovery and GPCR-ligand recognition. Methods 2020, 180, 89–110. [Google Scholar] [CrossRef]
  3. Congreve, M.; de Graaf, C.; Swain, N.A.; Tate, C.G. Impact of GPCR Structures on Drug Discovery. Cell 2020, 181, 81–91. [Google Scholar] [CrossRef]
  4. Hauser, A.S.; Attwood, M.M.; Rask-Andersen, M.; Schioth, H.B.; Gloriam, D.E. Trends in GPCR drug discovery: New agents, targets and indications. Nat. Rev. Drug Discov. 2017, 16, 829–842. [Google Scholar] [CrossRef] [PubMed]
  5. Bridges, T.M.; Lindsley, C.W. G-protein-coupled receptors: From classical modes of modulation to allosteric mechanisms. ACS Chem. Biol. 2008, 3, 530–541. [Google Scholar] [CrossRef]
  6. Feng, Z.; Hu, G.; Ma, S.; Xie, X.-Q. Computational Advances for the Development of Allosteric Modulators and Bitopic Ligands in G Protein-Coupled Receptors. AAPS J. 2015, 17, 1080–1095. [Google Scholar] [CrossRef] [Green Version]
  7. Sloop, K.W.; Emmerson, P.J.; Statnick, M.A.; Willard, F.S. The current state of GPCR-based drug discovery to treat metabolic disease. Br. J. Pharmacol. 2018, 175, 4060–4071. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  8. Conn, P.J.; Lindsley, C.W.; Meiler, J.; Niswender, C.M. Opportunities and challenges in the discovery of allosteric modulators of GPCRs for treating CNS disorders. Nat. Rev. Drug Discov. 2014, 13, 692–708. [Google Scholar] [CrossRef] [Green Version]
  9. Leach, K.; Sexton, P.M.; Christopoulos, A. Allosteric GPCR modulators: Taking advantage of permissive receptor pharmacology. Trends Pharmacol. Sci. 2007, 28, 382–389. [Google Scholar] [CrossRef]
  10. Lindsley, C.W.; Emmitte, K.A.; Hopkins, C.R.; Bridges, T.M.; Gregory, K.J.; Niswender, C.M.; Conn, P.J. Practical Strategies and Concepts in GPCR Allosteric Modulator Discovery: Recent Advances with Metabotropic Glutamate Receptors. Chem. Rev. 2016, 116, 6707–6741. [Google Scholar] [CrossRef] [Green Version]
  11. Nickols, H.H.; Conn, P.J. Development of allosteric modulators of GPCRs for treatment of CNS disorders. Neurobiol. Dis. 2014, 61, 55–71. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  12. Bian, Y.; Jun, J.J.; Cuyler, J.; Xie, X.-Q. Covalent allosteric modulation: An emerging strategy for GPCRs drug discovery. Eur. J. Med. Chem. 2020, 206, 112690. [Google Scholar] [CrossRef] [PubMed]
  13. Bian, Y.; Jing, Y.; Wang, L.; Ma, S.; Jun, J.J.; Xie, X.-Q. Prediction of Orthosteric and Allosteric Regulations on Cannabinoid Receptors Using Supervised Machine Learning Classifiers. Mol. Pharm. 2019, 16, 2605–2615. [Google Scholar] [CrossRef] [PubMed]
  14. Laprairie, R.B.; Bagher, A.M.; Kelly, M.E.; Denovan-Wright, E.M. Cannabidiol is a negative allosteric modulator of the cannabinoid CB1 receptor. Br. J. Pharmacol. 2015, 172, 4790–4805. [Google Scholar] [CrossRef] [Green Version]
  15. Wu, Y.; Tong, J.; Ding, K.; Zhou, Q.; Zhao, S. GPCR Allosteric Modulator Discovery. Adv. Exp. Med. Biol. 2019, 1163, 225–251. [Google Scholar] [CrossRef]
  16. Schneider, P.; Schneider, G. De Novo Design at the Edge of Chaos. J. Med. Chem. 2016, 59, 4077–4086. [Google Scholar] [CrossRef]
  17. Zhang, R.; Xie, X. Tools for GPCR drug discovery. Acta Pharmacol. Sin. 2012, 33, 372–384. [Google Scholar] [CrossRef] [Green Version]
  18. Finak, G.; Gottardo, R. Promises and Pitfalls of High-Throughput Biological Assays. Methods Mol. Biol. 2016, 1415, 225–243. [Google Scholar] [CrossRef]
  19. Evers, A.; Klabunde, T. Structure-based drug discovery using GPCR homology modeling: Successful virtual screening for antagonists of the alpha1A adrenergic receptor. J. Med. Chem. 2005, 48, 1088–1097. [Google Scholar] [CrossRef]
  20. Liu, L.; Jockers, R. Structure-Based Virtual Screening Accelerates GPCR Drug Discovery. Trends Pharmacol. Sci. 2020, 41, 382–384. [Google Scholar] [CrossRef]
  21. Petrucci, V.; Chicca, A.; Glasmacher, S.; Paloczi, J.; Cao, Z.; Pacher, P.; Gertsch, J. Pepcan-12 (RVD-hemopressin) is a CB2 receptor positive allosteric modulator constitutively secreted by adrenals and in liver upon tissue damage. Sci. Rep. 2017, 7, 9560. [Google Scholar] [CrossRef] [PubMed]
  22. Wang, H.; Duffy, R.A.; Boykow, G.C.; Chackalamannil, S.; Madison, V.S. Identification of novel cannabinoid CB1 receptor antagonists by using virtual screening with a pharmacophore model. J. Med. Chem. 2008, 51, 2439–2446. [Google Scholar] [CrossRef] [PubMed]
  23. Bemister-Buffington, J.; Wolf, A.J.; Raschka, S.; Kuhn, L.A. Machine Learning to Identify Flexibility Signatures of Class A GPCR Inhibition. Biomolecules 2020, 10, 454. [Google Scholar] [CrossRef] [Green Version]
  24. Bian, Y.; Wang, J.; Jun, J.J.; Xie, X.-Q. Deep Convolutional Generative Adversarial Network (dcGAN) Models for Screening and Design of Small Molecules Targeting Cannabinoid Receptors. Mol. Pharm. 2019, 16, 4451–4460. [Google Scholar] [CrossRef]
  25. Liang, G.; Fan, W.; Luo, H.; Zhu, X. The emerging roles of artificial intelligence in cancer drug development and precision therapy. Biomed. Pharmacother 2020, 128, 110255. [Google Scholar] [CrossRef]
  26. Ma, C.; Wang, L.; Xie, X.-Q. Ligand Classifier of Adaptively Boosting Ensemble Decision Stumps (LiCABEDS) and its application on modeling ligand functionality for 5HT-subtype GPCR families. J. Chem. Inf. Model. 2011, 51, 521–531. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  27. Ma, C.; Wang, L.; Yang, P.; Myint, K.Z.; Xie, X.-Q. LiCABEDS II. Modeling of ligand selectivity for G-protein-coupled cannabinoid receptors. J. Chem. Inf. Model. 2013, 53, 11–26. [Google Scholar] [CrossRef] [Green Version]
  28. Reda, C.; Kaufmann, E.; Delahaye-Duriez, A. Machine learning applications in drug development. Comput. Struct. Biotechnol. J. 2020, 18, 241–252. [Google Scholar] [CrossRef] [PubMed]
  29. Tsou, L.K.; Yeh, S.H.; Ueng, S.H.; Chang, C.P.; Song, J.S.; Wu, M.H.; Chang, H.F.; Chen, S.R.; Shih, C.; Chen, C.T.; et al. Comparative study between deep learning and QSAR classifications for TNBC inhibitors and novel GPCR agonist discovery. Sci. Rep. 2020, 10, 16771. [Google Scholar] [CrossRef]
  30. Bian, Y.; Xie, X.-Q. Generative chemistry: Drug discovery with deep learning generative models. J. Mol. Model. 2021, 27, 71. [Google Scholar] [CrossRef]
  31. Kumar, R.; Sharma, A.; Siddiqui, M.H.; Tiwari, R.K. Prediction of Human Intestinal Absorption of Compounds Using Artificial Intelligence Techniques. Curr. Drug Discov. Technol. 2017, 14, 244–254. [Google Scholar] [CrossRef] [PubMed]
  32. Jacob, L.; Vert, J.P. Protein-ligand interaction prediction: An improved chemogenomics approach. Bioinformatics 2008, 24, 2149–2156. [Google Scholar] [CrossRef] [Green Version]
  33. AlQuraishi, M. AlphaFold at CASP13. Bioinformatics 2019, 35, 4862–4865. [Google Scholar] [CrossRef]
  34. Senior, A.W.; Evans, R.; Jumper, J.; Kirkpatrick, J.; Sifre, L.; Green, T.; Qin, C.; Zidek, A.; Nelson, A.W.R.; Bridgland, A.; et al. Improved protein structure prediction using potentials from deep learning. Nature 2020, 577, 706–710. [Google Scholar] [CrossRef]
  35. Shen, Q.; Wang, G.; Li, S.; Liu, X.; Lu, S.; Chen, Z.; Song, K.; Yan, J.; Geng, L.; Huang, Z.; et al. ASD v3.0: Unraveling allosteric regulation with structural mechanisms and biological networks. Nucleic Acids Res. 2016, 44, D527–D535. [Google Scholar] [CrossRef] [Green Version]
  36. Irwin, J.J.; Shoichet, B.K. ZINC--a free database of commercially available compounds for virtual screening. J. Chem. Inf. Model. 2005, 45, 177–182. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  37. Raymond, E.; Carhart, D.H.S.; Venkataraghavan, R. Atom pairs as molecular features in structure-activity studies: Definition and applications. J. Chem. Inf. Comput. Sci. 1985, 25, 64–73. [Google Scholar] [CrossRef]
  38. Durant, J.L.; Leland, B.A.; Henry, D.R.; Nourse, J.G. Reoptimization of MDL keys for use in drug discovery. J. Chem. Inf. Comput. Sci. 2002, 42, 1273–1280. [Google Scholar] [CrossRef] [Green Version]
  39. Rogers, D.; Hahn, M. Extended-connectivity fingerprints. J. Chem. Inf. Model. 2010, 50, 742–754. [Google Scholar] [CrossRef]
  40. Steinbeck, C.; Hoppe, C.; Kuhn, S.; Floris, M.; Guha, R.; Willighagen, E.L. Recent developments of the chemistry development kit (CDK)—An open-source java library for chemo- and bioinformatics. Curr. Pharm. Des. 2006, 12, 2111–2120. [Google Scholar] [CrossRef] [Green Version]
  41. Cortes, C.; Vapnik, V. Support-Vector Networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar] [CrossRef]
  42. Isa, N.A.M.; Mamat, W.M.F.W. Clustered-hybrid multilayer perceptron network for pattern recognition application. Appl. Soft Comput. 2011, 11, 1457–1466. [Google Scholar]
  43. Safavian, S.R.; Landgrebe, D. A survey of decision tree classifier methodology. IEEE Trans. Syst. Man Cybern. 1991, 21, 660–674. [Google Scholar] [CrossRef] [Green Version]
  44. Plewczynski, D.; von Grotthuss, M.; Rychlewski, L.; Ginalski, K. Virtual high throughput screening using combined random forest and flexible docking. Comb. Chem. High Throughput Screen. 2009, 12, 484–489. [Google Scholar] [CrossRef]
  45. Friedman, N.; Geiger, D.; Goldszmidt, M. Bayesian Network Classifiers. Mach. Learn. 1997, 29, 131–163. [Google Scholar] [CrossRef] [Green Version]
  46. Tolles, J.; Meurer, W.J. Logistic Regression: Relating Patient Characteristics to Outcomes. JAMA 2016, 316, 533–534. [Google Scholar] [CrossRef]
  47. Van der Maaten, L.; Hinton, G. Visualizing Data using t-SNE. J. Mach. Learn. Res. 2008, 9, 2579–2605. [Google Scholar]
  48. Zhou, B.; Jin, W. Visualization of Single Cell RNA-Seq Data Using t-SNE in R. Methods Mol. Biol 2020, 2117, 159–167. [Google Scholar] [CrossRef]
  49. Carletta, J. Assessing Agreement on Classification Tasks: The Kappa Statistic. Comput. Linguist. 1996, 22, 249–254. [Google Scholar]
  50. Matthews, B.W. Comparison of the predicted and observed secondary structure of T4 phage lysozymeBiochim. Biophys. Acta Protein Struct. 1975, 405, 442–451. [Google Scholar] [CrossRef]
  51. Korkmaz, S. Deep Learning-Based Imbalanced Data Classification for Drug Discovery. J. Chem. Inf. Model. 2020, 60, 4180–4190. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Schematic illustration of the workflow. (A) Workflow of data collection and reconstruction. (B) Workflow of model training and validation. GPCRs abbreviations: Cannabinoid receptor 1 (CB1); Free fatty acid receptor 2 (FFA2); Muscarinic acetylcholine receptor M1 (mAchR M1); Sphingosine 1-phosphate receptor 3 (S1P3); Glucagon-like peptide 1 receptor (GLP1-R); Glucagon receptor (GCGR); Parathyroid hormone-related peptide receptor (PTHrP); Metabotropic glutamate receptor 2 (mGlu2); Metabotropic glutamate receptor 4 (mGlu4); Metabotropic glutamate receptor 5 (mGlu5).
Figure 1. Schematic illustration of the workflow. (A) Workflow of data collection and reconstruction. (B) Workflow of model training and validation. GPCRs abbreviations: Cannabinoid receptor 1 (CB1); Free fatty acid receptor 2 (FFA2); Muscarinic acetylcholine receptor M1 (mAchR M1); Sphingosine 1-phosphate receptor 3 (S1P3); Glucagon-like peptide 1 receptor (GLP1-R); Glucagon receptor (GCGR); Parathyroid hormone-related peptide receptor (PTHrP); Metabotropic glutamate receptor 2 (mGlu2); Metabotropic glutamate receptor 4 (mGlu4); Metabotropic glutamate receptor 5 (mGlu5).
Biomolecules 11 00870 g001
Figure 2. t-SNE method was used to visualize the chemical distribution of allosteric modulators from four subtypes of class A GPCRs according to Atom-pair, ECFP6, MACCS fingerprints, and molecular descriptors.
Figure 2. t-SNE method was used to visualize the chemical distribution of allosteric modulators from four subtypes of class A GPCRs according to Atom-pair, ECFP6, MACCS fingerprints, and molecular descriptors.
Biomolecules 11 00870 g002
Figure 3. t-SNE method was used to visualize the chemical distribution of allosteric modulators from three subtypes of class B GPCRs according to Atom-pair, ECFP6, MACCS fingerprints, and molecular descriptors.
Figure 3. t-SNE method was used to visualize the chemical distribution of allosteric modulators from three subtypes of class B GPCRs according to Atom-pair, ECFP6, MACCS fingerprints, and molecular descriptors.
Biomolecules 11 00870 g003
Figure 4. t-SNE method was used to visualize the chemical distribution of allosteric modulators from three subtypes of class C GPCRs according to Atom-pair, ECFP6, MACCS fingerprints, and molecular descriptors.
Figure 4. t-SNE method was used to visualize the chemical distribution of allosteric modulators from three subtypes of class C GPCRs according to Atom-pair, ECFP6, MACCS fingerprints, and molecular descriptors.
Biomolecules 11 00870 g004
Figure 5. Confusion matrix (m) representing SVM classification performance on test set constructed with ECFP6 feature. The element m(i, j) is the number of times an observation of the ith true class was predicted to be of the jth class. Each colored cell of the confusion matrix chart corresponds to one element of the confusion matrix. Drug-like molecules, CB1, FFA2, mAchR M1, S1P3, GLP1-R, GCGR, PTHrP, mGlu2, mGlu4, and mGlu5 were labeled as 0 to 10, respectively.
Figure 5. Confusion matrix (m) representing SVM classification performance on test set constructed with ECFP6 feature. The element m(i, j) is the number of times an observation of the ith true class was predicted to be of the jth class. Each colored cell of the confusion matrix chart corresponds to one element of the confusion matrix. Drug-like molecules, CB1, FFA2, mAchR M1, S1P3, GLP1-R, GCGR, PTHrP, mGlu2, mGlu4, and mGlu5 were labeled as 0 to 10, respectively.
Biomolecules 11 00870 g005
Table 1. Dataset Information.
Table 1. Dataset Information.
DatasetTraining SetValidation SetTest SetTotal
Drug-like compounds5691142317798893
CB1—Cannabinoid receptor 1 (Class A)12423113831936
FFA2—Free fatty acid receptor 2 (Class A)60151489
mAchR M1—Muscarinic acetylcholine receptor M1 (Class A)4019100512256249
S1PR3—Sphingosine 1-phosphate receptor 3 (Class A)32381114518
GLP1-R—Glucagon-like peptide 1 receptor (Class B)2025161314
GCGR—Glucagon receptor (Class B)2857265422
PTHrP—Parathyroid hormone/parathyroid hormone-related peptide receptor (Class B)56151788
mGlu2—Metabotropic glutamate receptor 2 (Class C)351988010755474
mGlu4—Metabotropic glutamate receptor 4 (Class C)13423364102088
mGlu5—Metabotropic glutamate receptor 5 (Class C)5295132417448363
Table 2. Results of the Test Set on Each Dataset with One Feature Type.
Table 2. Results of the Test Set on Each Dataset with One Feature Type.
DatasetsModelAUCACCBal_ACCf1-ScoreCKMCCPrecisionRecall
Atom-pair SVM0.9660.9720.9360.9490.9650.9650.9650.936
NB0.8200.6270.6790.6250.5500.5590.6440.679
MLP0.9600.9540.9250.9330.9430.9430.9440.925
LR0.9420.9120.8950.9070.8910.8910.9220.895
RF0.9420.9560.8900.9250.9460.9460.9710.890
DT0.8610.7700.7480.7430.7150.7160.7430.748
ECFP6SVM0.9740.976 0.9500.9630.9710.9710.9780.950
NB0.9160.8680.8470.8720.8350.8390.9110.847
MLP0.9570.9580.9190.9310.9480.9480.9480.919
LR0.9550.9400.9160.9320.9250.9250.9540.916
RF0.9460.9640.8970.9310.9560.9560.9800.897
DT0.8960.8220.8120.7890.7800.7800.7710.812
MACCSSVM0.9610.9630.9260.9340.9540.9540.9440.926
NB0.8220.6290.6840.5880.5510.5550.5620.684
MLP0.9540.9400.9140.9150.9250.9250.9180.914
LR0.8980.8390.8150.8340.8000.8000.8570.815
RF0.9450.9610.8950.9240.9510.9510.9620.895
DT0.8900.8540.7960.7910.8200.8200.7900.796
Molecular DescriptorsSVM0.8090.7880.6440.6820.7340.7360.7510.644
NB0.8240.6180.6900.5760.5380.5410.5310.690
MLP0.9430.9360.8930.8930.9200.9200.9020.893
LR0.8870.8400.7930.8100.8010.8010.8340.793
RF0.9410.9530.8870.9190.9410.9410.9660.887
DT0.8800.8280.7790.7670.7870.7880.7580.779
Table 3. Results of the Test Set on Each Dataset with Two Feature Types.
Table 3. Results of the Test Set on Each Dataset with Two Feature Types.
DatasetsModelAUCACCBal_ACCf1-ScoreCKMCCPrecisionRecall
Atom-pair and Molecular Descriptors SVM0.8250.8180.6720.7150.7710.7730.7910.672
NB0.8340.6600.7040.6670.5880.5970.6880.704
MLP0.9590.9450.9250.9180.9320.9320.9140.925
LR0.9590.9470.9230.9310.9340.9340.9410.923
RF0.9440.9620.8940.9270.9520.9530.9750.894
DT0.8750.8270.7700.7650.7850.7850.7620.770
ECFP6 and Molecular DescriptorsSVM0.8190.8090.6610.7030.7590.7610.7760.661
NB0.9220.8710.8590.8750.8380.8410.9010.859
MLP0.9730.9600.9510.9450.9500.9500.9390.951
LR0.9600.9610.9250.9390.9520.9520.9570.925
RF0.9490.9680.9020.9320.9600.9600.9780.902
DT0.8760.8140.7730.7590.7700.7700.7470.773
MACCS and Molecular DescriptorsSVM0.8120.7930.6490.6880.7390.7410.7600.649
NB0.8530.6590.7430.6210.5880.5920.5790.743
MLP0.9430.9510.8910.9130.9390.9390.9500.891
LR0.9220.8970.8560.8710.8720.8720.8960.856
RF0.9450.9610.8950.9260.9510.9510.9710.895
DT0.8940.8370.8060.7850.7990.7990.7670.806
Table 4. Results of the Test Set on Dataset with Four Feature Types.
Table 4. Results of the Test Set on Dataset with Four Feature Types.
DatasetsModelAUCACCBal_ACCf1-ScoreCKMCCPrecisionRecall
Atom-pair and ECFP6 and MACCS and Molecular Descriptors SVM0.8370.8410.6930.7350.8000.8020.8070.693
NB0.8840.7760.7920.7780.7250.7290.8000.792
MLP0.9620.9670.9410.9410.9600.9600.9580.928
LR0.9680.9670.9400.9500.9590.9590.9620.940
RF0.9500.9700.9040.9330.9620.9620.9760.904
DT0.8830.8230.7870.7740.7810.7810.7650.787
Table 5. Overall Performance Comparison of Different Feature Types (Average over Nine Datasets on All ML Models).
Table 5. Overall Performance Comparison of Different Feature Types (Average over Nine Datasets on All ML Models).
DatasetsAUCACCBal_ACCf1-ScoreCKMCCPrecisionRecall
Atom-pair 0.9150.8650.8460.8470.8350.8370.8650.846
ECFP60.9410.9210.8900.9030.9030.9030.9240.890
MACCS0.9120.8640.8380.8310.8340.8340.8390.838
Molecular Descriptors0.8810.8270.7810.7750.7870.7880.7900.781
Atom-pair and Molecular Descriptors0.8990.8600.8150.8210.8270.8290.8450.815
ECFP6 and Molecular Descriptors0.9170.8970.8450.8590.8720.8720.8830.845
MACCS and Molecular Descriptors0.8950.8500.8070.8010.8150.8160.8210.807
Atom-pair and ECFP6 and MACCS and Molecular Descriptors0.9140.8910.8430.8520.8650.8660.8780.841
Table 6. The f1-score of the Test Set on Each Dataset with One Feature Type on Each GPCR Class.
Table 6. The f1-score of the Test Set on Each Dataset with One Feature Type on Each GPCR Class.
DatasetsModelDrug-LikeCB1FFA2mAchR M1S1P3GLP1-RGCGRPTHrPmGlu2mGlu4mGlu5
Atom-pair SVM0.9540.9870.8890.9830.9910.8220.8701.0000.9790.9790.981
NB0.6090.6830.3730.7150.7200.4560.6730.9190.6760.3870.661
MLP0.9280.9670.9230.9710.9740.7890.8381.0000.9560.9400.975
LR0.8690.9520.8800.9340.9780.7650.8401.0000.9110.9130.936
RF0.9250.9700.7830.9720.9500.7920.8691.0000.9660.9660.978
DT0.6980.8020.6900.8240.7780.5580.5281.0000.7430.7110.840
ECFP6SVM0.9600.9890.9230.9810.9910.9030.9021.0000.9860.9730.987
NB0.8130.8120.8000.9010.9440.7850.8711.0000.8920.8680.906
MLP0.9340.9790.7500.9680.9870.8850.8471.0000.9730.9440.971
LR0.9020.9800.8330.9550.9780.8750.8851.0000.9540.9420.952
RF0.9380.9730.7830.9760.9820.7840.8781.0000.9790.9670.982
DT0.7570.8470.6250.8480.8610.6180.6451.0000.8360.7540.885
MACCSSVM0.9430.9780.8150.9750.9730.8470.8261.0000.9780.9670.972
NB0.6120.6480.2750.6650.7190.3170.4381.0000.6950.4280.670
MLP0.9090.9650.9290.9560.9640.7140.7891.0000.9520.9300.960
LR0.7990.8820.7200.8490.9420.7290.7751.0000.8340.7540.885
RF0.9360.9760.6960.9760.9640.8410.8661.0000.9720.9590.974
DT0.7910.8670.5000.8760.7930.6430.6530.9710.8750.8240.913
Molecular DescriptorsSVM0.7590.8120.0000.8360.8650.4650.5661.0000.7870.5680.839
NB0.6770.5700.3290.6710.7520.4700.4060.8290.5760.4020.652
MLP0.9090.9380.7330.9600.9650.7520.7591.0000.9490.9080.955
LR0.8280.8800.5830.8810.9000.7660.7051.0000.8380.6690.860
RF0.9250.9690.7270.9710.9590.8110.8701.0000.9660.9430.969
DT0.7840.8460.4830.8560.7870.6260.6530.9410.8230.7530.886
Table 7. The f1-score of the Test Set on Each Dataset with Two Feature Types on Each GPCR Class.
Table 7. The f1-score of the Test Set on Each Dataset with Two Feature Types on Each GPCR Class.
DatasetsModelDrug-LikeCB1FFA2mAchR M1S1P3GLP1-RGCGRPTHrPmGlu2mGlu4mGlu5
Atom-pair and Molecular Descriptors SVM0.7820.8320.0000.8640.8700.5780.6081.0000.8320.6380.860
NB0.6480.6990.5500.7420.7550.5100.7060.9190.7250.3930.693
MLP0.9220.9440.8890.9650.9110.8280.8370.9710.9530.9120.967
LR0.9260.9640.8460.9650.9740.8700.8591.0000.9520.9280.957
RF0.9360.9720.7270.9750.9680.8110.8981.0000.9750.9520.979
DT0.7630.8430.5000.8610.7730.6270.5800.9410.8250.8170.887
ECFP6 and Molecular DescriptorsSVM0.7730.8280.0000.8580.8650.5230.5961.0000.8180.6130.856
NB0.8270.8290.8000.9040.9640.7960.8941.0000.8850.8220.905
MLP0.9410.9790.9630.9750.9610.8280.8661.0000.9600.9420.975
LR0.9420.9810.8330.9740.9650.8620.9010.9710.9680.9560.971
RF0.9440.9770.7270.9780.9870.8080.8991.0000.9790.9730.983
DT0.7530.8820.3640.8320.8500.5920.6091.0000.8180.7760.869
MACCS and Molecular DescriptorsSVM0.7610.8130.0000.8390.8650.4880.5851.0000.7950.5810.842
NB0.6730.6450.3710.7180.7860.3770.4800.9710.6940.4280.690
MLP0.9300.9580.7270.9680.9540.8000.8371.0000.9670.9330.964
LR0.8780.9230.6960.9290.9220.8150.8061.0000.8940.8040.916
RF0.9370.9710.7270.9730.9680.8330.8751.0000.9730.9540.977
DT0.7800.8480.5000.8730.8270.5910.6961.0000.8460.7860.888
Table 8. The f1-score of the Test Set on Dataset with Four Feature Types on Each GPCR Class.
Table 8. The f1-score of the Test Set on Dataset with Four Feature Types on Each GPCR Class.
DatasetsModelDrug-LikeCB1FFA2mAchR M1S1P3GLP1-RGCGRPTHrPmGlu2mGlu4mGlu5
Atom-pair and ECFP6 and MACCS and Molecular Descriptors SVM0.8040.8520.0000.8890.9000.5780.6211.0000.8600.7020.877
NB0.7390.8030.6320.8480.8910.6780.8210.9710.8370.5120.823
MLP0.9530.9810.8000.9820.9910.8650.8701.0000.9700.9640.975
LR0.9530.9790.8800.9810.9740.8810.8941.0000.9720.9610.973
RF0.9480.9790.7270.9790.9820.8330.8821.0000.9810.9730.985
DT0.7530.8690.5000.8490.8300.5960.6171.0000.8410.7750.882
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Hou, T.; Bian, Y.; McGuire, T.; Xie, X.-Q. Integrated Multi-Class Classification and Prediction of GPCR Allosteric Modulators by Machine Learning Intelligence. Biomolecules 2021, 11, 870. https://doi.org/10.3390/biom11060870

AMA Style

Hou T, Bian Y, McGuire T, Xie X-Q. Integrated Multi-Class Classification and Prediction of GPCR Allosteric Modulators by Machine Learning Intelligence. Biomolecules. 2021; 11(6):870. https://doi.org/10.3390/biom11060870

Chicago/Turabian Style

Hou, Tianling, Yuemin Bian, Terence McGuire, and Xiang-Qun Xie. 2021. "Integrated Multi-Class Classification and Prediction of GPCR Allosteric Modulators by Machine Learning Intelligence" Biomolecules 11, no. 6: 870. https://doi.org/10.3390/biom11060870

APA Style

Hou, T., Bian, Y., McGuire, T., & Xie, X. -Q. (2021). Integrated Multi-Class Classification and Prediction of GPCR Allosteric Modulators by Machine Learning Intelligence. Biomolecules, 11(6), 870. https://doi.org/10.3390/biom11060870

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop