Next Article in Journal
Exploring the Potential of the Model Cyanobacteria Synechococcus PCC 7002 and PCC 7942 for the Photoproduction of High-Value Terpenes: A Comparison with Synechocystis PCC 6803
Next Article in Special Issue
Transformer-Based Multi-Modal Data Fusion Method for COPD Classification and Physiological and Biochemical Indicators Identification
Previous Article in Journal
Osteopontin: A Bone-Derived Protein Involved in Rheumatoid Arthritis and Osteoarthritis Immunopathology
Previous Article in Special Issue
G4Beacon: An In Vivo G4 Prediction Method Using Chromatin and Sequence Information
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Molecular Property Prediction by Combining LSTM and GAT

College of Computer Science and Technology, Qingdao University, Qingdao 266071, China
*
Author to whom correspondence should be addressed.
Biomolecules 2023, 13(3), 503; https://doi.org/10.3390/biom13030503
Submission received: 6 January 2023 / Revised: 10 February 2023 / Accepted: 6 March 2023 / Published: 9 March 2023
(This article belongs to the Special Issue Artificial Intelligence (AI) in Biomedicine)

Abstract

:
Molecular property prediction is an important direction in computer-aided drug design. In this paper, to fully explore the information from SMILE stings and graph data of molecules, we combined the SALSTM and GAT methods in order to mine the feature information of molecules from sequences and graphs. The embedding atoms are obtained through SALSTM, firstly using SMILES strings, and they are combined with graph node features and fed into the GAT to extract the global molecular representation. At the same time, data augmentation is added to enlarge the training dataset and improve the performance of the model. Finally, to enhance the interpretability of the model, the attention layers of both models are fused together to highlight the key atoms. Comparison with other graph-based and sequence-based methods, for multiple datasets, shows that our method can achieve high prediction accuracy with good generalizability.

1. Introduction

Traditional drug design [1] is a time-consuming and costly process. In the early stages of drug design, drug development usually relies on the experience of medicinal chemists for the design and validation of drug molecules. The whole process is very complex and lengthy, and the results are uncontrollable and unpredictable. Computer-aided drug design (CADD) [2] is a new approach that can reduce the time, cost, and risk factors involved in the process of drug design, with help of computer technology. The prediction of molecular properties [3] is an important task in CADD, which also is one of the key tasks in cheminformatics. As the data volume of molecular property prediction becomes larger and larger, how to fully utilize these data to improve the accuracy of prediction has received extensive attention.
Density flooding theory (DFT) [4] is a common computational method for molecular property prediction, however, its computational cost is very high, and takes one hour to calculate the properties of a molecule with only 20 atoms [5]. The success of deep learning (DL) in various fields, including image classification [6], video understanding [7], medical imaging [8], and bioinformatics [9,10], shows that deep learning is a powerful tool for learning features from data and task-relevant prediction. In recent years, with the development of big data in biology [11], chemistry [12], and medicine [13], various deep learning algorithms have been optimized and applied to the field of CADD, which has significantly improved the efficiency of drug design and discovery processes.
For the molecular property prediction, how to select an appropriate form of molecule is important, and also affects the DL model selection. There are two dominant forms of molecular representation, namely sequences and molecular graphs, as shown in Figure 1. For the sequential representation, the simplified molecular input line entry system [14] (SMILES) is the mostly used method, which encodes molecules into meaningful sequences using certain simple syntactic rules. Recurrent neural networks (RNN) and their variants, such as long- and short-term memory units (LSTM), are widely used to capture features of sequence-based molecular data [15,16]. LSTM units generally consist of input gates, forget gates, output gates, and memory blocks, and they selectively pass information through the gate mechanism by merging the memory units that learn the pattern, in order to forget the previous hidden state and to update the hidden state. Thus, high-level features can be extracted from SMILES for property prediction by LSTM models. In the field of natural language processing (NLP), Word2vec [17] is used to learn high-dimensional embeddings of words. Similarly, Mol2vec [18] is also used to generate embeddings of molecules in the field of biochemistry. Lv et al. [19] mined structural information within atoms and semantic information between atoms using a multilayer LSTM with a hierarchical structure. Datta et al. [20] predicted the lipophilicity of molecules by processing sequences with Mol2vec and LSTM. The model stacked two layers of bidirectional LSTM (BiLSTM), in order to process the molecular substructure. Wu et al. [21] utilized the BiLSTM attention network with a novel multi-step attention mechanism, in order to extract key features from SMILES strings. Zheng et al. [22] proposed a deep learning method, SALSTM, with a self-attention mechanism for processing SMILES, in order to find the relationship between molecular substructures and properties derived from SMILES.
In addition, some studies have applied convolutional neural networks (CNN) to sequence data for molecular feature extraction. For example, Wang et al. [23] used 1D CNN to learn hidden data in sequences, in order to mine deeper information. Oyewola et al. [24] first preprocessed and normalized signal data from molecular properties, then built a 1D CNN to extract the characteristics of the normalized molecular property of the sequence data.
However, there are still a few problems using SMILES to represent a molecule. First, two atoms that are originally adjacent in the molecular structure may be relatively far apart in the sequence. For example, a ring may be broken into a SMILES sequence, in which the connected atoms are the first and last characters in the string. This may affect the accuracy and the performance of the prediction. Second, the SMILES strategy has the problem of non-uniqueness. A molecule may correspond to multiple different SMILES representations. For example, the SMILES of molecule 3-(3,4-dichlorophenyl)-1,1-dimethylurea is CN(C)C(=O)NC1=CC(=C(C=C1)Cl)Cl, which also could be expressed as c1(NC(=O)N(C)C)cc(c(cc1)Cl)Cl or c1(cc(Cl)c(Cl)cc1)NC(=O)N(C)C. This problem also provides a data augmentation method to enhance the performance and generalization of molecular representation. For example, Wu et al. [21] used the SMILES enumeration method in order to increase the number and diversity of training samples. Kimber et al. [25] explored five strategies to augment a single SMILES string into multiple SMILES strings: no augmentation, repeated augmentation, no repeated augmentation, reduced repeated augmentation, and estimated maximum augmentation. They found that the effect of data augmentation is affected by the model and the size of dataset.
Compared with SMILES, molecular graph representation can retain the molecular structure and topological information [26], which can provide richer information for DL models. With the widespread employing of graph neural networks (GNN), which treat atoms and bonds in molecules as nodes and edges, there are many types of GNN used to extract features of graphs, including GCN (graph convolutional networks) [27], GGNN (gated graph neural networks) [28] and DMPNN (directed message passing neural networks) [29]. For the molecular property prediction, Deng et al. [30] proposed the integrated framework XGraphBoost, which addressed the problem of a large number of samples for training in traditional GNN models. Wang et al. [31] proposed a convolutional spatial graph embedding layer (C-SGEL) and a stacked multiple C-SGEL, in order to construct convolutional spatial graph-embedding networks for learning features from molecular graphs.
In addition, the interpretability of a model is also an important aspect, indicating whether the model fits the chemical context, as well as in terms of its representation of key atoms, ascribing certain molecular properties. Weber et al. [32] proposed a simplified and interpretable graph convolutional neural network, which used saliency techniques to highlight molecular substructures associated with the corresponding property. To highlight the importance of specific atoms, Jiménez-Luna et al. [33] proposed the interpretability method, in order to iteratively mask individual atoms and compute molecular fingerprints. In addition, the GraSeq [34] combine both graph encoder and sequence encoder for molecular property.
Although the graph-based [35] and SMILES-based [36] methods for molecular property prediction have developed rapidly, there are still several ways to improve the performance of molecular property prediction.
First, both sequence and graph data provide different views of the molecule, making it important to utilize both of them together to improve the performance. However, just concatenating the information derived from sequence and graph models directly cannot effectively mine the correlation between them. In order to utilize these two forms of molecular representations more effectively, we need to explore a better way to couple these two models more closely, in order to predict molecular properties.
Second, the interpretability of machine learning is one of the most important concerns in this field. For the drug discovery and property prediction, the interpretability is more important, since we need to discover the relationship between the substructure and its property. For property predictions, different functional groups play different roles; thus, how to effectively use the parameters of sequence and graph models to explain the relationship between different functional groups is also an important issue.
Based on this, the main contribution of this paper is as follows: first, in order to fuse the sequence information and graph information of molecules, inspired by GraSeq [34], a method coupling both SALSTM and the graph attention network (GAT) is proposed in this paper. For the same molecule, the information of its SMILES sequence and molecular graph is combined to improve the comprehensiveness and generalization of the molecular representation. Next, to solve the problem of ambiguity in the SMILES sequence, the data augmentation is introduced through the SMILES enumeration method in the preprocessing stage. Finally, attention layers from the SALSTM and GAT methods are grouped together, in order to improve the performance and interpretability of the model. Through the comparative experiments, it is demonstrated that the performance of the proposed model is superior to the SALSTM or GAT model in multiple benchmark datasets.

2. Materials and Methods

2.1. Overview

In this paper, a model incorporating SALSTM and GAT is proposed, combining two strategies in order to improve the performance and interpretability of the model. The overall flow of the model is shown in Figure 2. First, for each molecule, the data augmentation is implemented to generate multiple and non-repetitive SMILES sequences. The corresponding molecular graph is also generated, and, at the same time, the adjacency matrix and the initial feature vector of each atom are obtained based on the molecular graph. Next, the augmented sequence is fed into a SALSTM for feature learning. The feature vector of each atom is updated by the SALSTM, and it is combined with the initial feature vector of atom in the molecular graph as the input of GAT, which ensures that the proposed model combines both the semantic information obtained from SALSTM and the structure information obtained from GAT. Moreover, the attention layer in SALSTM is combined with the attention layer in GAT to highlight the key atoms.

2.2. Data Pre-Processing and Augmentation

Each molecule corresponds to a unique molecular structure. Due to the ambiguity of the SMILES models, there are multiple SMILES strings that correspond to the same molecule, as shown in Figure 3, which provides an efficient way to enlarge the training dataset to meet the need for DL model training. Specifically, for molecular property prediction tasks, there are two main benefits for using data augmentation. The first one is expanding the datasets and improving the performance and accuracy of model; the second one is improving the generalization of the model to learn the features of the canonical and non-canonical SMILES of the molecule. There are already some methods [37,38,39] that introduce the data augmentation method for molecular property prediction; we used a similar method, with randomly generated SMILES strings.
First, the dataset is split into a training set, a validation set, and a test set in the ratio of 8:1:1. The RDKit chemistry toolkit [40] is used to obtain the canonical SMILES of the input molecule, and its non-canonical SMILES strings are also generated until the number of SMILES strings reach a pre-defined number, which is set as 5 or 10 in this paper according to the different datasets.
Before entering the model, the SMILES string needs to be encoded as a vector for processing. After data augmentation, the dataset is enumerated and each character is marked as a number to build a dictionary, which is same as in the BiLSTM attention network. For some atoms represented by two characters, such as Cl, Br, Si, etc., treating them as two characters will introduce noisy data into the embedding. Instead, the characters R, L, M, etc., are used to represent these atoms. Assuming that the input molecule is a sequence of length N, each atom is represented by a number X n .

2.3. SALSTM Model

After acquiring the input sequence of each molecule, we first extract the embedding of each atom from the SMILES string by using the SALSTM [22] method.
First, each element of SMILES obtains a corresponding embedding vector through a word embedding algorithm [41], which is then fed into the BiLSTM to obtain the dependencies between neighbor atoms. The BiLSTM is adopted to ensure that each atom could learn the forward and backward information.
In order to improve the interpretability and performance of the whole model, a self-attention mechanism is introduced in the SALSTM, which assigns a corresponding weight to each atom through a set of summation weight vectors for the LSTM hidden states. These multiple attention states are used in SALSTM. In order to simplify the operation, we used a single attention state in this paper. To be specific, the attention mechanism takes the whole BiLSTM hidden state H ϵ R n B × d B as an input, and outputs a weight vector α B through the following equation:
α B = s o f t m a x w 2 t a n h W 1 H T
where W 1 ϵ R d H × d B and w 2 ϵ R 1 × d H denote the weight matrix and vector, respectively, d B is the dimension of the hidden vector of each element, and n B is the number of elements. The attention weight vector α B is expanded to A B ϵ R n B × d B , and assigned to each element through the following equation:
O B = A B H
where denotes Hadamard products. Since the O B and A B are used as the input for GAT and for interpretability respectively, the features of punctuations and H atoms were removed in O B and A B , and the final output of the SALSTM is denoted as O B ϵ R n × d B and A B ϵ R n × d B , where n is the number of atoms of each molecule.

2.4. GAT Model

A large number of existing methods focus on using single-sequence-based [42] or graph-based [43,44] models independently for molecular property prediction. However, we believe that combining them together can improve the accuracy of the model. In this paper, the GAT is introduced to process the molecular graph. Moreover, the output O B obtained from the SALSTM is also used as a part of the input of GAT to better learn the local and global structural features of molecules. At the same time, the attention mechanism embedded in the GAT, combined with the attention mechanism in SALSTM, was used to compute the attention scores of the nodes of the molecular graph, in order to obtain the important sub-structure of the graph.
First, we used the RDKit toolkit to obtain the molecular graph, with its adjacency matrix derived from the corresponding SMILES sequence. For each molecular graph, atoms are regarded as nodes and bonds as edges. Atom features are also extracted by the RDKit, and detailed information is shown in Table 1.
The initial feature matrix is O G ϵ R n × d G , where n denotes the number of atoms and d G denotes the initial feature dimension of each atom. The O G is concatenated with O B to obtain the input I ϵ R n × d B + d G of GAT. Figure 4 shows the process of the GAT.
The graph attention network [45] is used to extract the final features of molecules, with the help of a self-attention mechanism for each node. Firstly, the attention coefficient e i j is calculated through the following equation:
e i j = L e a k y R e L U ( a T W I i | | W I j ] ,     j ϵ N i
where e i j denotes the attention coefficient of node i with respect to node j , N i denotes the neighborhood of node i , a and W are shared learnable parameters, and | | denotes the concatenating operation.
The e i j is fed into a s o f t m a x function for normalization.
α i j = s o f t m a x e i j = e x p e i j k ϵ N i e x p e i k
Finally, the node features are updated through aggregating node weight information through the following equation:
I i = σ j ϵ N i α i j W I j
where σ denotes the nonlinear activation function. I i is a new feature vector of node i after one attention layer.
The weight vectors of each atom i are obtained through the following equation:
α i G = j ϵ N i α i j
After GAT processing, molecular features are processed by the dense layer for property prediction. For classification tasks, a sigmoid is used to obtain the final output.

2.5. Interpretability

To improve the interpretability of the proposed model, we utilized both attention layers from SALSTM and GAT to highlight the important elements of the molecule. Since the attention scores of two models are different, instead of using a single layer from a sequence-based model or graph-based model, independently, combining them together could help in discovering the importance of atoms, not only in sequence data, but also in terms of molecular structure. In this paper, the attention score of each atom i in the SALSTM attention layer is a i B , and the attention score of each atom i in the GAT attention layer is a i G ; each score is normalized through the min-max normalization as a ˜ B and a ˜ G , and are added together to obtain the final fusing scores through the following equation:
a i = a ˜ i B + a ˜ i G

3. Results

3.1. Dataset

To fully demonstrate the advantage of the proposed model, several datasets are introduced to evaluate the proposed model in tasks of regression and classification, including ESOL, Lipophilicity, heRG, etc. For each dataset, the data are divided into training, validation, and test sets in a ratio of 8:1:1. Table 2 shows the details of the datasets. ClinTox and Tox21 are multi-classification tasks, in which the input sample corresponds to multiple labels.

3.1.1. Regression Task

For the regression task, we used the root mean square error (RMSE) as the evaluation metric; the lower the RMSE value, the better the prediction performance of the model. We selected three datasets including ESOL, FreeSolv and Lipophilicity datasets, from MoleculeNet (https://moleculenet.org/, accessed on 5 June 2022) [46], and the heRG dataset for the regression task. ESOL dataset [47] contains water solubility data for 1128 compounds. FreeSolv dataset [48] provides experimental and calculated hydration free energies for approximately 642 small molecules in water. Lipophilicity dataset [49] contains 4200 compounds’ experimental results, in terms of their octanol/water distribution coefficient. heRG dataset [50] contains the information of heRG blockers for 6993 compounds.

3.1.2. Classification Task

For the classification task, the ROC–AUC is used as the evaluation metric, and the higher the AUC value, the better the classification performance. We selected three datasets, including the BACE, ClinTox and Tox21 dataset, from MoleculeNet, as well as the Mutagenesis dataset, for the classification task. BACE dataset [51] provides a set of quantitative IC_50 for human β-secretase 1 (BACE-1) inhibitors. ClinTox dataset [52] contains clinical trial toxicity results for 1491 drug compounds. Tox21 dataset [53] contains qualitative toxicity measurements for 7831 compounds across 12 different targets. Mutagenesis dataset [54] contains mutagenesis information for 6506 compounds.

3.2. Experiment

In the comparative experiment, we compared the proposed method with various other sequence-based and graph-based approaches, in which FCNN [55], N-GRAM [56], RNNS2S [42], SMILES Transformer [57], FP2VEC [58] and TranGRU [59] are sequence-based model, and SGCN [60], MPNN [61], DMPNN [29], MGCN [62], AttentionFP [63], PreGNN [64], and GraSeq [34] are graph-based models. The results were obtained from public literature and shown in Table 3 and Table 4. The 5-fold cross-validation was used for evaluation.
Compared with other models, our model achieves good results on both regression and classification tasks. In comparison with the previous best results, the improvements are 0.5%, both for the BACE and ClinTox classification task, compared to the best results from N-GRAM and MPNN, respectively. At the same time, the RMSE value of our model decreased by 0.116 in the FreeSolv task compared with that of MPNN.

3.3. Ablation Experiment

To further research the factors affecting the performance of the proposed model, several ablation experiments with different parameters were conducted on both classification and regression datasets. For the regression task, we used ESOL, Lipophilicity and heRG datasets for evaluation, since the sizes of these datasets are distributed widely, from 1128 to 4813, which can better explore the differences between datasets. For the classification task, we used BACE, Mutagenesis and Tox21 datasets for evaluation, for the same reason, since the size range of these datasets is from 1513 to 7831.

3.3.1. Comparison with SALSTM and GAT

We combined both SALSTM and GAT together to improve the accuracy of the prediction in this paper. To demonstrate the performance of the proposed model, we compared it with SALSTM and GAT, both in regression and classification tasks, and the results are shown in Figure 5. It can be seen that GAT generally outperforms SALSTM, and after combining both, our model further improved the performance. Figure 5a shows the performance of the three models in the regression task. Compared with SALSTM, our model is improved by 3.2%, 12.5% and 13.0%, and the improvement compared to GAT is 2.1%, 21.7% and 1.6%, respectively. For the classification task, it can be seen from Figure 5b that the proposed method still showed superiority when compared to other methods.
Moreover, we compared the differences in training loss of the three methods between the ESOL, BACE and Lipop dataset, as shown in Figure 6. For the ESOL dataset, at the 30th epoch, the losses of our model and GAT are the same. Although the value of loss is similar between our model and GAT, the final loss (epoch 100) of our method is lower than that of GAT. SALSTM is similar to our model in the early stage of training, but the final training loss is higher than GAT and our model. For the BACE dataset, we observed that our model showed a good performance from the early stage of training, the convergence speed is higher than that of GAT and SALSTM, and it tends to be stable at the 60th epoch. For the Lipop dataset, it can be clearly seen from the Figure 6c that, as the number of training rounds increases, our model training loss decreases more significantly than that of SALSTM and GAT. At epoch 40, the loss of our model is lower than that of SALSTM and GAT.

3.3.2. Evaluation on Different Data Augmentation Methods

Data augmentation is important in enlarging the training dataset and improving the generalization of the model. To demonstrate the contribution of the data augmentation method, we tested the performance of the model with the augmentation method, including 5 samples and 10 samples, and without the augmentation method; the results are shown in Figure 7. As can be seen from Figure 7a, the model tested with augmentation methods is improved, in terms of the ESOL, Lipophilicity and heRG datasets, when compared to the model tested without augmentation methods. Moreover, the improvements are different, depending on the size of the datasets and the amount of data augmentation. When the number of augmented samples is five, the improvement is about 22.5%, 15.2% and 23.1% for the ESOL, Lipophilicity and heRG datasets, respectively. This shows that data augmentation is very significant for the performance improvement of the model. It can also be seen from Figure 7b that data augmentation improves the model significantly in classification tasks. On the other hand, there is no obvious difference between the methods when using 5 augmented samples and 10 augmented samples. The performance of the model using 10 augmented samples is better than that when using 5 augmented samples, in ESOL and Mutagenesis only, but is worse for the other four datasets. This shows that, in terms of the amount of data augmentation, more is not better. It is possible that the large amount of data augmentation will result in the overfitting of the model, which will affect the performance of the model.

3.3.3. Impact of Adding Attention Mechanism to the Proposed Model

We tested the model’s performance with and without the attention mechanism in the regression task and classification task, as shown in Figure 8a,b, respectively. It can be seen that, after adding the attention mechanism, it performed better than when no attention mechanism was added, for the ESOL, Lipophilicity and heRG regression tasks, as well as the BACE, Mutagenesis and Tox21 classification tasks. Through adding the attention mechanism, the molecular embedding obtained is not only semantic information, requiring longer relations, but also structural information between local adjacent atoms.

3.4. Interpretability

Besides improving the accuracy of the prediction, the attention mechanism provides an effect way to improve the interpretability of the model. By using the Equation (7) we combined the both attention values from SALSTM and GAT together to highlight the important atoms. We collected the attention weights of the nodes, in order to color each atom of the molecule. The attention weights represent the importance of atoms for the corresponding molecular property, with darker colors representing higher attention weights. Figure 9 shows eight molecules randomly selected from the ESOL and BACE datasets. For ESOL datasets, different functional groups have different importance, in terms of the water solubility of the molecule. It can be obtained from the figure that the functional groups with large attention weights are basically hydrophilic groups, such as hydroxyl(-OH), carbonyl(C=O), etc. This proves that our model can effectively learn molecular representations, and this can provide an effective way for the interpretability of the model. The BACE-1 inhibitors with their attention values are shown in the bottom part of Figure 9. According to the existing studies [65,66], 2-aminoimidazole and isocytosine have inhibitory effects against BACE-1. As shown in the Figure 9, the attention values of the isocytosine or 2-aminoimidazole in the molecule is greater, which is consistent with these studies.

4. Conclusions

In this paper, we proposed a model that combines semantic information in molecule sequences and structural information in molecular graphs for predicting molecular properties. Attention mechanisms in the sequence-based model and graph-based model were also fused, in order to learn the key atomic information of the molecule and visualize it. Our model was tested on different datasets, including using regression and classification tasks, and the results demonstrate the generalizability and robustness of the proposed method.

Author Contributions

Conceptualization, L.X. (Lei Xu) and Z.L.; methodology, L.X. (Lei Xu); validation, S.P.; dataset, L.X. (Leiming Xia); writing—original draft preparation, L.X. (Lei Xu); writing—review and editing, Z.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work has been supported by Shandong Key Science and Technology Innovation Project [2021CXGC011003] and Qingdao Key Technology Research and Industrialization Projects [22-3-2-qljh-8-gx].

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Mouchlis, V.D.; Afantitis, A.; Serra, A.; Fratello, M.; Papadiamantis, A.G.; Aidinis, V.; Lynch, I.; Greco, D.; Melagraki, G. Advances in de Novo Drug Design: From Conventional to Machine Learning Methods. Int. J. Mol. Sci. 2021, 22, 1676. [Google Scholar] [CrossRef] [PubMed]
  2. Gurung, A.B.; Ali, M.A.; Lee, J.; Farah, M.A.; Al-Anazi, K.M. An Updated Review of Computer-Aided Drug Design and Its Application to COVID-19. BioMed Res. Int. 2021, 2021, 8853056. [Google Scholar] [CrossRef]
  3. Li, Z.; Jiang, M.; Wang, S.; Zhang, S. Deep Learning Methods for Molecular Representation and Property Prediction. Drug Discov. Today 2022, 27, 103373. [Google Scholar] [CrossRef] [PubMed]
  4. Ndagi, U.; Falaki, A.A.; Abdullahi, M.; Lawal, M.M.; Soliman, M.E. Antibiotic Resistance: Bioinformatics-Based Understanding as a Functional Strategy for Drug Design. RSC Adv. 2020, 10, 18451–18468. [Google Scholar] [CrossRef]
  5. Raghavachari, K.; Stefanov, B.B.; Curtiss, L.A. Accurate Density Functional Thermochemistry for Larger Molecules. Mol. Phys. 1997, 91, 555–560. [Google Scholar] [CrossRef]
  6. Jena, B.; Saxena, S.; Nayak, G.K.; Saba, L.; Sharma, N.; Suri, J.S. Artificial Intelligence-Based Hybrid Deep Learning Models for Image Classification: The First Narrative Review. Comput. Biol. Med. 2021, 137, 104803. [Google Scholar] [CrossRef]
  7. Pradhyumna, P.; Shreya, G.P.; Mohana. Graph Neural Network (GNN) in Image and Video Understanding Using Deep Learning for Computer Vision Applications. In Proceedings of the 2nd International Conference on Electronics and Sustainable Communication Systems, ICESC 2021, Coimbatore, India, 4–6 August 2021; pp. 1183–1189. [Google Scholar]
  8. Pandey, B.; Kumar Pandey, D.; Pratap Mishra, B.; Rhmann, W. A Comprehensive Survey of Deep Learning in the Field of Medical Imaging and Medical Natural Language Processing: Challenges and Research Directions. J. King Saud Univ.-Comput. Inf. Sci. 2022, 34, 5083–5099. [Google Scholar] [CrossRef]
  9. Yi, H.C.; You, Z.H.; Huang, D.S.; Kwoh, C.K. Graph Representation Learning in Bioinformatics: Trends, Methods and Applications. Brief. Bioinform. 2022, 23, bbab340. [Google Scholar] [CrossRef]
  10. Berrar, D.; Dubitzky, W. Deep Learning in Bioinformatics and Biomedicine. Brief. Bioinform. 2021, 22, 1513–1514. [Google Scholar] [CrossRef]
  11. Xia, J.; Wang, J.; Niu, S. Research Challenges and Opportunities for Using Big Data in Global Change Biology. Glob. Chang. Biol. 2020, 26, 6040–6061. [Google Scholar] [CrossRef]
  12. Tetko, I.V.; Engkvist, O. From Big Data to Artificial Intelligence: Chemoinformatics Meets New Challenges. J. Cheminform. 2020, 12, 12–14. [Google Scholar] [CrossRef]
  13. Awrahman, B.J.; Aziz Fatah, C.; Hamaamin, M.Y. A Review of the Role and Challenges of Big Data in Healthcare Informatics and Analytics. Comput. Intell. Neurosci. 2022, 2022, 5317760. [Google Scholar] [CrossRef] [PubMed]
  14. Segota, S.B.; Andelic, N.; Lorencin, I.; Musulin, J.; Stifanic, D.; Car, Z. Preparation of Simplified Molecular Input Line Entry System Notation Datasets for Use in Convolutional Neural Networks. In Proceedings of the BIBE 2021—21st IEEE International Conference on BioInformatics and BioEngineering, Kragujevac, Serbia, 25–27 October 2021; Volume 9, pp. 12–17. [Google Scholar]
  15. Kachalkin, M.N.; Ryazanova, T.K.; Sokolova, I.V.; Voronin, A.V. Prediction of COX-2 Inhibitory Activity Using LSTM-Network. In Proceedings of the 2022 Ural-Siberian Conference on Computational Technologies in Cognitive Science, Genomics and Biomedicine, CSGB 2022, Novosibirsk, Russia, 4–8 July 2022; pp. 160–163. [Google Scholar]
  16. Segler, M.H.S.; Kogej, T.; Tyrchan, C.; Waller, M.P. Generating Focused Molecule Libraries for Drug Discovery with Recurrent Neural Networks. ACS Cent. Sci. 2018, 4, 120–131. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  17. Mikolov, T.; Chen, K.; Corrado, G.; Dean, J. Efficient Estimation of Word Representations in Vector Space. In Proceedings of the 1st International Conference on Learning Representations, ICLR 2013—Workshop Track Proceedings, Scottsdale, AZ, USA, 2–4 May 2013; pp. 1–12. [Google Scholar]
  18. Jaeger, S.; Fulle, S.; Turk, S. Mol2vec: Unsupervised Machine Learning Approach with Chemical Intuition. J. Chem. Inf. Model. 2018, 58, 27–35. [Google Scholar] [CrossRef] [PubMed]
  19. Lv, Q.; Chen, G.; Zhao, L.; Zhong, W.; Yu-Chian Chen, C. Mol2Context-Vec: Learning Molecular Representation from Context Awareness for Drug Discovery. Brief. Bioinform. 2021, 22, bbab317. [Google Scholar] [CrossRef]
  20. Datta, R.; Das, D.; Das, S. Efficient Lipophilicity Prediction of Molecules Employing Deep-Learning Models. Chemom. Intell. Lab. Syst. 2021, 213, 104309. [Google Scholar] [CrossRef]
  21. Wu, C.K.; Zhang, X.C.; Yang, Z.J.; Lu, A.P.; Hou, T.J.; Cao, D.S. Learning to SMILES: BAN-Based Strategies to Improve Latent Representation Learning from Molecules. Brief. Bioinform. 2021, 22, bbab327. [Google Scholar] [CrossRef]
  22. Zheng, S.; Yan, X.; Yang, Y.; Xu, J. Identifying Structure-Property Relationships through SMILES Syntax Analysis with Self-Attention Mechanism. J. Chem. Inf. Model. 2019, 59, 914–923. [Google Scholar] [CrossRef] [Green Version]
  23. Wang, X.; Liu, J.; Zhang, C.; Wang, S. SSGraphCPI: A Novel Model for Predicting Compound-Protein Interactions Based on Deep Learning. Int. J. Mol. Sci. 2022, 23, 3780. [Google Scholar] [CrossRef]
  24. Oyewola, D.O.; Dada, E.G.; Emebo, O.; Oluwagbemi, O.O. Using Deep 1D Convolutional Grated Recurrent Unit Neural Network to Optimize Quantum Molecular Properties and Predict Intramolecular Coupling Constants of Molecules of Potential Health Medications and Other Generic Molecules. Appl. Sci. 2022, 12, 7228. [Google Scholar] [CrossRef]
  25. Kimber, T.B.; Gagnebin, M.; Volkamer, A. Maxsmi: Maximizing Molecular Property Prediction Performance with Confidence Estimation Using SMILES Augmentation and Deep Learning. Artif. Intell. Life Sci. 2021, 1, 100014. [Google Scholar] [CrossRef]
  26. Wang, S.; Jiang, M.; Zhang, S.; Wang, X.; Yuan, Q.; Wei, Z.; Li, Z. Mcn-cpi: Multiscale Convolutional Network for Compound–Protein Interaction Prediction. Biomolecules 2021, 11, 1119. [Google Scholar] [CrossRef]
  27. Sun, M.; Zhao, S.; Gilvary, C.; Elemento, O.; Zhou, J.; Wang, F. Graph Convolutional Networks for Computational Drug Development and Discovery. Brief. Bioinform. 2020, 21, 919–935. [Google Scholar] [CrossRef]
  28. Beck, D.; Haffari, G.; Cohn, T. Graph-to-Sequence Learning Using Gated Graph Neural Networks. In Proceedings of the ACL 2018—56th Annual Meeting of the Association for Computational Linguistics, Proceedings of the Conference (Long Papers), Melbourne, Australia, 15–20 July 2018; Volume 1, pp. 273–283. [Google Scholar]
  29. Ma, H.; Bian, Y.; Rong, Y.; Huang, W.; Xu, T.; Xie, W.; Ye, G.; Huang, J. Dual Message Passing Neural Network for Molecular Property Prediction. arXiv 2020, arXiv:2005.13607. [Google Scholar]
  30. Deng, D.; Chen, X.; Zhang, R.; Lei, Z.; Wang, X.; Zhou, F. XGraphBoost: Extracting Graph Neural Network-Based Features for a Better Prediction of Molecular Properties. J. Chem. Inf. Model. 2021, 61, 2697–2705. [Google Scholar] [CrossRef] [PubMed]
  31. Wang, X.; Li, Z.; Jiang, M.; Wang, S.; Zhang, S.; Wei, Z. Molecule Property Prediction Based on Spatial Graph Embedding. J. Chem. Inf. Model. 2019, 59, 3817–3828. [Google Scholar] [CrossRef]
  32. Weber, J.K.; Morrone, J.A.; Bagchi, S.; Pabon, J.D.E.; Kang, S.G.; Zhang, L.; Cornell, W.D. Simplified, Interpretable Graph Convolutional Neural Networks for Small Molecule Activity Prediction. J. Comput.-Aided Mol. Des. 2021, 36, 391–404. [Google Scholar] [CrossRef]
  33. Jiménez-Luna, J.; Skalic, M.; Weskamp, N.; Schneider, G. Coloring Molecules with Explainable Artificial Intelligence for Preclinical Relevance Assessment. J. Chem. Inf. Model. 2021, 61, 1083–1094. [Google Scholar] [CrossRef]
  34. Guo, Z.; Yu, W.; Zhang, C.; Jiang, M.; Chawla, N.V. GraSeq: Graph and Sequence Fusion Learning for Molecular Property Prediction. In Proceedings of the International Conference on Information and Knowledge Management, Virtual, 19–23 October 2020; pp. 435–443. [Google Scholar]
  35. Jin, Y.; Lu, J.; Shi, R.; Yang, Y. EmbedDTI: Enhancing the Molecular Representations via Sequence Embedding and Graph Convolutional Network for the Prediction of Drug-Target Interaction. Biomolecules 2021, 11, 1783. [Google Scholar] [CrossRef]
  36. Shrivastava, A.D.; Swainston, N.; Samanta, S.; Roberts, I.; Muelas, M.W.; Kell, D.B. Massgenie: A Transformer-Based Deep Learning Method for Identifying Small Molecules from Their Mass Spectra. Biomolecules 2021, 11, 1793. [Google Scholar] [CrossRef]
  37. Li, C.; Feng, J.; Liu, S.; Yao, J. A Novel Molecular Representation Learning for Molecular Property Prediction with a Multiple SMILES-Based Augmentation. Comput. Intell. Neurosci. 2022, 2022, 8464452. [Google Scholar] [CrossRef]
  38. Sumner, D.; He, J.; Thakkar, A.; Engkvist, O.; Bjerrum, E.J. Levenshtein Augmentation Improves Performance of SMILES Based Deep-Learning Synthesis Prediction. ChemRxiv 2020. [Google Scholar] [CrossRef]
  39. Arús-Pous, J.; Johansson, S.V.; Prykhodko, O.; Bjerrum, E.J.; Tyrchan, C.; Reymond, J.L.; Chen, H.; Engkvist, O. Randomized SMILES Strings Improve the Quality of Molecular Generative Models. J. Cheminform. 2019, 11, 71. [Google Scholar] [CrossRef]
  40. Landrum, G. RDKit: Open-Source Cheminformatics. 2006. Available online: http://www.rdkit.org/ (accessed on 10 May 2022).
  41. Yang, K.K.; Wu, Z.; Bedbrook, C.N.; Arnold, F.H. Learned Protein Embeddings for Machine Learning. Bioinformatics 2018, 34, 2642–2648. [Google Scholar] [CrossRef] [Green Version]
  42. Xu, Z.; Wang, S.; Zhu, F.; Huang, J. Seq2seq Fingerprint: An Unsupervised Deep Molecular Embedding for Drug Discovery. In Proceedings of the 8th ACM International Conference on Bioinformatics, Computational Biology, and Health Informatics, Boston, MA, USA, 20–23 August 2017; pp. 285–294. [Google Scholar]
  43. Yang, K.; Swanson, K.; Jin, W.; Coley, C.; Eiden, P.; Gao, H.; Guzman-Perez, A.; Hopper, T.; Kelley, B.; Mathea, M.; et al. Analyzing Learned Molecular Representations for Property Prediction. J. Chem. Inf. Model. 2019, 59, 3370–3388. [Google Scholar] [CrossRef] [Green Version]
  44. Li, R.; Wang, S.; Zhu, F.; Huang, J. Adaptive Graph Convolutional Neural Networks. In Proceedings of the AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018; Volume 32, pp. 3546–3553. [Google Scholar]
  45. Veličković, P.; Casanova, A.; Liò, P.; Cucurull, G.; Romero, A.; Bengio, Y. Graph Attention Networks. arXiv 2018, arXiv:1710.10903. [Google Scholar]
  46. Wu, Z.; Ramsundar, B.; Feinberg, E.N.; Gomes, J.; Geniesse, C.; Pappu, A.S.; Leswing, K.; Pande, V. MoleculeNet: A Benchmark for Molecular Machine Learning. Chem. Sci. 2018, 9, 513–530. [Google Scholar] [CrossRef] [Green Version]
  47. Delaney, J.S. ESOL: Estimating Aqueous Solubility Directly from Molecular Structure. J. Chem. Inf. Comput. Sci. 2004, 44, 1000–1005. [Google Scholar] [CrossRef]
  48. Mobley, D.L.; Guthrie, J.P. FreeSolv: A Database of Experimental and Calculated Hydration Free Energies, with Input Files. J. Comput.-Aided Mol. Des. 2014, 28, 711–720. [Google Scholar] [CrossRef] [Green Version]
  49. Waring, M.J. Lipophilicity in Drug Discovery. Expert Opin. Drug Discov. 2010, 5, 235–248. [Google Scholar] [CrossRef]
  50. Garrido, A.; Lepailleur, A.; Mignani, S.M.; Dallemagne, P.; Rochais, C. HERG Toxicity Assessment: Useful Guidelines for Drug Design. Eur. J. Med. Chem. 2020, 195, 112290. [Google Scholar] [CrossRef]
  51. Subramanian, G.; Ramsundar, B.; Pande, V.; Denny, R.A. Computational Modeling of β-Secretase 1 (BACE-1) Inhibitors Using Ligand Based Approaches. J. Chem. Inf. Model. 2016, 56, 1936–1949. [Google Scholar] [CrossRef] [PubMed]
  52. Boria, I.; Garelli, E.; Gazda, H.T.; Aspesi, A.; Quarello, P.; Pavesi, E.; Ferrante, D.; Meerpohl, J.J.; Kartal, M.; Da Costa, L.; et al. The Ribosomal Basis of Diamond-Blackfan Anemia: Mutation and Database Update. Hum. Mutat. 2010, 31, 1269–1279. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  53. Wu, L.; Huang, R.; Tetko, I.V.; Xia, Z.; Xu, J.; Tong, W. Trade-off Predictivity and Explainability for Machine-Learning Powered Predictive Toxicology: An in-Depth Investigation with Tox21 Data Sets. Chem. Res. Toxicol. 2021, 34, 541–549. [Google Scholar] [CrossRef]
  54. Gray, V.E.; Hause, R.J.; Luebeck, J.; Shendure, J.; Fowler, D.M. Quantitative Missense Variant Effect Prediction Using Large-Scale Mutagenesis Data. Cell Syst. 2018, 6, 116–124.e3. [Google Scholar] [CrossRef] [Green Version]
  55. Zhang, Y.; Lee, J.D.; Wainwright, M.J.; Jordan, M.I. On the Learnability of Fully-Connected Neural Networks. PMLR 2017, 54, 83–91. [Google Scholar]
  56. Liu, S.; Demirel, M.F.; Liang, Y. N-Gram Graph: Simple Unsupervised Representation for Graphs, with Applications to Molecules. Adv. Neural Inf. Process. Syst. 2019, 32, 1–13. [Google Scholar]
  57. Honda, S.; Shi, S.; Ueda, H.R. SMILES Transformer: Pre-Trained Molecular Fingerprint for Low Data Drug Discovery. arXiv 2019, arXiv:1911.04738. [Google Scholar]
  58. Jeon, W.; Kim, D. FP2VEC: A New Molecular Featurizer for Learning Molecular Properties. Bioinformatics 2019, 35, 4979–4985. [Google Scholar] [CrossRef]
  59. Jiang, J.; Zhang, R.; Ma, J.; Liu, Y.; Yang, E.; Du, S.; Zhao, Z.; Yuan, Y. TranGRU: Focusing on Both the Local and Global Information of Molecules for Molecular Property Prediction. Appl. Intell. 2022, 52, 1–15. [Google Scholar] [CrossRef]
  60. Li, J.; Zhang, T.; Tian, H.; Jin, S.; Fardad, M.; Zafarani, R. SGCN: A Graph Sparsifier Based on Graph Convolutional Networks. In Proceedings of the Pacific-Asia Conference on Knowledge Discovery and Data Mining 2020, Singapore, 11–14 May 2020; Volume 12084, pp. 275–287. [Google Scholar]
  61. Gilmer, J.; Schoenholz, S.S.; Riley, P.F.; Vinyals, O.; Dahl, G.E. Neural Message Passing for Quantum Chemistry. In Proceedings of the 34th International Conference on Machine Learning, ICML 2017, Sydney, Australia, 6–11 August 2017; Volume 3, pp. 2053–2070. [Google Scholar]
  62. Lu, C.; Liu, Q.; Wang, C.; Huang, Z.; Lin, P.; He, L. Molecular Property Prediction: A Multilevel Quantum Interactions Modeling Perspective. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 17 July 2019; Volume 33, pp. 1052–1060. [Google Scholar]
  63. Xiong, Z.; Wang, D.; Liu, X.; Zhong, F.; Wan, X.; Li, X.; Li, Z.; Luo, X.; Chen, K.; Jiang, H.; et al. Pushing the Boundaries of Molecular Representation for Drug Discovery with the Graph Attention Mechanism. J. Med. Chem. 2020, 63, 8749–8760. [Google Scholar] [CrossRef] [PubMed]
  64. Hu, W.; Liu, B.; Gomes, J.; Zitnik, M.; Liang, P.; Pande, V.; Leskovec, J. Strategies for Pre-Training Graph Neural Networks. arXiv 2019, arXiv:1905.12265. [Google Scholar]
  65. Chiriano, G.; De Simone, A.; Mancini, F.; Perez, D.I.; Cavalli, A.; Bolognesi, M.L.; Legname, G.; Martinez, A.; Andrisano, V.; Carloni, P.; et al. A Small Chemical Library of 2-Aminoimidazole Derivatives as BACE-1 Inhibitors: Structure-Based Design, Synthesis, and Biological Evaluation. Eur. J. Med. Chem. 2012, 48, 206–213. [Google Scholar] [CrossRef] [PubMed]
  66. Mureddu, L.G.; Vuister, G.W. Fragment-Based Drug Discovery by NMR. Where Are the Successes and Where Can It Be Improved? Front. Mol. Biosci. 2022, 9, 834453. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Two representation methods of the same molecule, namely SMILES and molecular graph.
Figure 1. Two representation methods of the same molecule, namely SMILES and molecular graph.
Biomolecules 13 00503 g001
Figure 2. The overall framework.
Figure 2. The overall framework.
Biomolecules 13 00503 g002
Figure 3. The canonical and non-canonical SMILES representations of the molecule 3-(3,4-dichlorophenyl)-1,1-dimethylurea.
Figure 3. The canonical and non-canonical SMILES representations of the molecule 3-(3,4-dichlorophenyl)-1,1-dimethylurea.
Biomolecules 13 00503 g003
Figure 4. The process of GAT.
Figure 4. The process of GAT.
Biomolecules 13 00503 g004
Figure 5. The performance of SALSTM, GAT and our model; the number of augmented samples is 5. (a) The performance in the regression task. (b) The performance in the classification task. Error bars indicate standard deviation under 5-fold cross-validation.
Figure 5. The performance of SALSTM, GAT and our model; the number of augmented samples is 5. (a) The performance in the regression task. (b) The performance in the classification task. Error bars indicate standard deviation under 5-fold cross-validation.
Biomolecules 13 00503 g005
Figure 6. Training loss. (a) Training losses of different models for the ESOL dataset. (b) Training losses of different models for the BACE dataset. (c) Training losses of different models for the Lipop dataset.
Figure 6. Training loss. (a) Training losses of different models for the ESOL dataset. (b) Training losses of different models for the BACE dataset. (c) Training losses of different models for the Lipop dataset.
Biomolecules 13 00503 g006
Figure 7. Impact of data augmentation on performance. (a) The performance of the regression task. (b) The performance of the classification task. Error bars indicate standard deviation under 5-fold cross-validation.
Figure 7. Impact of data augmentation on performance. (a) The performance of the regression task. (b) The performance of the classification task. Error bars indicate standard deviation under 5-fold cross-validation.
Biomolecules 13 00503 g007
Figure 8. Comparing the effect of attention on the performance of the model. (a) The performance of the regression task. (b) The performance of the classification task. Error bars indicate standard deviation under 5-fold cross-validation.
Figure 8. Comparing the effect of attention on the performance of the model. (a) The performance of the regression task. (b) The performance of the classification task. Error bars indicate standard deviation under 5-fold cross-validation.
Biomolecules 13 00503 g008
Figure 9. The visualization of molecular attention weights in the ESOL and BACE datasets. The darker the color of the atom, the greater the influence of the atom on the corresponding property.
Figure 9. The visualization of molecular attention weights in the ESOL and BACE datasets. The darker the color of the atom, the greater the influence of the atom on the corresponding property.
Biomolecules 13 00503 g009
Table 1. Atomic Features.
Table 1. Atomic Features.
FeatureDescription
Atomic numberAtomic number
DegreeNumber of directly bonded neighbors (one-hot)
Formal chargeInteger electronic charge (one-hot)
Chiral tagChirality information of atoms (one-hot)
Hs numNumber of hydrogen atoms (one-hot)
HybridizationSP,SP2,SP3,SP3D,SP3D2 (one-hot)
AromaticityWhether the atom is in an aromatic hydrocarbon
MassAtomic mass
Table 2. The details of the datasets.
Table 2. The details of the datasets.
DatasetTaskTask Type#MoleculeSplitsMetric
ESOL1Regression1128RandomRMSE
FreeSolv1Regression642RandomRMSE
Lipophilicity1Regression4200RandomRMSE
heRG1Regression4813RandomRMSE
BACE1Classification1513RandomROC-AUC
Mutagenesis1Classification6506RandomROC-ACU
ClinTox2Classification1478RandomROC-AUC
Tox2112Classification7831RandomROC-AUC
Table 3. RMSE scores of regression tasks on test sets. The values in boldface represent the best prediction performances with the corresponding datasets.
Table 3. RMSE scores of regression tasks on test sets. The values in boldface represent the best prediction performances with the corresponding datasets.
FreeSolvESOLLipophilicity
Sequence-basedFCNN1.87 ± 0.071.12 ± 0.150.86 ± 0.01
N-GRAM2.512 ± 0.1901.100 ± 0.1600.876 ± 0.033
RNNS2S2.9871.3171.219
SMILES Transformers2.2461.1441.169
FP2VEC2.512 ± 0.1901.100 ± 0.1600.876 ± 0.033
Graph-basedSGCN2.158 ± 0.0491.345 ± 0.0191.074 ± 0.007
MPNN1.327 ± 0.2790.700 ± 0.0730.673 ± 0.038
DMPNN2.1770.980/
MGCN3.349 ± 0.0971.266 ± 0.1471.113 ± 0.041
AttentionFP2.030 ± 0.4200.853 ± 0.0600.650 ± 0.030
our method1.211 ± 0.1920.885 ± 0.0670.709 ± 0.023
Table 4. ROC–AUC scores of classification tasks on test sets. The values in boldface represent the best prediction performances with the corresponding datasets.
Table 4. ROC–AUC scores of classification tasks on test sets. The values in boldface represent the best prediction performances with the corresponding datasets.
BACEClinToxTox21
Sequence-basedN-GRAM0.876 ± 0.0350.855 ± 0.0370.769 ± 0.027
RNNS2S0.717\0.702
SMILES Transformers0.719\0.706
TranGRU0.790\0.813
Graph-basedSGCN\0.820 ± 0.0090.766 ± 0.002
MPNN0.793 ± 0.0310.879 ± 0.0540.809 ± 0.017
MGCN0.734 ± 0.0300.634 ± 0.0420.707 ± 0.016
AttentionFP0.863 ± 0.0150.796 ± 0.0050.807 ± 0.020
PreGNN0.845\0.781
GraSeq0.838\0.820
our method0.880 ± 0.0090.883 ± 0.0250.774 ± 0.005
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Xu, L.; Pan, S.; Xia, L.; Li, Z. Molecular Property Prediction by Combining LSTM and GAT. Biomolecules 2023, 13, 503. https://doi.org/10.3390/biom13030503

AMA Style

Xu L, Pan S, Xia L, Li Z. Molecular Property Prediction by Combining LSTM and GAT. Biomolecules. 2023; 13(3):503. https://doi.org/10.3390/biom13030503

Chicago/Turabian Style

Xu, Lei, Shourun Pan, Leiming Xia, and Zhen Li. 2023. "Molecular Property Prediction by Combining LSTM and GAT" Biomolecules 13, no. 3: 503. https://doi.org/10.3390/biom13030503

APA Style

Xu, L., Pan, S., Xia, L., & Li, Z. (2023). Molecular Property Prediction by Combining LSTM and GAT. Biomolecules, 13(3), 503. https://doi.org/10.3390/biom13030503

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop