Next Article in Journal
Concentrations of Glypican-4, Irisin and Total Antioxidant Status in Women with Metabolic Syndrome: Influence of Physical Activity
Next Article in Special Issue
DRpred: A Novel Deep Learning-Based Predictor for Multi-Label mRNA Subcellular Localization Prediction by Incorporating Bayesian Inferred Prior Label Relationships
Previous Article in Journal
DeepIMAGER: Deeply Analyzing Gene Regulatory Networks from scRNA-seq Data
Previous Article in Special Issue
A Review for Artificial Intelligence Based Protein Subcellular Localization
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

mRCat: A Novel CatBoost Predictor for the Binary Classification of mRNA Subcellular Localization by Fusing Large Language Model Representation and Sequence Features

1
School of Computer Science and Technology, Zhengzhou University of Light Industry, Zhengzhou 450002, China
2
Henan Provincial Key Laboratory of Data Intelligence for Food Safety, Zhengzhou University of Light Industry, Zhengzhou 450002, China
3
School of Electronic Information, Zhengzhou University of Light Industry, Zhengzhou 450002, China
*
Author to whom correspondence should be addressed.
Biomolecules 2024, 14(7), 767; https://doi.org/10.3390/biom14070767
Submission received: 9 May 2024 / Revised: 23 June 2024 / Accepted: 25 June 2024 / Published: 27 June 2024
(This article belongs to the Special Issue Artificial Intelligence (AI) in Biomedicine)

Abstract

:
The subcellular localization of messenger RNAs (mRNAs) is a pivotal aspect of biomolecules, tightly linked to gene regulation and protein synthesis, and offers innovative insights into disease diagnosis and drug development in the field of biomedicine. Several computational methods have been proposed to predict the subcellular localization of mRNAs within cells. However, there remains a deficiency in the accuracy of these predictions. In this study, we propose an mRCat predictor based on the gradient boosting tree algorithm specifically to predict whether mRNAs are localized in the nucleus or in the cytoplasm. This predictor firstly uses large language models to thoroughly explore hidden information within sequences and then integrates traditional sequence features to collectively characterize mRNA gene sequences. Finally, it employs CatBoost as the base classifier for predicting the subcellular localization of mRNAs. The experimental validation on an independent test set demonstrates that mRCat obtained accuracy of 0.761, F1 score of 0.710, MCC of 0.511, and AUROC of 0.751. The results indicate that our method has higher accuracy and robustness compared to other state-of-the-art methods. It is anticipated to offer deep insights for biomolecular research.

1. Introduction

Several studies have suggested that the intracellular localization of RNA plays a role in cellular function mechanisms [1,2,3]. In the regulation process of gene expression, the subcellular localization of messenger RNAs (mRNAs) is a crucial step. This process exerts precise control of the site of protein synthesis, thereby influencing the expression and regulation of relevant cellular functions [4,5,6,7]. The subcellular localization of mRNAs is primarily governed by a dynamic regulation of transport mechanisms between the cell nucleus and the cytoplasm [8]. Once mRNAs are synthesized within the cell nucleus and undergo appropriate RNA splicing, they must traverse the nuclear membrane through the nuclear pore complex to enter the cytoplasm. This nuclear translocation process is one of the key regulatory steps in gene expression, as it determines the distribution of mRNAs in the cytoplasm [9,10,11,12]. Furthermore, the subcellular localization of mRNAs plays a crucial role in pharmaceutical applications, particularly in areas such as vaccine development, gene therapy, and drug discovery [13,14,15]. mRNA vaccines represent a novel type of vaccines that utilize mRNA molecules containing encoded specific antigens to induce the immune system to produce antibodies [16]. In vaccine development, the precise control of mRNA subcellular localization is paramount to ensure the effective expression and presentation of the synthesized proteins (antigens) to the immune system. By modulating mRNA localization within the cytoplasm, vaccine researchers can optimize antigen expression and the intensity of the immune responses, thereby enhancing vaccine efficacy. mRNAs can serve as a tool for delivering therapeutic information in gene therapy [17,18,19]. By modulating mRNA localization within cells, researchers can achieve the targeted expression of therapeutic genes, thereby more precisely controlling treatment outcomes. For example, by introducing mRNAs encoding therapeutic genes into the cell nucleus or cytoplasm, gene expression and protein synthesis can be more effectively achieved in gene therapy. In the process of drug development, the subcellular localization of mRNAs plays a critical role in the study of synthesizing drugs or drug targets [20]. By adjusting the intracellular localization of mRNAs, the synthesis location of specific proteins can be influenced, thereby regulating related biological processes. Therefore, understanding the subcellular localization of mRNAs is of significant importance for studying its functionality.
Presently, the subcellular localization of RNA is primarily studied through wet-lab experiments and computational methods. RNA fluorescence in situ hybridization (FISH) enables the accurate detection of RNA subcellular localization at single-RNA resolution and in live cells [21,22,23]. However, wet-lab experiments typically require specific experimental conditions and materials and significant time investment, which makes them not only expensive but also time-consuming. In contrast, computational methods leverage computer algorithms and mathematical models to efficiently analyze large-scale datasets and recognize patterns, thereby predicting RNA subcellular localization. The advantages of this approach lie in its high throughput, automation, and relatively lower costs. Various computational methods have been applied to determine RNA subcellular localization, such as mRNALoc [24], SubLocEP [5], RNATracker [25], and RNAlight [26]. The mRNALoc model was used to investigate the localization of mRNAs within both the cytoplasm and the nucleus. It was accomplished by employing a support vector machine (SVM) [27] to analyze sequence data and predict the subcellular localization of mRNAs. The SubLocEP is a two-layer integrated prediction model leveraging comprehensive feature attributes and LightGBM [28] classifiers, demonstrating superior accuracy and generalization on independent datasets for eukaryotic mRNA subcellular localization prediction. The RNATracker is a computational method developed based on an LSTM recurrent neural network with attention mechanism and integrates mRNA sequence and secondary structure information encoded by 4 bit and 6 bit one-hot encoding or mRNA subcellular localization prediction. RNAlight is a machine learning model based on LightGBM, designed to identify nucleotide k-mers that contribute to mRNA subcellular localization, demonstrating excellent performance in determining mRNA subcellular localization. The utilization of these computational methods provides powerful tools and resources for research into mRNA subcellular localization. However, these computational methods, which primarily rely on handcrafted features represented in a traditional manner, do not fully capture the complex and ill-defined patterns and relationships within biological sequences. For instance, nucleotide k-mer features focus solely on the local characteristics of a sequence, neglecting the order of nucleotides beyond the length of the k-mer used and thus losing the global features. Therefore, this method limits the comprehensive understanding of RNA gene sequences, and the accuracy of mRNA subcellular localization prediction still needs improvement.
In this study, we propose a new machine learning-based subcellular localization predictor for mRNAs, named mRCat. It is worth noting that in this study, only two primary subcellular localizations, the nucleus and the cytoplasm, were utilized for model training and prediction. mRNAs in the cytosol, insoluble, and membrane fractions of general cytoplasmic extracts were considered to be located in the cytoplasm, and mRNAs in the nucleus, nucleolus, lamina, and nuclear pore were considered to be located in the nucleus. This predictor initially harnesses large language models to deeply explore the implicit information within a sequence. It then amalgamates traditional sequence characteristics for a comprehensive portrayal of mRNA gene sequences. Ultimately, it utilizes CatBoost [29] as the foundational classifier to predict the subcellular localization of mRNAs. We performed 100-time 5-fold cross-validation to evaluate the model performance. Additionally, to further validate the reliability of the model, we utilized an independent dataset for additional validation, assessing the model’s performance on unseen data. The results indicate that large language models can characterize mRNA gene sequences more effectively than traditional methods, and compared to other state-of-the-art approaches, our method showed higher accuracy and robustness.

2. Materials and Methods

2.1. Datasets

High-quality datasets are essential for constructing robust models. This study constructed a prediction model for mRNA subcellular localization based on the dataset established by Guo-Hua Yuan et al. [26]. CeFra-seq [30] is a technique used to study RNA–protein interactions. This technique involves crosslinking RNA with proteins in live cells, followed by separating the crosslinked complexes into different subcellular structures. Ultimately, high-throughput sequencing is employed to reveal the subcellular localization information of RNA. APEX-Seq [31] is a technique used to investigate the localization of RNA within subcellular structures. It integrates protein biology and localization analysis, accurately marking the positions of RNA within subcellular compartments using protein tags. The subcellular localization information of mRNAs is identified through datasets generated by two advanced techniques, CeFra-seq and APEX-Seq. Considering variations under different conditions (e.g., different cell lines or experimental methods), some RNAs exhibit multiple inconsistent localizations in publicly available datasets. Given this inconsistency, further data processing was conducted, including the removal of RNAs inconsistent with multiple localizations [30]. After this filtering step, we obtained a comprehensive dataset, comprising 2256 mRNA sequences located in the nucleus and 2924 mRNA sequences located in the cytoplasm. Finally, the dataset was divided into training and testing sets at a ratio of 9:1. The training set consisted of 4662 mRNA gene sequences, with 2028 mRNA sequences located in the nucleus and 2634 mRNA sequences located in the cytoplasm. The testing set comprised 518 mRNA gene sequences, with 228 mRNA sequences located in the nucleus and 290 mRNA sequences located in the cytoplasm. The testing set was used as an independent test set for independent testing, and 5-fold cross-validation was conducted on the training set.

2.2. Framework of mRCat

We constructed an mRCat model based on the CatBoost framework for mRNA subcellular localization and compared its performance with a series of other machine learning algorithms. The specific workflow is depicted in Figure 1. The machine learning algorithms we employed included K-nearest neighbor (KNN) [32], logistic regression (LR) [33], random forest (RF) [34], and support vector machine (SVM). When handling both the training and the testing sets, we utilized the same feature encoding method and represented features using both traditional handcrafted features and learned features from large language models. Finally, we integrated these two types of features into the model training and the evaluation process. We utilized library functions from the scikit-learn [35] standard machine learning library to train and test the model in Jupyter (iPython) Notebooks. The training and testing sets can be obtained from https://zenodo.org/records/12044998 (accessed on 18 June 2024)
The CatBoost algorithm, a variant of gradient boosting, is the third iteration in the GBDT family following XGBoost [36] and LightGBM. It specializes in solving classification and regression problems. CatBoost utilizes the gradient boosting tree algorithm for model training, iteratively fitting a series of weak classifiers (typically, decision trees). Each iteration trains a new classifier based on the residuals from the previous round, gradually enhancing the model performance. Additionally, we employed early stopping during training. Training halted when the model failed to show performance improvement within 50 rounds, conserving unnecessary training time and computational resources, and preventing overfitting to the training data, thereby enhancing the model generalization. Considering these advantages, along with CatBoost’s superior classification capabilities compared to other machine learning algorithms, we chose CatBoost as the base classifier for the model framework to train.

2.3. Feature Encoding

2.3.1. Nucleotide Composition

To obtain feature information from mRNA gene sequences, we started by analyzing their nucleotide composition [37]. In this study, five calculation methods were employed: Z-curve, GC content, AT/GC ratio, GC skew, and AT skew. The calculation formulas are as follows:
Z C u r v e = X   =   F A + F G F C + F T Y   =   F A + F C F G + F T Z   =   F A + F T F G + F C
G C   c o n t e n t = F G + F C F A + F T + F G + F C
A T / G C = F A + F T F G + F C
G C s k e w = F G F C F G + F C
A T s k e w = F A F T F A + F T
where F X represents the frequency of occurrence of the bases ATGC in the gene sequence.
It is noteworthy that we could directly extract the aforementioned feature information using the PyFeat (1.0) [38] tool.

2.3.2. Three-Tuple Nucleotide Electron–Ion Interaction Pseudopotential

The energy of delocalized electrons in amino acids and nucleotides is calculated as Electron–Ion Interaction Pseudopotential (EIIP) [39]. In this approach, we employed k-mers and Electron–Ion Interaction Pseudopotentials (EIIPs) to characterize mRNA gene sequences. k-mers are crucial in biological sequence analysis, and the calculation of k-mer frequencies is considered an important method for extracting local features of gene sequences. An RNA gene sequence of length L contains Lk + 1 k-mer subsequences, with 4k possible k-mer subsequences. An RNA gene sequence of length L is assumed to be as follows:
S = N 1 , N 2 , N 3 , N 4 , N 5 , N 6 , N 7 , N 8 , N 9 , · · · , N L 1 , N L
where Nj represents any one of the nucleotides ATGC.
In this approach, we utilized 3-tuple nucleotides. When k is 3, the set of nucleotide compositions for a sequence is as follows:
T = N 1 N 2 N 3 , N 2 N 3 N 4 , N 3 N 4 N 5 , N 4 N 5 N 6 , · · · , N L 2 N L 1 N L
where Nj represents any one of the nucleotides ATGC.
The set T contains a total of L − 3 + 1 subsequences. When k is 3, there are a total of 43 = 64 possible k-mer subsequences, such as the set W = {AAA, AAT, AAG, AAC, …, CCC}. Therefore, the 3-mer features of this RNA gene sequence can be represented by a 64-dimensional vector as follows:
V = X 1 , X 2 , X 3 , X 4 , · · · , X 64
where Xi represents the frequency of occurrence of each k-mer subsequence in set T, and each k-mer subsequence is from set W.
The EIIP values for each nucleotide are shown in Table 1. The sum of the EIIP values of each nucleotide in the 3-tuple was calculated as the EIIP value of the 3-tuple. Finally, the EIIP value of the 3-tuple was multiplied by its frequency to construct a 64-dimensional vector to characterize the mRNA gene sequence.

2.3.3. Large Language Model Features

Relying solely on handcrafted features may lead to the loss of positional information for nucleotides in gene sequences. To address this, we introduced the pre-trained language model DNABERT-2 [40]. DNABERT-2 is a multi-species genomic base model tokenized using Byte Pair Encoding (BPE) [41], employing multiple strategies to overcome input length restrictions, reduce time and memory consumption, and enhance model capacity. Specifically, DNABERT-2 employs Attention with Linear Biases (ALiBi) [42] as a replacement for the conventional positional embeddings [43] utilized by mainstream language models to overcome input length limitations, coupled with Flash Attention [44] to increase the computational efficiency.
DNABERT-2 represents an iterative advancement of DNABERT [43], adapting the transformer encoder architecture similar to BERT [45], and contains 12 transformer layers. Each layer is equipped with 768 hidden units and 12 attention heads. DNABERT-2 first employs BPE to tokenize mRNA gene sequences as input, and each sequence is represented as a matrix M by embedding each token into a numerical vector. Subsequently, it utilizes a multi-head self-attention mechanism to capture contextual information on M:
M u l t i H e a d M = C o n c a t h e a d 1 , h e a d 2 , h e a d 3 , , h e a d h W O
h e a d i = s o f t m a x M W i Q M W i K T d k · M W i V
where W O , W i Q , W i K , and W i V W i Q , W i K , W i V i = 0 h are learned parameters for linear projection.
Head calculates the next hidden states of M by first computing the attentions scores between every two tokens and then utilizing them as weights to sum up lines in M W i V · M u l t i H e a d ( ) concatenates the results of each independent head with different sets of W i Q , W i K , W i V . The iteration occurs L times, aligning with the number of layers within the model architecture.
To address the challenge of explicit positional information required in attention-based models, DNABERT-2 integrates Attention with Linear Biases (ALiBi) within its multi-head self-attention mechanism. Rather than adding positional embeddings to the input, ALiBi incorporates positional information into attention scores by introducing a fixed set of static, non-learned biases to each attention calculation.
The sequence tokens undergo processing through 12 transformer blocks, each of which incorporates multi-head self-attention mechanisms. Subsequently, we extracted the mean value of the first output from the last hidden state to represent the mRNA gene sequence information for mRNA subcellular localization. Each sequence is represented as a 768-dimensional feature vector.

2.4. Performance Evaluation Metrics

We adopted a systematic and comprehensive model evaluation approach to ensure the accurate and reliable assessment of the proposed model’s performance in addressing the specific problem. The model was primarily evaluated using the following five metrics:
A c c u r a c y = T P + T N T P + T N + F P + F N
R e c a l l = T P T P + F N
P r e c i s i o n = T P T P + F P
F 1 s c o r e = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l
M C C = T P × T N F P × F N T P + F P T P + F N T N + F P T N + F N
where TP refers to instances where the model correctly classifies positive samples as positive, TN refers to instances where the model correctly classifies negative samples as negative, FP refers to instances where the model incorrectly classifies negative samples as positive, and FN refers to instances where the model incorrectly classifies positive samples as negative.

3. Results and Discussion

3.1. Comparison of Different Classifiers

Cross-validation is a commonly used technique for model evaluation, based on the principle of slicing a dataset into different training and validation sets [46]. In this study, we performed 100-time 5-fold cross-validation on the training set, comparing the mean and standard deviation of the performance metrics for different base classifiers. Figure 2 shows the performance of each base classifiers in terms of accuracy, precision, recall, and F1 score.
From Figure 2, we can observe that CatBoost significantly outperformed the other base classifiers. Specifically, the mean accuracy of CatBoost was 0.691, which was higher than the mean accuracy of KNN (0.682), LR (0.672), RF (0.683), and SVM (0.671) by 1.3%, 2.8%, 1.2% and 2.9%, respectively. The mean F1 score of CatBoost was 0.617, which was higher than the mean F1 score of KNN (0.562), LR (0.588), RF (0.589), and SVM (0.535) by 9.8%, 4.9%, 4.8%, and 15.3%, respectively. The mean precision and recall also followed the same trend. This indicates that CatBoost has superior classification performance. This may be due to CatBoost’s use of a symmetric tree structure, which ensures that each feature has the opportunity to be used at every split point in the tree, thereby better capturing nonlinear relationships in the data. The standard deviations of accuracy, precision, and recall with CatBoost (0.013, 0.020, 0.022) were comparable to those with KNN (0.012, 0.023, 0.023), LR (0.013, 0.019, 0.023), RF (0.013, 0.021, 0.025), and SVM (0.012, 0.024, 0.023). However, the standard deviation of the F1 scores with CatBoost (0.017) was lower than those obtained with KNN (0.020), LR (0.018), RF (0.020), and SVM (0.021). This suggests that CatBoost exhibits robustness and stability.
Through observation of the 100-time 5-fold cross-validation results, we discovered that CatBoost demonstrates very strong capability and robustness in mRNA subcellular localization. Therefore, choosing CatBoost as the base classifier for our mRCat model architecture was an excellent decision.

3.2. Comparison of Different Encoding Schemes

Utilizing effective feature representations in sequence-based prediction tasks was demonstrated to enhance model performance. In particular, in gene sequence analysis, effective features can identify specific sequence patterns and motifs. To achieve a more effective representation of mRNA gene sequence features, we employed both handcrafted features (HFs) and learned features (LFs) by large language models for sequence characterization. Traditional handcrafted feature encodings, including Z-curve, GC content, AT/GC ratio, GC skew, AT skew, and 3-tuple ElIP, are predominantly based on the nucleotide composition and the physical properties of nucleotides within mRNA gene sequences. Learned features, which are obtained through the deep learning of mRNA gene sequences by the DNABERT-2 large language model that has been pre-trained on an extensive corpus of genetic sequences, capture the interrelationships between motifs and the underlying patterns among the nucleotides that constitute a sequence. The performance metrics of CatBoost combined with various types of features are shown in Table 2. We observed that learned features (LFs) by large language models exhibited a stronger representational capacity compared to traditional handcrafted features (HFs). CatBoost + LFs obtained accuracy of 0.724 and F1 score of 0.659, while CatBoost + HFs obtained accuracy of 0.714 and F1 score of 0.649. This is because large language models are capable of learning long-distance dependencies and complex biological patterns, thereby enhancing the accuracy of predictions. Furthermore, our observations indicated that the fusion of handcrafted features with learned features facilitated a more robust characterization of mRNA gene sequences. mRCat obtained accuracy of 0.761 and F1 score of 0.710. These results demonstrate that large language models excel at learning sequence information and underscore the superiority of our model architecture.
To further demonstrate the advantages of mRCat and large language models in capturing sequence information, we conducted a comparative analysis on the performance metrics of Matthew’s correlation coefficient (MCC) and area under the receiver operating characteristic curve (AUROC) for traditional handcrafted features (HFs), features learned (LFs) by large language models, and the combination of HFs and LFs. The experimental results in Figure 3 show that large language models provide superior representation of mRNA gene sequences compared to traditional methods. Additionally, combining HFs with LFs enabled a more comprehensive learning of sequence characteristics. Specifically, on an independent test set, the MCC and AUROC of LFs surpassed those of HFs individually. Moreover, the combination of HFs and LFs in mRCat demonstrated significantly higher MCC and AUROC compared to using HFs or LFs alone. On average, mRCat showed a 23.1% increase in MCC compared to HFs and a 17.5% improvement compared to LFs. Additionally, mRCat demonstrated a 7% increase in AUROC over HFs and a 5.63% increase over LFs. The integration of handcrafted and learned features capitalizes on the advantages of each. Handcrafted features grant a clear and interpretable insight into the data, enhancing our understanding of the underlying mechanisms. In contrast, learned features, particularly those derived from large language models, excel at discerning subtle and complex patterns that may not be immediately apparent. This synergistic approach allows for a more comprehensive and nuanced analysis, ultimately improving the predictive power and interpretability of models. The comprehensive results substantiate the superior capability of mRCat in learning sequence information.

3.3. Comparison with Other Predictors

To further evaluate the performance of mRCat in predicting mRNA subcellular localization, we compared mRCat with several existing state-of-the-art predictors by using an independent test set. From Table 3, we can observe that mRCat significantly outperformed existing predictors. Specifically, mRCat consistently surpassed other existing predictors across all performance metrics. mRCat obtained accuracy of 0.761 and F1 score of 0.710.

4. Conclusions

The subcellular spatial distribution pattern of mRNAs represents a pivotal avenue for elucidating mRNA regulation, functionality, and associated pathologies, thereby furnishing crucial insights into the intricate landscape of intracellular gene expression regulatory networks. In this study, we propose mRCat, a gradient-boosted tree-based architecture for mRNA subcellular localization. To identify an exceptional base classifier for mRCat, we conducted 100-time 5-fold cross-validation on five classifiers: k-nearest neighbor (KNN), logistic regression (LR), random forest (RF), support vector machine (SVM), and CatBoost. This rigorous evaluation was performed to gauge the classifiers’ performance on the training sets, thus informing the selection of the most suitable base classifier for mRCat. Through observation of the 100-time 5-fold cross-validation results, we discovered that CatBoost demonstrated the strongest capability in mRNA subcellular localization. Therefore, choosing CatBoost as the base classifier for our mRCat model architecture was an excellent decision. By effectively characterizing mRNA gene sequences, the performance of model architectures in predicting mRNA subcellular localization can be significantly enhanced. In addition to employing traditional methods such as Z-curve, GC content, AT/GC, GC skew, AT skew, and 3-tuple EIIP for sequence feature extraction, we also incorporated the advanced large language model DNABERT-2 to automatically learn latent information within the sequences. Using the independent test set, the results showed accuracy, precision, recall, and F1 score values as shown in Table 2 and MCC and AUROC as shown in Figure 3. We found that utilizing sequence features learned by large language models offered superior characterization of mRNA gene sequences compared to the use of features extracted through traditional methods. Ultimately, the fusion of handcrafted features extracted through traditional methods with learned features from the large language model resulted in a further enhancement of our model’s performance, underscoring the superiority of our feature extraction methodology.
In conclusion, the subcellular spatial distribution pattern of mRNAs serves as a critical foundation for understanding mRNA regulation, functionality, and associated pathologies. Our proposed model, mRCat, leverages a gradient boosted tree-based architecture to accurately predict mRNA subcellular localization. Through 100-time 5-fold cross-validation, CatBoost was selected as the optimal classifier for our model architecture. Additionally, the integration of advanced techniques like DNABERT-2 further enhanced the model’s predictive capabilities by automatically capturing latent information within mRNA sequences. Overall, our findings underscore the importance of leveraging both traditional and state-of-the-art methodologies to advance our understanding of mRNA biology and subcellular localization dynamics.

Author Contributions

Conceptualization, L.Y.; methodology, L.Y.; supervision, X.W. and R.W.; validation, X.W.; writing—original draft, L.Y.; writing—review and editing, X.W. and R.W. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by funds from the Key Research Project of Colleges and Universities of Henan Province (No. 22A520013, No. 23B520004), the Key Science and Technology Development Program of Henan Province (No. 232102210020, No. 202102210144), and the Training Program of Young Backbone Teachers in Colleges and Universities of Henan Province (No. 2019GGJS132).

Data Availability Statement

The source codes and data for mRCat are available at https://zenodo.org/records/12044998 (accessed on 18 June 2024).

Conflicts of Interest

The authors declare there are no conflicts of interest.

References

  1. Kloc, M.; Zearfoss, N.R.; Etkin, L.D. Mechanisms of subcellular mRNA localization. Cell 2002, 108, 533–544. [Google Scholar] [CrossRef] [PubMed]
  2. Holt, C.E.; Bullock, S.L. Subcellular mRNA localization in animal cells and why it matters. Science 2009, 326, 1212–1216. [Google Scholar] [CrossRef] [PubMed]
  3. Chen, L.L. Towards higher-resolution and in vivo understanding of lncRNA biogenesis and function. Nat. Methods 2022, 19, 1152–1155. [Google Scholar] [CrossRef]
  4. Meyer, C.; Garzia, A.; Tuschl, T. Simultaneous detection of the subcellular localization of RNAs and proteins in cultured cells by combined multicolor RNA-FISH and IF. Methods 2017, 118–119, 101–110. [Google Scholar] [CrossRef]
  5. Li, J.; Zhang, L.; He, S.; Guo, F.; Zou, Q. SubLocEP: A novel ensemble predictor of subcellular localization of eukaryotic mRNA based on machine learning. Brief. Bioinform. 2021, 22, bbaa401. [Google Scholar] [CrossRef]
  6. Kejiou, N.S.; Palazzo, A.F. mRNA localization as a rheostat to regulate subcellular gene expression. Wiley Interdiscip. Rev. RNA 2017, 8, e1416. [Google Scholar] [CrossRef] [PubMed]
  7. Peer, E.; Moshitch-Moshkovitz, S.; Rechavi, G.; Dominissini, D. The Epitranscriptome in Translation Regulation. Cold Spring Harb. Perspect. Biol. 2019, 11, a032623. [Google Scholar] [CrossRef] [PubMed]
  8. Medioni, C.; Mowry, K.; Besse, F. Principles and roles of mRNA localization in animal development. Development 2012, 139, 3263–3276. [Google Scholar] [CrossRef] [PubMed]
  9. Stefanini, M.; Lovino, M.; Cucchiara, R.; Ficarra, E. Predicting gene and protein expression levels from DNA and protein sequences with Perceiver. Comput. Methods Programs Biomed. 2023, 234, 107504. [Google Scholar] [CrossRef] [PubMed]
  10. Martin, K.C.; Ephrussi, A. mRNA localization: Gene expression in the spatial dimension. Cell 2009, 136, 719–730. [Google Scholar] [CrossRef]
  11. Di Liegro, C.M.; Schiera, G.; Di Liegro, I. Regulation of mRNA transport, localization and translation in the nervous system of mammals (Review). Int. J. Mol. Med. 2014, 33, 747–762. [Google Scholar] [CrossRef] [PubMed]
  12. Bergalet, J.; Lécuyer, E. The functions and regulatory principles of mRNA intracellular trafficking. Adv. Exp. Med. Biol. 2014, 825, 57–96. [Google Scholar] [PubMed]
  13. Cooper, T.A.; Wan, L.; Dreyfuss, G. RNA and disease. Cell 2009, 136, 777–793. [Google Scholar] [CrossRef] [PubMed]
  14. Liu, H.; Zhang, W.; Zou, B.; Wang, J.; Deng, Y.; Deng, L. DrugCombDB: A comprehensive database of drug combinations toward the discovery of combinatorial therapy. Nucleic Acids Res. 2020, 48, D871–D881. [Google Scholar] [PubMed]
  15. Fagerberg, L.; Hallström, B.M.; Oksvold, P.; Kampf, C.; Djureinovic, D.; Odeberg, J.; Habuka, M.; Tahmasebpoor, S.; Danielsson, A.; Edlund, K.; et al. Analysis of the human tissue-specific expression by genome-wide integration of transcriptomics and antibody-based proteomics. Mol. Cell. Proteom. 2014, 13, 397–406. [Google Scholar] [CrossRef] [PubMed]
  16. Zhivaki, D.; Gosselin, E.A.; Sengupta, D.; Concepcion, H.; Arinze, C.; Chow, J.; Nikiforov, A.; Komoroski, V.; MacFarlane, C.; Sullivan, C.; et al. mRNAs encoding self-DNA reactive cGAS enhance the immunogenicity of lipid nanoparticle vaccines. mBio 2023, 14, e0250623. [Google Scholar] [CrossRef] [PubMed]
  17. Riedmayr, L.M.; Hinrichsmeyer, K.S.; Thalhammer, S.B.; Mittas, D.M.; Karguth, N.; Otify, D.Y.; Böhm, S.; Weber, V.J.; Bartoschek, M.D.; Splith, V.; et al. mRNA trans-splicing dual AAV vectors for (epi)genome editing and gene therapy. Nat. Commun. 2023, 14, 6578. [Google Scholar] [CrossRef] [PubMed]
  18. Gai, C.; Pomatto, M.A.C.; Deregibus, M.C.; Dieci, M.; Piga, A.; Camussi, G. Edible Plant-Derived Extracellular Vesicles for Oral mRNA Vaccine Delivery. Vaccines 2024, 12, 200. [Google Scholar] [CrossRef] [PubMed]
  19. Lei, J.; Qi, S.; Yu, X.; Gao, X.; Yang, K.; Zhang, X.; Cheng, M.; Bai, B.; Feng, Y.; Lu, M.; et al. Development of Mannosylated Lipid Nanoparticles for mRNA Cancer Vaccine with High Antigen Presentation Efficiency and Immunomodulatory Capability. Angew. Chem. Int. Ed. 2024, 63, e202318515. [Google Scholar] [CrossRef] [PubMed]
  20. Hori, H.; Yoshida, F.; Ishida, I.; Matsuo, J.; Ogawa, S.; Hattori, K.; Kim, Y.; Kunugi, H. Blood mRNA expression levels of glucocorticoid receptors and FKBP5 are associated with depressive disorder and altered HPA axis. J. Affect. Disord. 2024, 349, 244–253. [Google Scholar] [CrossRef] [PubMed]
  21. Cabili, M.N.; Dunagin, M.C.; McClanahan, P.D.; Biaesch, A.; Padovan-Merhar, O.; Regev, A.; Rinn, J.L.; Raj, A. Localization and abundance analysis of human lncRNAs at single-cell and single-molecule resolution. Genome Biol. 2015, 16, 20. [Google Scholar] [CrossRef] [PubMed]
  22. Kochan, J.; Wawro, M. Immunofluorescence Combined with Single-Molecule RNA Fluorescence In Situ Hybridization for Concurrent Detection of Proteins and Transcripts in Stress Granules. Methods Mol. Biol. 2024, 2752, 127–141. [Google Scholar] [PubMed]
  23. Dresselhaus, T.; Bleckmann, A. Tagging and Application of RNA Probes for Sequence-Specific Visualization of RNAs by Fluorescent In Situ Hybridization. Methods Mol. Biol. 2020, 2166, 3–21. [Google Scholar] [PubMed]
  24. Garg, A.; Singhal, N.; Kumar, R.; Kumar, M. mRNALoc: A novel machine-learning based in-silico tool to predict mRNA subcellular localization. Nucleic Acids Res. 2020, 48, W239–W243. [Google Scholar] [CrossRef] [PubMed]
  25. Yan, Z.; Lécuyer, E.; Blanchette, M. Prediction of mRNA subcellular localization using deep recurrent neural networks. Bioinformatics 2019, 35, i333–i342. [Google Scholar] [CrossRef] [PubMed]
  26. Yuan, G.H.; Wang, Y.; Wang, G.Z.; Yang, L. RNAlight: A machine learning model to identify nucleotide features determining RNA subcellular localization. Brief. Bioinform. 2023, 24, bbac509. [Google Scholar] [CrossRef]
  27. Cortes, C.; Vapnik, V. Support-vector networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar] [CrossRef]
  28. Ke, G.L.; Meng, Q.; Finley, T.; Wang, T.; Chen, W.; Ma, W.; Ye, Q.; Liu, T. LightGBM: A highly efficient gradient boosting decision tree. In Proceedings of the 31st Conference on Neural Information Processing System, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  29. Prokhorenkova, L.; Gusev, G.; Vorobev, A.; Dorogush, A.V.; Gulin, A. CatBoost: Unbiased boosting with categorical features. In Proceedings of the 32nd International Conference on Neural Information Processing Systems, Montreal, QC, Canada, 2–8 December 2018. [Google Scholar]
  30. Benoit Bouvrette, L.P.; Cody, N.A.L.; Bergalet, J.; Lefebvre, F.A.; Diot, C.; Wang, X.; Blanchette, M.; Lécuyer, E. CeFra-seq reveals broad asymmetric mRNA and noncoding RNA distribution profiles in Drosophila and human cells. RNA 2018, 24, 98–113. [Google Scholar] [CrossRef]
  31. Fazal, F.M.; Han, S.; Parker, K.R.; Kaewsapsak, P.; Xu, J.; Boettiger, A.N.; Chang, H.Y.; Ting, A.Y. Atlas of subcellular RNA localization revealed by APEX-Seq. Cell 2019, 178, 473–490.e426. [Google Scholar] [CrossRef]
  32. Cover, T.; Hart, P. Nearest neighbor pattern classification. IEEE Trans. Inf. Theory 1967, 13, 21–27. [Google Scholar] [CrossRef]
  33. LaValley, M.P. Logistic regression. Circulation 2008, 117, 2395–2399. [Google Scholar] [CrossRef] [PubMed]
  34. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  35. Pedregosa, F.; Varoquaux, G.; Gramfort, A. Scikit-learn: Machine learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
  36. Chen, T.; Guestrin, C. XGBoost: A Scalable Tree Boosting System. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016. [Google Scholar]
  37. Chen, W.; Lin, H.; Chou, K.C. Pseudo nucleotide composition or PseKNC: An effective formulation for analyzing genomic sequences. Mol. Biosyst. 2015, 11, 2620–2634. [Google Scholar] [CrossRef] [PubMed]
  38. Muhammod, R.; Ahmed, S.; Md Farid, D.; Shatabda, S.; Sharma, A.; Dehzangi, A. PyFeat: A Python-based effective feature generation tool for DNA, RNA and protein sequences. Bioinformatics 2019, 35, 3831–3833. [Google Scholar] [CrossRef] [PubMed]
  39. Nair, A.S.; Sreenadhan, S.P. A coding measure scheme employing electron-ion interaction pseudopotential (EIIP). Bioinformation 2006, 1, 197–202. [Google Scholar] [PubMed]
  40. Zhou, Z.; Ji, Y.; Li, W.; Dutta, P.; Davuluri, R.; Liu, H. Dnabert-2: Efficient foundation model and benchmark for multi-species genome. arXiv 2023, arXiv:2306.15006. [Google Scholar]
  41. Sennrich, R.; Haddow, B.; Birch, A. Neural machine translation of rare words with subword units. arXiv 2015, arXiv:1508.07909. [Google Scholar]
  42. Press, O.; Smith, N.A.; Lewis, M. Train Short, Test Long: Attention with Linear Biases Enables Input Length Extrapolation. arXiv 2021, arXiv:2108.12409. [Google Scholar]
  43. Ji, Y.; Zhou, Z.; Liu, H.; Davuluri, R.V. DNABERT: Pre-trained Bidirectional Encoder Representations from Transformers model for DNA-language in genome. Bioinformatics 2021, 37, 2112–2120. [Google Scholar] [CrossRef] [PubMed]
  44. Dao, T.; Fu, D.Y.; Ermon, S.; Rudra, A.; Ré, C. FlashAttention: Fast and Memory-Efficient Exact Attention with IO-Awareness. In Advances in Neural Information Processing Systems; Neural Information Processing Systems Foundation, Inc.: La Jolla, CA, USA, 2022; Volume 35, pp. 16344–16359. [Google Scholar]
  45. Devlin, J.; Chang, M.W.; Lee, K.; Toutanova, K. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. Bioinformatics 2021, 37, 2112–2120. [Google Scholar]
  46. Arlot, S.; Celisse, A. A survey of cross-validation procedures for model selection. Stat. Surv. 2010, 4, 40–79. [Google Scholar] [CrossRef]
Figure 1. Framework of mRCat. mRCat consists of the three modules of feature encoding, classifier selection, and subcellular localization prediction. The feature encoding module combines two different feature encoding methods: traditional manual encoding and large language model encoding. Traditional manual encoding includes Z-curve, GC content, AT/GC ratio, GC skew, AT skew, and 3-tuple ElIP. Encoding in the large language model utilizes DNABERT-2. Ultimately, the sequences are represented as feature vectors. The classifier selection module involves choosing from five different classifiers: KNN, LR, CatBoost, RF, and SVM. CatBoost was ultimately chosen as the base classifier for mRCat. The subcellular localization prediction module utilizes the base classifier for subcellular localization prediction.
Figure 1. Framework of mRCat. mRCat consists of the three modules of feature encoding, classifier selection, and subcellular localization prediction. The feature encoding module combines two different feature encoding methods: traditional manual encoding and large language model encoding. Traditional manual encoding includes Z-curve, GC content, AT/GC ratio, GC skew, AT skew, and 3-tuple ElIP. Encoding in the large language model utilizes DNABERT-2. Ultimately, the sequences are represented as feature vectors. The classifier selection module involves choosing from five different classifiers: KNN, LR, CatBoost, RF, and SVM. CatBoost was ultimately chosen as the base classifier for mRCat. The subcellular localization prediction module utilizes the base classifier for subcellular localization prediction.
Biomolecules 14 00767 g001
Figure 2. Violin plots, showing the performance of each base classifiers in terms of (A) accuracy, (B) precision, (C) recall, and (D) F1 score, were verified by 100-time 5-fold cross-validation on the training set.
Figure 2. Violin plots, showing the performance of each base classifiers in terms of (A) accuracy, (B) precision, (C) recall, and (D) F1 score, were verified by 100-time 5-fold cross-validation on the training set.
Biomolecules 14 00767 g002
Figure 3. Comparison of Matthews’ correlation coefficient (MCC) and area under the receiver operating characteristic curve (AUROC) performance of different features on the independent test set.
Figure 3. Comparison of Matthews’ correlation coefficient (MCC) and area under the receiver operating characteristic curve (AUROC) performance of different features on the independent test set.
Biomolecules 14 00767 g003
Table 1. Electron–ion interaction pseudopotentials of nucleotides.
Table 1. Electron–ion interaction pseudopotentials of nucleotides.
NucleotideEIIP
A0.1260
C0.1340
G0.0806
T0.1335
Table 2. Effects of different features on the predictive performance of mRNA subcellular localization on the independent test set.
Table 2. Effects of different features on the predictive performance of mRNA subcellular localization on the independent test set.
Evaluation Metrics
ModelAccuracyPrecisionRecallF1 Score
CatBoost + HFs0.7140.7060.6010.649
CatBoost + LFs0.7240.7230.6050.659
CatBoost + HFs + LFs (mRCat)0.7610.7600.6670.710
Table 3. Comparison with state-of-the-art models on the independent test set.
Table 3. Comparison with state-of-the-art models on the independent test set.
Evaluation Metrics
ModelAccuracyPrecisionRecallF1 Score
mRCat (CatBoost)0.7610.7600.6670.710
RNAlight0.730.750.590.66
SubLocEP0.6550.660.650.655
RNATracker0.5160.5950.5190.554
mRNALoc0.5910.5450.490.516
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, X.; Yang, L.; Wang, R. mRCat: A Novel CatBoost Predictor for the Binary Classification of mRNA Subcellular Localization by Fusing Large Language Model Representation and Sequence Features. Biomolecules 2024, 14, 767. https://doi.org/10.3390/biom14070767

AMA Style

Wang X, Yang L, Wang R. mRCat: A Novel CatBoost Predictor for the Binary Classification of mRNA Subcellular Localization by Fusing Large Language Model Representation and Sequence Features. Biomolecules. 2024; 14(7):767. https://doi.org/10.3390/biom14070767

Chicago/Turabian Style

Wang, Xiao, Lixiang Yang, and Rong Wang. 2024. "mRCat: A Novel CatBoost Predictor for the Binary Classification of mRNA Subcellular Localization by Fusing Large Language Model Representation and Sequence Features" Biomolecules 14, no. 7: 767. https://doi.org/10.3390/biom14070767

APA Style

Wang, X., Yang, L., & Wang, R. (2024). mRCat: A Novel CatBoost Predictor for the Binary Classification of mRNA Subcellular Localization by Fusing Large Language Model Representation and Sequence Features. Biomolecules, 14(7), 767. https://doi.org/10.3390/biom14070767

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop