Next Article in Journal
Multi-Criteria Decision Support System for Automatically Selecting Photovoltaic Sets to Maximise Micro Solar Generation
Previous Article in Journal
Verification of Control System Runtime Using an Executable Semantic Model
Previous Article in Special Issue
Hybrid Machine Learning Algorithms to Evaluate Prostate Cancer
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Supervised Dynamic Correlated Topic Model for Classifying Categorical Time Series

by
Namitha Pais
1,†,
Nalini Ravishanker
1,† and
Sanguthevar Rajasekaran
2,*,†
1
Department of Statistics, University of Connecticut, Storrs, CT 06269, USA
2
Department of Computer Science and Engineering, University of Connecticut, Storrs, CT 06269, USA
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Algorithms 2024, 17(7), 275; https://doi.org/10.3390/a17070275
Submission received: 30 April 2024 / Revised: 11 June 2024 / Accepted: 20 June 2024 / Published: 22 June 2024
(This article belongs to the Special Issue Hybrid Intelligent Algorithms)

Abstract

:
In this paper, we describe the supervised dynamic correlated topic model (sDCTM) for classifying categorical time series. This model extends the correlated topic model used for analyzing textual documents to a supervised framework that features dynamic modeling of latent topics. sDCTM treats each time series as a document and each categorical value in the time series as a word in the document. We assume that the observed time series is generated by an underlying latent stochastic process. We develop a state-space framework to model the dynamic evolution of the latent process, i.e., the hidden thematic structure of the time series. Our model provides a Bayesian supervised learning (classification) framework using a variational Kalman filter EM algorithm. The E-step and M-step, respectively, approximate the posterior distribution of the latent variables and estimate the model parameters. The fitted model is then used for the classification of new time series and for information retrieval that is useful for practitioners. We assess our method using simulated data. As an illustration to real data, we apply our method to promoter sequence identification data to classify E. coli DNA sub-sequences by uncovering hidden patterns or motifs that can serve as markers for promoter presence.

1. Introduction

Data on multiple time series or sequences are ubiquitous in various domains, and their classification finds applications in numerous areas, such as human motion classification [1], earthquake prediction [2], and heart attack detection [3]. The classification of multiple real-valued time series have been well studied in the statistical literature (see the detailed review in [4]). However, for analyzing categorical time series, most statistical methods are primarily focused on examining a single time series. A few methods include the Markov chain model [5], the link function approach [6], and the likelihood-based method [7]. In the computer science literature, a number of sequence classification methods have been developed that are black-box in nature and may be difficult to interpret. These include the minimum edit distance classifier with sequence alignment [8,9] and Markov chain-based classifiers [10]. Overall, the classification of multiple categorical time series has not received much attention. Recent work has discussed a novel approach to the classification of categorical time series in the supervised learning framework, using a spectral envelope and optimal scalings [11,12]. We present an alternative approach for classifying multiple categorical time series in the topic modeling framework.
Topic modeling algorithms generally examine a set of documents referred to as a corpus in the natural language processing (NLP) domain. Often, the sets of words observed in documents appear to represent a coherent theme or topic. Topic models analyze the words in the documents in order to uncover themes that run through them. Many topic modeling algorithms have been developed over time, including non-negative matrix factorization [13], latent Dirichlet allocation (LDA) [14], and structural topic models [15]. Topic modeling in the dynamic framework has been extensively studied to model the dynamic evolution of document collections over time [16,17]. However, the dynamic evolution of words over time within each document has not been addressed. The topic modeling literature typically assumes that words are interchangeable. We relax this assumption to model word time series (documents). We build a family of probabilistic dynamic topic models by extending the correlated topic model (CTM) in a supervised framework and modeling the dynamic evolution of the underlying latent stochastic process that reflects the thematic structure of the time series collection, and we classify the documents. Topic models like latent Dirichlet allocation, correlated topic models, and dynamic topic models are unsupervised, as only documents are used to identify topics by maximizing the likelihood (or the posterior probability) of the collection. In such modeling frameworks, we hope that the topics will be useful for categorization and are useful when no response is available and we want to infer useful information from the observed documents. However, when the main goal is prediction, a supervised topic modeling framework is beneficial, as jointly modeling the documents and the responses can identify latent topics that can predict the response variables for future unlabeled documents. The sDCTM framework is attractive because it provides a supervised dynamic topic modeling framework to (i) estimate the evolution of the latent process that captures the dynamic evolution of words in the document; and (ii) classify the time series.
We apply the sDCTM method for promoter sequence identification in E. coli DNA sub-sequences. A promoter is a region of DNA where RNA polymerase begins to transcribe a gene, and it is usually located upstream or at the 5′ end of the transcription initiation site in DNA. DNA promoters are proven to be the primary cause of numerous human diseases, including diabetes [18] and Huntington’s disease [19]. Thus, the identification of DNA sequences containing promoters has gained significant attention from researchers in the field of bioinformatics. Several computational methods and tools have been developed to analyze DNA sequences and predict potential promoter regions. These include the classification of promoter DNA sequences using a robust deep learning model, DeePromoter [20], deep learning, and the combination of continuous FastText N-grams [21] and the position-correlation scoring matrix (PCSM) algorithm [22]. We use the sDCTM model to analyze the E. coli DNA sub-sequences by treating each DNA sub-sequence as a document and the presence/absence of promoter regions as the associated class label.
The format of this paper is as follows: In Section 2, we describe the framework of the supervised dynamic correlated topic model, along with details on inference and parameter estimation. Section 3 discusses the simulation study conducted to assess our method. In Section 4, we present the results of applying the sDCTM method for promoter sequence identification in E. coli DNA subsequences and compare our method with the various classification techniques.

2. Supervised Dynamic Correlated Topic Model

Topic models are traditionally developed by treating words as interchangeable to identify semantic themes within each document. Our method aims to develop a family of probabilistic time series models to analyze the evolution of words over time within document collections. We assume that each document, represented by a categorical time series of words, arises from a generative process that includes latent variables. sDCTM provides a supervised framework for time series classification, allowing it to model class labels associated with each word time series. Alternatively, by removing the response component of the generative process, we can derive an unsupervised dynamic correlated topic model (DCTM).
Suppose we have a corpus C consisting of M documents. We represent the dth document as D d = ( w d , 1 , w d , 2 , , w d , T ) , where w d , t corresponds to a word observed at time point t on the dth document for times t = 1 , 2 , , T . Each word is represented by a V-dimensional unit (basis) vector, i.e.,  w d , t = ( w d , t 1 , , w d , t V ) such that,
w d u = 1 if u = v , 0 u v ,
when w d , t corresponds to the vth word from the vocabulary for t = 1 , , T and V denote the number of levels associated with the categorical word time series (document). Additionally, let y d be a multinomial response associated with dth document for d = 1 , 2 , , M . Each y d is represented by a C × 1 unit (basis) vector, i.e., y d = ( y d , 1 , , y d , C ) such that,
y d g = 1 if g = c , 0 g c ,
where y d corresponds to the cth class for c { 1 , 2 , C } , where C denotes the number of levels associated with the response variable. Suppose J is the number of assumed latent topics for capturing the hidden thematic structure in the corpus; then, δ d , t is a J dimensional vector corresponding to the latent topic structure for the dth document at time t.
In the promoter identification data, we treat each DNA sub-sequence to be a word time series (document) with V = 4 levels (indicating nucleotides a , g , c , and t). The response variable y d for d = 1 , 2 , , M indicates the presence/absence of a promoter region with C = 2 levels. The number of assumed topics J in the sDCTM model capture the hidden patterns within the observed DNA sub-sequences. Under the sDCTM framework, the dth document (DNA sub-sequence) D d and its associated class label (presence/absence of a promoter) y d for d = 1 , 2 , , M arises from the following generative process. For d = 1 , 2 , , M ,
  • For t = 1 , 2 , , T
    • Choose δ d , t δ d , t 1 N ( Φ δ d , t 1 , Σ ) .
    • Choose a topic Z { d , t } δ d , t M u l t ( 1 , f ( δ d , t ) ) , where
    • f ( δ d , t , j ) = exp ( δ d , t , j ) l = 1 J exp ( δ d , t , l ) for j = 1 , 2 , , J .
    • Choose a word w d , t { Z { d , t } , β } M u l t ( 1 , β Z { d , t } ) .
  • Draw class label y d Z d , 1 : T , η s o f t m a x ( Z ¯ d , η ) , where
    Z ¯ d = 1 T t = 1 T Z { d , t } ,
    is the topic frequencies for the dth document. The per-word topic indicator for each word w d , t is represented by Z d , t for t = 1 , , T and d = 1 , , M . The model parameters include the latent VAR(1) parameter, the J × J matrices Φ and Σ , the J × V matrix of word probabilities β , and the J × C matrix of regression coefficients η . In Section 2.1, we discuss approximate inference techniques based on variational methods and construct the lower bound (Constructing the Lower Bound) to estimate the variational parameters, which can be used to approximate the posterior. In Section 2.2, we discuss the variational EM algorithm to estimate the model parameters. We also discuss simulated annealing (Section 2.2.1), a probabilistic technique employed in optimizing problems to estimate certain model parameters, along with details on selecting initial values (Section 2.2.2) for using this optimization technique.

2.1. Approximate Inference

Given the dth document D d and the associated response y d for d = 1 , 2 , , M , the posterior distribution of the latent variables ( δ d , 1 : T , Z d , 1 : T ) is    
p ( δ d , 1 : T , Z d , 1 : T | D d , y d , Φ , Σ , β , η ) = p ( δ d , Z d , D d , y d Φ , Σ , β , η ) p ( D d , y d Φ , Σ , β , η ) ,
for d = 1 , 2 , , M . The numerator term in Equation (3), p ( δ d , Z d , D d , y d Φ , Σ , β , η ) , corresponds to the joint distribution of the latent variables ( δ d , 1 : T , Z d , 1 : T ) and the observed document–response pair ( D d , y d ) and is given by
p ( δ d , Z d , D d , y d Φ , Σ , β , η ) = t = 1 T ( p ( w d , t | Z d , t , β ) × p ( Z d , t | δ d , t ) × p ( δ d , t δ d , t 1 , Φ , Σ ) ) × p ( y d Z d , 1 : T , η ) ,
for d = 1 , 2 , , M . This joint distribution of the latent variables and the document–response pair can be factorized due to the sDCTM framework. The denominator term in Equation (3), i.e.,  p ( D d , y d Φ , Σ , β , η ) , represents the joint distribution of the dth document–response pair ( D d , y d ) and is given by
p ( D d , y d Φ , Σ , β , η ) = ( Z d , 1 : T ( t = 1 T p ( w d , t | Z d , t , β ) × p ( Z d , t | δ d , t ) p ( δ d , t δ d , t 1 , Φ , Σ ) ) × p ( y d Z d , 1 : T , η ) ) d δ d , 1 : T ,
for d = 1 , 2 , , M .
Computing the posterior distribution in Equation (3) is not analytically tractable. Hence, we use variational methods that consider a simple family of distributions over the latent variables, indexed by free variational parameters. The variational parameters are estimated to minimize the Kullback–Leibler (KL) divergence [23] between the variational distribution and the true posterior distribution. The latent variables in the sDCTM framework include the per-word topic structure δ d , t and the per-word topic assignment Z d , t for t = 1 , 2 , T and d = 1 , 2 , M . We define the approximate variational posterior for the dth document as
q ( δ d , 1 : T , Z d , 1 : T ) = q ( δ d , 1 , δ d , 2 , , δ d , T δ ^ d , 1 , δ ^ d , 2 , , δ ^ d , T , σ ^ 2 ) × t = 1 T q ( Z d , t γ d , t ) ,
for d = 1 , 2 , , M , where Z d , t γ d , t is M u l t ( 1 , γ d , t ) , and we obtain q ( δ d , 1 : T δ ^ d , 1 : T ) using a dynamic model with Gaussian “variational observations” { δ ^ d , 1 , δ ^ d , 2 , , δ ^ d , T } . This dynamic model is defined using a variational Kalman filter, where the variational parameters δ ^ d , 1 : T are treated as observations. Using the variational distribution, we form a variational state space model as follows. For  t = 1 , 2 , T ,
  • Observation Equation:
δ ^ d , t δ d , t N ( δ d , t , σ ^ 2 I ) .
  • State Equation:
δ d , t δ d , t 1 N ( Φ δ d , t 1 , Σ ) .
The Gaussian variational forward filtering distribution p ( δ d , t δ ^ d , 1 : t ) using standard Kalman filter calculations is characterized as follows. For t = 1 , 2 , , T ,
δ d , t δ ^ d , 1 : t N J ( m d , t , V d , t ) , m d , t = ( I K d , t ) Φ m d , t 1 + K d , t δ ^ d , t , V d , t = ( I K d , t ) [ Φ V d , t 1 Φ + Σ ] ,
where
K d , t = [ Φ V d , t 1 Φ + Σ ] ( Φ V d , t 1 Φ + Σ + σ ^ 2 I ) 1 ,
is the Kalman gain matrix and the initial conditions are specified by m d , 0 and V d , 0 .
Similarly, the Gaussian variational backward smoothing distribution p ( δ d , t 1 δ ^ d , 1 : T ) using standard Kalman smoothing calculations is characterized as follows. For  t = T ,   T 1 , , 1 ,
δ d , t 1 δ ^ d , 1 : T N J ( m ˜ d , t 1 , V ˜ d , t 1 ) , m ˜ d , t 1 = ( I J d , t 1 Φ ) m d , t 1 + J d , t 1 m d , t , V ˜ d , t 1 = V d , t 1 + J d , t 1 [ V ˜ d , t Φ V d , t 1 Φ Σ ] J d , t 1 ,
where
J d , t 1 = V d , t 1 Φ [ Φ V d , t 1 Φ + Σ ] 1 ,
and the initial conditions are specified by m d , T = m ˜ d , T and V d , T = V ˜ d , T .
Using the Kalman filter equations, the approximate variational posterior for the dth document is given by
q ( δ d , 1 : T , Z d , 1 : T ) = q ( δ d , t δ ^ d , 1 : T , σ ^ 2 ) × t = 1 T q ( Z d , t γ d , t ) ,
where Z d , t γ d , t is M u l t ( 1 , γ d , t ) and δ d , t δ ^ d , 1 : T , σ ^ 2 N ( m ˜ d , t , V ˜ d , t ) for t = 1 , 2 , , T and d = 1 , 2 , , M .

Constructing the Lower Bound

Given the variational distribution defined in Equation (4), our goal is to estimate the variational parameters γ d , t , δ ^ d , t , and  σ ^ 2 for t = 1 , 2 , , T and d = 1 , 2 , , M in order to minimize the KL divergence between the variational distribution q ( . ) defined in Equation (4) and the true posterior p ( . ) defined in Equation (3). This optimization problem of minimizing the KL divergence is equivalent to maximizing the evidence lower bound (ELBO) [24] defined below.
log p ( D d , y d | Φ , Σ , β , η ) = log Z d , 1 : T p ( Z d , 1 : T , δ d , 1 : T , D d , y d Φ , Σ , β , η ) d δ d , 1 : T , = log Z d , 1 : T p ( Z d , 1 : T , δ d , 1 : T , D d , y d Φ , Σ , β , η ) q ( δ d , 1 : T , Z d , 1 : T ) × q ( δ d , 1 : T , Z d , 1 : T ) d δ d , 1 : T E q log p ( Z d , 1 : T , δ d , 1 : T , D d , y d Φ , Σ , β , η ) E q log q ( δ d , 1 : T , Z d , 1 : T ) , = L ( δ ^ d , 1 : T , σ ^ 2 , γ d ; Φ , Σ , β , η , D d , y d ) .
We estimate the variational parameters γ d , t , δ ^ d , t , and  σ ^ 2 for t = 1 , 2 , , T and d = 1 , 2 , , M by maximizing the ELBO for approximate inference. In Appendix A, we derive the ELBO and provide the update equations in order to estimate the variational parameters. We estimate the variational parameter γ d , t using a fixed-point update, σ ^ 2 using a constrained optimization using the DBCPOL routine available in the IMSL library [25] for Fortran 77, and  δ ^ d , 1 : T using simulated annealing, which we describe in Section 2.2.1.

2.2. Parameter Estimation

In this section, we present a method for parameter estimation. In particular, given the corpus C with M documents and their associated responses denoted by ( D , y ) 1 : M , we wish to estimate the model parameters Φ J × J , Σ J × J , β J × V , and η J × C in order to maximize the log likelihood of the data given by
l ( Φ , Σ , β , η ( D , y ) 1 : M ) = d = 1 M log p ( D d , y d | Φ , Σ , β , η ) .
Because the log likelihood is analytically intractable, we consider a tractable lower bound on the log likelihood by summing the ELBO for log p ( D d , y d | Φ , Σ , β , η ) , defined in Section 2.1 over M documents. We estimate the model parameters via an alternating variational EM (VEM) procedure [26] described below:
  • E-Step: For each document, we find the optimizing values of the variational parameters described in Section 2.1.
  • M-Step: Maximize the resulting lower bound on the log likelihood with respect to the model parameters.
In Appendix A, we derive the lower bound on the log likelihood defined in Equation (8) and provide the updated equations in order to estimate the model parameters. We estimate model parameter β using a fixed-point update. The model parameters Φ J × J (with the stationarity condition), Σ (with the positive definite condition), and  η are estimated using a randomized search optimization technique, simulated annealing, which we describe in the next section.
For parameter estimation, we use frequentist methods to estimate the unknown parameters of the sDCTM model. As an alternative, one can set up a Bayesian model for parameters estimation in the sDCTM model by employing the Minnesota prior on Φ [27], inverse Wishart or Lewandowski–Kurowicka–Joe (LKJ) prior on Σ , Dirichlet prior on β , and multivariate normal prior on η .

2.2.1. Simulated Annealing

Simulated annealing (SA) is a probabilistic technique employed for optimizing problems to find a good approximation to the global optimum of a given function [28,29]. It is often used when the search space is discrete and is inspired by the annealing technique in metallurgy, which involves the heating and controlled cooling of a material to increase the size of its crystals and reduce their defects. The algorithm starts with a randomly generated solution and iteratively explores its neighboring states to find better solutions. The acceptance probability mechanism considers solutions that are worse than the current one in order to avoid getting stuck in a local optimum. Algorithm 1 presents a structured pseudocode to execute the simulated annealing technique for optimization problems, and Table 1 describes the initial parameters.
Algorithm 1 Simulated Annealing for Optimization
  • procedure SIMULATED ANNEALING( x , f )
  •    x opt = x and f opt = f , c o u n t = 0
  •   repeat
  •    for  N T times do
  •      for  N v times do
  •         x i = x i + r × v i , r i U ( 1 , 1 ) for i = 1 , 2 , , n
  •         f = CALCULATE f ( x )
  •        if  f < f  then
  •           x = x and f = f
  •           c o u n t = c o u n t + 1
  •        end if
  •        if  f < f opt  then
  •           x = x and f = f
  •           x opt = x and f opt = f
  •        end if
  •        if  f f  then
  •           p = exp f f T and p U ( 0 , 1 )
  •          if  p > p  then
  •             x = x and f = f
  •          end if
  •        end if
  •        Adjust v such that 50 % of the moves are accepted
  •        Set r a t i o = c o u n t N v
  •        if  r a t i o 0.6  then
  •           v i = v i × 1 + c × ( r a t i o 0.6 ) 0.4 for i = 1 , 2 , , n
  •        end if
  •        if  r a t i o 0.4  then
  •           v i = v i 1 + c × ( 0.4 r a t i o ) 0.4 for i = 1 , 2 , , n
  •        end if
  •      end for
  •      if  | f f | < ϵ and | f opt old f opt | < ϵ  then
  •        REPORT  x opt , f opt
  •      else
  •        Set x = x opt and T = r T × T
  •      end if
  •     end for
  •   until convergence or N times
  • end procedure

2.2.2. Selecting Initial Values of the Model Parameters

We use the SA technique, one of the top methods for non-derivative-based random search optimization, to estimate the model parameters η , Φ , and Σ in the sDCTM framework. To ensure that the optimization works effectively, it is important to select appropriate initial values. We describe the steps to select initial values for η , Φ , and Σ below.
  • Step 1. We select an initial search space within R n , where n represents the number of variables involved in the estimation (for instance, n = C × J for η and n = J × J for Φ and Σ ). The initial search space is chosen based on our belief of where the estimates are likely to be found. Then, we divide this search space into 2 n smaller subspaces.
  • Step 2. For Φ and Σ , we evaluate the optimization function at the midpoint within each of the subspaces and choose the subspace with the highest (or lowest) value of the function based on the optimization problem as our new search space. In cases where multiple subspaces optimize the function, we consider each of them in the next step for further exploration. In the sDCTM framework, for η , we consider the training accuracy as our optimization (maximization) function, as η plays a crucial role in predicting the response variable. For model parameters Φ and Σ , we consider the lower bound on the log likelihood as the optimization (maximization) function.
  • Step 3. We repeat steps 1 and 2 until no significant improvement in the optimization function value is seen.
  • Step 4. If a single subspace is chosen as a search space to select the initial values for optimization, we choose our starting values from this subspace and estimate the parameters using the SA technique. If there are multiple subspaces selected as the initial search space, we choose the initial values from each of these subspaces and run the SA to obtain the parameter estimates. We choose the final parameter estimates from the search space yielding the highest (or lowest) value of the function based on the optimization problem. The sDCTM model is implemented using a standalone code in Fortran 77 [30]. The code for model estimation is posted on https://github.com/NamithaVionaPais/sDTCM (accessed on 10 June 2024).

3. Simulation Study

To evaluate the performance of our model, we conducted a simulation study. We generated a dataset consisting of M = 120 documents, where each document is represented by a word time series of length T = 100 . The vocabulary size was set to V = 6 , and the documents were divided into C = 2 classes. Each word time series was governed by an underlying latent topic structure with J = 3 topics. Our model was evaluated using a six-fold cross-validation with M T r = 100 train and M T s t = 20 test documents. We conducted our analysis in Fortran 77. Given the observed word time series from M documents, we fitted the sDCTM model to estimate the model parameters, β , Σ , Φ , and η . As the model parameters are matrices, we assessed the estimation error using the squared Frobenius norm. The squared Frobenius norm lies between two matrices, such as M and N , and is defined as
d ( M , N ) = tr { ( M N ) ( M N ) }
The average squared Frobenius norm (over the six-fold cross-validation) for each of the model parameters is shown in Table 2. The low values of the squared Frobenius norm on the model parameters suggest that we are able to estimate the matrices well.
We used the estimated model parameter η and the variational parameter γ d , t for t = 1 , 2 , , T and d = 1 , 2 , , M T r to predict the response associated with the training data containing M T r = 100 documents as follows.
c d = argmax c { 1 , C } η c T γ ¯ d ,
for d = 1 , 2 , , M T r and γ ¯ d = T t = 1 T γ d , t . Table 3 shows the confusion matrix obtained on train data for one of the cross-validations. We were able to achieve a training accuracy of 100 % , and the average training accuracy across all k = 6 cross-validation datasets was also 100 % .
While classifying the test data, we estimated the variational parameters on the test documents given the estimated model parameters using variational inference. However, because we assume that we do not know the true label in the test documents, we replace the terms in the likelihood function associated with the response variable by the pseudo-label. The pseudo-label assigned to a test document is the class label associated with the nearest train document. We used the hamming distance to calculate the distance between the test and train sequence. Given the word time series, the associated pseudo-label, and the model parameters estimated using the train data, we estimated the variational parameters for each test document. We then obtained the test predictions using Equation (10). Table 4 shows the confusion matrix obtained on the test data for one of the cross-validation test datasets with a test accuracy of 80 % . The average test accuracy across all k = 6 cross-validation datasets was 70.83 % .
The model parameter, denoted as β , provides an estimate of how words are distributed across a set of J latent topics. Figure 1 illustrates the proportions of words across each of the J = 3 latent topics. We observed that Word 1 was highly probable within Topic 1, Words 3, 4, and 6 were highly probable within Topic 2, and Word 4 was highly probable within Topic 3.

4. Application for Promoter Sequence Identification in E. coli DNA Sub-Sequences

Proteins are one of the most important classes of biological molecules, being the carriers of the message contained in the DNA. An important process involved in the synthesis of proteins is called transcription. During transcription, a single-stranded RNA molecule, called messenger RNA, is synthesized (using the complementarity of the bases) from one of the strands of DNA corresponding to a gene (a gene is a segment of the DNA that codes for a type of protein). This process begins with the binding of an enzyme called RNA polymerase to a certain location on the DNA molecule. This exact site, which determines which of the two strands of DNA will be a transcript and in which direction, is recognized by the RNA polymerase due to the existence of certain regions of DNA placed near the transcription start site of a gene, called promoters. Because determining the promoter region in the DNA is an important step in the process of detecting genes, the problem of promoter identification is of major importance within the field of bioinformatics.
We considered a dataset of 106 E. coli DNA sub-sequences that is publicly available at https://archive.ics.uci.edu/dataset/67/molecular+biology+promoter+gene+sequences (accessed on 12 March 2023), among which 53 DNA sub-sequences contain promoters and the remaining 53 DNA sub-sequences do not contain promoters. Each sub-sequence consists of 57 nucleotides, represented by the four nucleotide bases: adenine (a), guanine (g), cytosine (c), and thymine (t). Given the two sets of DNA sub-sequences, with and without the presence of promoter regions, we used sDCTM to uncover the hidden patterns or motifs that could serve as markers for promoter presence to classify the sub-sequences. A detailed data description is provided in [31].
We present the results of the sDCTM model applied to analyze the E. Coli DNA sub-sequences. We treated each DNA sub-sequence as a document and the presence/absence of promoter regions as the associated class label. Each word in the document was represented by the nucleotide observed at that position in the sub-sequence. We divided the data into an 80–20 train–test split in order to assess the model performance. We trained the model on data with M t r = 86 (number of train DNA sub-sequences) and set T = 57 (length of each DNA sub-sequence), V = 4 (number of unique nucleotide levels), and C = 2 (indicating the levels of the response variable). To choose the number of topics J, we ran the model for different values of J and chose J based on the highest test accuracy. We set the number of latent topics as J = 3 . The confusion matrices obtained on the train and test data with M T s t = 20 (number of test DNA sub-sequences) using the sDCTM model are shown in Table 5 and Table 6. We were able to achieve a training accuracy of 100 % and a test accuracy of 90 % .
The distribution of nucleotides accross the three topics identified by the sDCTM model is shown in Figure 2. We observed that nucleotide levels a, c, and g were highly probable within Topic 1, nucleotide level t was highly probable within Topic 2, and nucleotide level a was highly probable within Topic 3. In addition, based on the model predictions, we also constructed a sequence logo plot [32] to identify the diversity of sequences between the two predicted groups. The sequence logo plot based on the train model predictions are shown in Figure 3. We can see that that the promoter presence plot demonstrated significantly higher conservation at various positions compared with the promoter absence plot. The higher bits values in the promoter presence plot highlight critical motifs essential for promoter activity. In contrast, the promoter absence plot showed much smaller variability, suggesting fewer conserved elements.

Comparison with Other Methods

We compared our sDCTM model with existing classification techniques, such as SVM, k-NN, and classification trees [33]. We used the k-NN classification technique, which is a popular approach used for sequence classification [10] on the observed data. We compared our approach to popular machine learning techniques like SVM and classification trees on the extracted features from the DNA sequence. We used the Haar wavelet transform and the simplest discrete wavelet transform (DWT) to extract features from the observed time series and build a classification model based on these features. We implemented the k-NN classification technique using the R package class [34], which identifies K = 5 nearest neighbors (in Euclidean distance) to classify the train and test data. We used the R package wavelets [35] to extract the DWT (with Haar filter) coefficients. The classification tree was implemented using the R package party [36], and the SVM was implemented using the R package e1071 [37]. We ran each of these methods using an 80–20 train–test split to assess model performance. The train and test accuracies are shown in Table 7. In comparison with the the other classification techniques, sDCTM performed better on both the train and test data.

5. Discussion

In this paper, we introduced the sDCTM model framework, which aims to classify categorical time series by dynamically modeling the underlying latent topics that capture the hidden thematic structure of the time series collection. We demonstrated the applicability of our method to real data by applying it to the promoter identification data to classify E. coli DNA sub-sequences based on promoter presence. Using sDCTM, we estimated the dynamic latent topic structure that serves as markers for promoter presence in a DNA sub-sequence and classify the sub-sequences. Among the J = 3 latent topics obtained using sDCTM, we observed that nucleotide levels a, c, and g were highly probable within Topic 1, nucleotide level t was highly probable within Topic 2, and nucleotide level a was highly probable within Topic 3. In addition, the logoplot based on the model predictions showed higher conservation at various positions in DNA sub-sequences with predicted promoter presence in comparison with DNA sub-sequences with predicted promoter absence. We compared our method with the k-NN, SVM, and classification tree method on a train–test data setup. Our comparative study results indicated that the sDCTM model performed better than the other classification techniques in both training and test datasets. This indicates that the estimated underlying latent topic structure captured by the sDCTM model is able to identify the promoter presence/absence in a DNA sub-sequence better in comparison with the other classification approaches. An extension of the sDCTM model accommodating word time series of varying lengths can be derived as a part of future work.

Author Contributions

Conceptualization, N.P., N.R. and S.R.; methodology, N.P., N.R. and S.R.; formal analysis, N.P., N.R. and S.R.; investigation, N.P., N.R. and S.R.; data curation, N.P.; writing—original draft preparation, N.P., N.R. and S.R.; writing—review and editing, N.P., N.R. and S.R.; visualization, N.P., N.R. and S.R.; supervision, N.R. and S.R.; project administration, N.R. and S.R. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data and analysis code are available at https://github.com/NamithaVionaPais/sDTCM (accessed on 10 June 2024), Github link.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A. Approximate Inference for sDCTM

We derive the evidence lower bound (ELBO) based on our choice of variational distribution defined in (4). Then, we derive the variational inference algorithm for the sDCTM model by maximizing the ELBO to estimate the variational parameters.

Appendix A.1. ELBO

L ( δ ^ d , 1 : T , σ ^ 2 , γ d ; Φ , Σ , β , η ) = E q log p ( Z d , 1 : T , δ d , 1 : T , D d , y d Φ , Σ , β , η ) E q log q ( δ d , 1 : T , Z d , 1 : T ) = t = 1 T E q log p ( w d , t Z d , t , β ) + t = 1 T E q log p ( Z d , t δ d , t ) + t = 1 T E q log p ( δ d , t δ d , t 1 , Φ , Σ ) + E q log p ( y d = 1 c Z d , 1 : T , η ) t = 1 T E q log q ( δ d , t δ ^ d , 1 : T , σ ^ 2 ) t = 1 T E q log q ( Z d , t γ d , t ) t = 1 T j = 1 J i = 1 V γ d t j w d t i log β j i + t = 1 T j = 1 J γ d t j m ˜ d t j log l = 1 J exp [ m ˜ d t l + v ˜ d t l 2 ] T J 2 log ( 2 π ) T 2 log | Σ | 1 2 t = 1 T t r Σ 1 ( V ˜ d t + Φ V ˜ d t Φ ) 1 2 t = 1 T ( m ˜ d t Φ m ˜ d t 1 ) Σ 1 ( m ˜ d t Φ m ˜ d t 1 ) + 1 T η c T γ d t log l = 1 C t = 1 T j = 1 J γ d t j exp ( 1 T η l j ) + T J 2 log ( 2 π ) + 1 2 t = 1 T log | V ˜ d , t | + T J 2 t = 1 T j = 1 J γ d t j l o g ( γ d t j ) t = 1 T γ d t log ( β ) w d t + t = 1 T γ d t m ˜ d t log J exp [ m ˜ d t l + d i a g V ˜ d t 2 ] T 2 log | Σ | 1 2 t = 1 T t r Σ 1 ( V ˜ d t + Φ V ˜ d t Φ ) 1 2 t = 1 T ( m ˜ d t Φ m ˜ d t 1 ) Σ 1 ( m ˜ d t Φ m ˜ d t 1 ) + 1 T η c T γ d t log l = 1 C t = 1 T exp η l T γ d t + 1 2 t = 1 T log | V ˜ d , t | + T J 2 t = 1 T γ d t log ( γ d t )

Appendix A.2. Variational Multinomial

L [ γ d t j ] t = 1 T j = 1 J i = 1 V γ d t j w d t i log β j i + t = 1 T j = 1 J γ d t j m ˜ d t j log l = 1 J exp [ m ˜ d t l + v ˜ d t l 2 ] + 1 T η c T γ d t log l = 1 C t = 1 T j = 1 J γ d t j exp ( 1 T η l j ) γ d t j log ( γ d t j ) = t = 1 T j = 1 J i = 1 V γ d t j w d t i log β j i + t = 1 T j = 1 J γ d t j m ˜ d t j log l = 1 J exp [ m ˜ d t l + v ˜ d t l 2 ] + 1 T η c j γ d t j log ( h T γ d t ) γ d t j log ( γ d t j ) t = 1 T j = 1 J i = 1 V γ d t j w d t i log β j i + t = 1 T j = 1 J γ d t j m ˜ d t j log l = 1 J exp [ m ˜ d t l + v ˜ d t l 2 ] + 1 T η c j γ d t j ( h T γ t o l d ) 1 h T γ t log ( h T γ t o l d ) + 1 γ d t j log ( γ d t j ) + Λ t ( j = 1 J γ t j 1 ) .
Result used: We know that
log ( x ) ζ 1 x + log ( ζ ) 1 , x > 0 , ζ > 0
with equality holding iff x = ζ . Then,
δ L [ γ d t j ] γ d t j = log β j v + m ˜ d t j log l = 1 J exp [ m ˜ d t l + v ˜ d t l 2 ] + 1 T η c j ( h T γ d t o l d ) 1 h j log ( γ d t j ) 1 + Λ t .
Equating to zero, we have a fixed-point estimate:
γ d t j β j v exp [ m ˜ d t j + 1 T η c j ( h T γ d t o l d ) 1 h j ] .
We can then normalize γ d t j so that j γ d t j = 1 .

Appendix A.3. Variational Gaussian

L [ δ ^ d t j ] τ = t 1 T j = 1 J γ d τ j m ˜ d τ j log l = 1 J exp [ m ˜ d τ l + v ˜ d τ l 2 ] 1 2 τ = t 1 T ( m ˜ d τ Φ m ˜ d τ 1 ) Σ 1 ( m ˜ d τ Φ m ˜ d τ 1 ) .
We estimate δ ^ d t j using simulated annealing.
L [ σ ^ 2 ] t = 1 T j = 1 J γ d t j m ˜ d t j log l = 1 J exp [ m ˜ d t l + v ˜ d t l 2 ] 1 2 t = 1 T t r Σ 1 ( V ˜ d t + Φ V ˜ d t Φ ) 1 2 t = 1 T ( m ˜ d t Φ m ˜ d t 1 ) Σ 1 ( m ˜ d t Φ m ˜ d t 1 ) + 1 2 t = 1 T log | V ˜ d , t | .
We obtain σ ^ 2 using DBCPOL in Fortran, which minimizes a function using a direct search complex algorithm.

Appendix A.4. Estimation for sDCTM

In this section, we provide update equations to estimate the model parameters based on the lower bound on the log likelihood of the data.
l ( Φ , Σ , β , η ) = d = 1 M log p ( D d , y d | Φ , Σ , β , η ) .
Then,
l ( Φ , Σ , β , η ) d = 1 M t = 1 T j = 1 J i = 1 V γ d t j w d t i log β j i + d = 1 M t = 1 T j = 1 J γ d t j m ˜ d t j log l = 1 J exp [ m ˜ d t l + v ˜ d t l 2 ] M T J 2 log ( 2 π ) M T 2 log | Σ | 1 2 d = 1 M t = 1 T t r Σ 1 ( V ˜ d t + Φ V ˜ d t Φ ) 1 2 d = 1 M t = 1 T ( m ˜ d t Φ m ˜ d t 1 ) Σ 1 ( m ˜ d t Φ m ˜ d t 1 ) + d = 1 M η c d T γ ¯ d log l = 1 C t = 1 T j = 1 J γ d t j exp ( 1 T η l j ) + M T J 2 log ( 2 π ) + 1 2 d = 1 M t = 1 T log | V ˜ d , t | + M T J 2 d = 1 M t = 1 T j = 1 J γ d t j log ( γ d t j ) .

Appendix A.5. Conditional Gaussian

L [ Φ ] d = 1 M t = 1 T j = 1 J γ d t j m ˜ d t j log l = 1 J exp [ m ˜ d t l + v ˜ d t l 2 ] 1 2 d = 1 M t = 1 T t r Σ 1 ( V ˜ d t + Φ V ˜ d t Φ ) 1 2 d = 1 M t = 1 T ( m ˜ d t Φ m ˜ d t 1 ) Σ 1 ( m ˜ d t Φ m ˜ d t 1 ) + 1 2 d = 1 M t = 1 T log | V ˜ d , t | .
We estimate Φ using simulated annealing.
L [ Σ ] d = 1 M t = 1 T j = 1 J γ d t j m ˜ d t j log l = 1 J exp [ m ˜ d t l + v ˜ d t l 2 ] M T 2 log | Σ | 1 2 d = 1 M t = 1 T t r Σ 1 ( V ˜ d t + Φ V ˜ d t Φ ) 1 2 d = 1 M t = 1 T ( m ˜ d t Φ m ˜ d t 1 ) Σ 1 ( m ˜ d t Φ m ˜ d t 1 ) + 1 2 d = 1 M t = 1 T log | V ˜ d , t | .
We estimate Σ using simulated annealing.

Appendix A.6. Conditional Multinomial

L [ β j i ] = d = 1 M t = 1 T j = 1 J i = 1 V γ d t j w t i log β j i + j = 1 J Λ j ( i = 1 V β j i 1 ) L β j i = d = 1 M t = 1 T γ d t j w t i 1 β j i + Λ j .
This leads to a fixed-point update:
β j i d = 1 M t = 1 T γ d t j w t i
We can then normalize β j .

Appendix A.7. Softmax Regression

L [ η 1 : C ] = d = 1 M η c d T γ ¯ d log l = 1 C t = 1 T j = 1 J γ d t j exp ( 1 T η l j ) .
We estimate the η using simulated annealing. We add an L 2 regularization term to prevent overfitting and prevent the inflation of parameter values.

References

  1. Le Nguyen, T.; Gsponer, S.; Ilie, I.; O’reilly, M.; Ifrim, G. Interpretable Time Series Classification using Linear Models and Multi-resolution Multi-domain Symbolic Representations. Data Min. Knowl. Discov. 2019, 33, 1183–1222. [Google Scholar] [CrossRef]
  2. Neuhauser, D.S.; Allen, R.M.; Zuzlewski, S. Northern California Earthquake Data Center: Data Sets and Data Services. In AGU Fall Meeting Abstracts; American Geophysical Union: Washington, DC, USA, 2015; Volume 2015, p. S53A-2774. [Google Scholar]
  3. Olszewski, R.T. Generalized Feature Extraction for Structural Pattern Recognition in Time-Series Data; Carnegie Mellon University: Pittsburgh, PA, USA, 2001. [Google Scholar]
  4. Bagnall, A.; Lines, J.; Bostrom, A.; Large, J.; Keogh, E. The Great Time Series Classification Bake Off: A Review and Experimental Evaluation of Recent Algorithmic Advances. Data Min. Knowl. Discov. 2017, 31, 606–660. [Google Scholar] [CrossRef]
  5. Billingsley, P. Statistical Methods in Markov Chains. In The Annals of Mathematical Statistics; Institute of Mathematical Statistics: Hayward, CA, USA, 1961; pp. 12–40. [Google Scholar]
  6. Fahrmeir, L.; Kaufmann, H. Regression Models for Non-Stationary Categorical Time Series. J. Time Ser. Anal. 1987, 8, 147–160. [Google Scholar] [CrossRef]
  7. Fokianos, K.; Kedem, B. Prediction and Classification of Non-Stationary Categorical Time Series. J. Multivar. Anal. 1998, 67, 277–296. [Google Scholar] [CrossRef]
  8. Navarro, G. A Guided Tour to Approximate String Matching. ACM Comput. Surv. (CSUR) 2001, 33, 31–88. [Google Scholar] [CrossRef]
  9. Jurafsky, D.; Martin, J.H. Naïve Bayes Classifier Approach to Word Sense Disambiguation. In Computational Lexical Semantics; University of Groningen: Groningen, The Netherlands, 2009. [Google Scholar]
  10. Deshpande, M.; Karypis, G. Evaluation of Techniques for Classifying Biological Sequences. In Proceedings of the Pacific-Asia Conference on Knowledge Discovery and Data Mining, Taipei, Taiwan, 6–8 May 2002; Springer: Berlin/Heidelberg, Germany, 2002; pp. 417–431. [Google Scholar]
  11. Stoffer, D.S.; Tyler, D.E.; Wendt, D.A. The Spectral Envelope and its Applications. In Statistical Science; Institute of Mathematical Statistics: Hayward, CA, USA, 2000; pp. 224–253. [Google Scholar]
  12. Li, Z.; Bruce, S.A.; Cai, T. Classification of Categorical Time Series Using the Spectral Envelope and Optimal Scalings. arXiv 2021, arXiv:2102.02794. [Google Scholar]
  13. Yan, X.; Guo, J.; Liu, S.; Cheng, X.; Wang, Y. Learning Topics in short texts by Non-Negative Matrix Factorization on Term Correlation Matrix. In Proceedings of the 2013 SIAM International Conference on Data Mining. SIAM, Austin, TX, USA, 2–4 May 2013; pp. 749–757. [Google Scholar]
  14. Blei, D.M.; Ng, A.Y.; Jordan, M.I. Latent Dirichlet Allocation. J. Mach. Learn. Res. 2003, 3, 993–1022. [Google Scholar]
  15. Roberts, M.E.; Stewart, B.M.; Tingley, D.; Airoldi, E.M. The Structural Topic Model and Applied Social Science. In Advances in Neural Information Processing Systems Workshop on Topic Models: Computation, Application, and Evaluation. Harrahs and Harveys, Lake Tahoe; Harvard University: Cambridge, MA, USA, 2013; Volume 4, pp. 1–20. [Google Scholar]
  16. Blei, D.M.; Lafferty, J.D. Dynamic Topic Models. In Proceedings of the 23rd International Conference on Machine Learning, Pittsburgh, PA, USA, 25–29 June 2006; pp. 113–120. [Google Scholar]
  17. Wang, C.; Blei, D.; Heckerman, D. Continuous Time Dynamic Topic Models. arXiv 2012, arXiv:1206.3298. [Google Scholar]
  18. Ionescu-Tîrgovişte, C.; Gagniuc, P.A.; Guja, C. Structural properties of gene promoters highlight more than two phenotypes of diabetes. PLoS ONE 2015, 10, e0137950. [Google Scholar] [CrossRef] [PubMed]
  19. Coles, R.; Caswell, R.; Rubinsztein, D.C. Functional analysis of the Huntington’s disease (HD) gene promoter. Hum. Mol. Genet. 1998, 7, 791–800. [Google Scholar] [CrossRef]
  20. Oubounyt, M.; Louadi, Z.; Tayara, H.; Chong, K.T. DeePromoter: Robust Promoter Predictor using Deep Learning. Front. Genet. 2019, 10, 286. [Google Scholar] [CrossRef] [PubMed]
  21. Le, N.Q.K.; Yapp, E.K.Y.; Nagasundaram, N.; Yeh, H.Y. Classifying Promoters by Interpreting the Hidden Information of DNA Sequences via Deep Learning and Combination of Continuous FastText N-Grams. Front. BIoeng. BIotechnol. 2019, 7, 305. [Google Scholar] [CrossRef] [PubMed]
  22. Li, Q.Z.; Lin, H. The Recognition and Prediction of σ70 Promoters in Escherichia Coli K-12. J. Theor. Biol. 2006, 242, 135–141. [Google Scholar] [CrossRef] [PubMed]
  23. Kullback, S.; Leibler, R.A. On Information and Sufficiency. Ann. Math. Stat. 1951, 22, 79–86. [Google Scholar] [CrossRef]
  24. Blei, D.M.; Kucukelbir, A.; McAuliffe, J.D. Variational Inference: A Review for Statisticians. J. Am. Stat. Assoc. 2017, 112, 859–877. [Google Scholar] [CrossRef]
  25. Visual Numerics. IMSL® Fortran Numerical Math Library; Visual Numerics Inc.: Houston, TX, USA, 2007. [Google Scholar]
  26. Dempster, A.P.; Laird, N.M.; Rubin, D.B. Maximum Likelihood from Incomplete Data via the EM Algorithm. J. R. Stat. Soc. Ser. B (Methodol.) 1977, 39, 1–22. [Google Scholar] [CrossRef]
  27. Lütkepohl, H. New Introduction to Multiple Time Series Analysis; Springer Science & Business Media: New York, NY, USA, 2005. [Google Scholar]
  28. Goffe, W.L.; Ferrier, G.D.; Rogers, J. Global Optimization of Statistical Functions with Simulated Annealing. J. Econom. 1994, 60, 65–99. [Google Scholar] [CrossRef]
  29. Rajasekaran, S. On Simulated Annealing and Nested Annealing. J. Glob. Optim. 2000, 16, 43–56. [Google Scholar] [CrossRef]
  30. Brainerd, W. Fortran 77. Commun. ACM 1978, 21, 806–820. [Google Scholar] [CrossRef]
  31. Czibula, G.; Bocicor, M.I.; Czibula, I.G. Promoter sequences prediction using relational association rule mining. Evol. Bioinform. 2012, 8, EBO-S9376. [Google Scholar] [CrossRef] [PubMed]
  32. Schneider, T.D.; Stephens, R.M. Sequence logos: A new way to display consensus sequences. Nucleic Acids Res. 1990, 18, 6097–6100. [Google Scholar] [CrossRef] [PubMed]
  33. Zhao, Y. R and Data Mining: Examples and Case Studies; Academic Press: New York, NY, USA, 2012. [Google Scholar]
  34. Venables, W.N.; Ripley, B.D. R Package Class Modern Applied Statistics with S, 4th ed.; Springer: New York, NY, USA, 2002; ISBN 0-387-95457-0. [Google Scholar]
  35. Percival, D.B.; Walden, A.T. Wavelet Methods for Time Series Analysis; Cambridge University Press: Cambridge, UK, 2000; Volume 4. [Google Scholar]
  36. Strobl, C.; Malley, J.; Tutz, G. An introduction to recursive partitioning: Rationale, application, and characteristics of classification and regression trees, bagging, and random forests. Psychol. Methods 2009, 14, 323. [Google Scholar] [CrossRef] [PubMed]
  37. Meyer, D.; Wien, F. Support Vector Machines. R News 2001, 1, 23–26. [Google Scholar]
Figure 1. Estimated distribution of words over each of the J = 3 latent topics on the simulated data.
Figure 1. Estimated distribution of words over each of the J = 3 latent topics on the simulated data.
Algorithms 17 00275 g001
Figure 2. Estimated distribution of V = 4 nucleotides over each of the J = 3 latent topics.
Figure 2. Estimated distribution of V = 4 nucleotides over each of the J = 3 latent topics.
Algorithms 17 00275 g002
Figure 3. Sequence logo plot based on model predictions on the train data.
Figure 3. Sequence logo plot based on model predictions on the train data.
Algorithms 17 00275 g003
Table 1. Initial parameters for simulated annealing.
Table 1. Initial parameters for simulated annealing.
ParameterDescription
x n dimensional initial vector.
v n dimensional initial step length vector.
fObjective function to minimize.
ϵ Convergence criteria.
TTemperature
r T Temperature reduction factor.
NMaximum number of function evaluations.
N T Number of function evaluations before T adjustment.
N v Number of function evaluations before v adjustment.
c o u n t Calculate the number of acceptances.
cControls how fast v adjusts.
Table 2. Average squared Frobenius norm between the true and estimated model parameters using the sDCTM model on the simulated dataset.
Table 2. Average squared Frobenius norm between the true and estimated model parameters using the sDCTM model on the simulated dataset.
ParameterSquared Frobenius Norm
Φ 0.5270
Σ 0.3697
β 0.1622
η 0.2385
Table 3. Confusion matrix obtained from the sDCTM model on the test data of M = 20 documents based on a simulation.
Table 3. Confusion matrix obtained from the sDCTM model on the test data of M = 20 documents based on a simulation.
Actual
PredictedClass 1Class 2
Class 1800
Class 2020
Table 4. Confusion matrix obtained from the sDCTM model on the test data in a simulation.
Table 4. Confusion matrix obtained from the sDCTM model on the test data in a simulation.
Actual
PredictedClass 1Class 2
Class 190
Class 247
Table 5. Confusion matrix obtained from the sDCTM model on the promoter identification train data with M = 86 DNA sub-sequences.
Table 5. Confusion matrix obtained from the sDCTM model on the promoter identification train data with M = 86 DNA sub-sequences.
Actual
PredictedPromoter PresencePromoter Absence
Promoter presence430
Promoter absence043
Table 6. Confusion matrix obtained from the sDCTM model on the promoter identification test data with M = 20 DNA sub-sequences.
Table 6. Confusion matrix obtained from the sDCTM model on the promoter identification test data with M = 20 DNA sub-sequences.
Actual
PredictedPromoter PresencePromoter Absence
Promoter presence100
Promoter absence28
Table 7. Comparing accuracy using the sDCTM, k-NN, classification tree, and SVM methods on the train and test data.
Table 7. Comparing accuracy using the sDCTM, k-NN, classification tree, and SVM methods on the train and test data.
MethodTrain AccuracyTest Accuracy
sDCTM100%90%
k-NN84.8%85%
Classification tree87.2%85%
SVM100%55%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Pais, N.; Ravishanker, N.; Rajasekaran, S. Supervised Dynamic Correlated Topic Model for Classifying Categorical Time Series. Algorithms 2024, 17, 275. https://doi.org/10.3390/a17070275

AMA Style

Pais N, Ravishanker N, Rajasekaran S. Supervised Dynamic Correlated Topic Model for Classifying Categorical Time Series. Algorithms. 2024; 17(7):275. https://doi.org/10.3390/a17070275

Chicago/Turabian Style

Pais, Namitha, Nalini Ravishanker, and Sanguthevar Rajasekaran. 2024. "Supervised Dynamic Correlated Topic Model for Classifying Categorical Time Series" Algorithms 17, no. 7: 275. https://doi.org/10.3390/a17070275

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop