Next Article in Journal
Stower-13: A Multi-View Inspection Image Dataset for the Automatic Classification and Naming of Tension Towers
Next Article in Special Issue
Bias and Cyberbullying Detection and Data Generation Using Transformer Artificial Intelligence Models and Top Large Language Models
Previous Article in Journal
Outdoor-to-Indoor mmWave Relaying with Massive MIMO: Impact of Imperfect Channel Estimation
Previous Article in Special Issue
Zero-FVeinNet: Optimizing Finger Vein Recognition with Shallow CNNs and Zero-Shuffle Attention for Low-Computational Devices
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Hidden Variable Models in Text Classification and Sentiment Analysis

Concordia Institute for Information Systems Engineering (CIISE), Montreal, QC H3G 1M8, Canada
*
Author to whom correspondence should be addressed.
Electronics 2024, 13(10), 1859; https://doi.org/10.3390/electronics13101859
Submission received: 25 March 2024 / Revised: 30 April 2024 / Accepted: 7 May 2024 / Published: 10 May 2024
(This article belongs to the Special Issue Emerging Artificial Intelligence Technologies and Applications)

Abstract

:
In this paper, we are proposing extensions to the multinomial principal component analysis (MPCA) framework, which is a Dirichlet (Dir)-based model widely used in text document analysis. The MPCA is a discrete analogue to the standard PCA (it operates on continuous data using Gaussian distributions). With the extensive use of count data in modeling nowadays, the current limitations of the Dir prior (independent assumption within its components and very restricted covariance structure) tend to prevent efficient processing. As a result, we are proposing some alternatives with flexible priors such as generalized Dirichlet (GD) and Beta-Liouville (BL), leading to GDMPCA and BLMPCA models, respectively. Besides using these priors as they generalize the Dir, importantly, we also implement a deterministic method that uses variational Bayesian inference for the fast convergence of the proposed algorithms. Additionally, we use collapsed Gibbs sampling to estimate the model parameters, providing a computationally efficient method for inference. These two variational models offer higher flexibility while assigning each observation to a distinct cluster. We create several multitopic models and evaluate their strengths and weaknesses using real-world applications such as text classification and sentiment analysis.

1. Introduction

In this fast-paced world of technological advances, one of the most significant contributing factors has been the emergence of various digital data forms, opening opportunities in different fields to gather helpful information. Everyday, massive amounts of digital data are stored in digital data archives. The same distinction can be made for the enormous quantity of textual data available on the Internet. Therefore, it is critical to developing effective and scalable statistical models to extract hidden knowledge from such rich data [1].
One of the main challenges in the statistical analysis of textual data is capturing and representing their complexity. Different approaches have been applied to deal with this problem. Furthermore, due to information technology’s rapid development, vast quantities of scientific documents are now freely available to be mined. Thus, the analysis and mining of scientific documents have been very active research areas for many years.
Data projection and clustering are crucial for document analysis, with projections aimed at creating low-dimensional, meaningful data representations and clustering and grouping similar data patterns [2,3]. Traditionally, these methods are studied separately, but they intersect in many applications [3]. K-means clustering, though widely used for creating compact cluster representations, does not fully capture document semantics. This gap has led to the adoption of machine learning and deep learning for text mining challenges, including text classification [4], summarization [5], segmentation [6], topic modeling [7], and sentiment analysis [8].
In this paper, we will focus on topic modeling aspects. Topic models are generally classified into two categories: those based on matrix decomposition, like singular value decomposition (SVD), and generative models [9]. The matrix decomposition approach, such as probabilistic latent semantic analysis (PLSA) [10,11], analyzes text via mining and requires a deep understanding of the corpus structure. PLSA, also known as probabilistic latent semantic indexing (pLSI) [11], represents documents as a mix of topics by performing matrix decomposition on the term–document matrix and is effective in identifying relevant words for each topic. In contrast, the generative approach of topic modeling focuses on the context of words across the entire document corpus. These models use latent variable models, and a document is treated as a combination of various topics, each represented by a random vector of words [3].
Meanwhile, the research by [12] indicates that while the probabilistic latent semantic indexing (pLSI) model offers some insights, it falls short in clustering and as a generative model due to its inability to generalize to new documents. To address these limitations, latent Dirichlet allocation (LDA) [12] was introduced, enhancing pLSI using a Dirichlet distribution for topic mixtures. LDA stands out as a more effective generative model, though it still lacks robust clustering capabilities [3]. The integration of clustering and projection into a single framework has been a recent focus in this field, recognizing the need to combine these two approaches [13,14].
The LDA model [12] has been proposed to solve these shortcomings, and this model has proved to be an efficient and scalable data processing method [15,16].
The main issue with current text analysis models is their failure to clearly define a probability model encompassing hidden variables and assumptions [11,17,18,19]. To address this, variational Expectation Maximization (EM) has been utilized, notably in Multinomial PCA (MPCA), which links topics to latent mixture proportions in a probabilistic matrix factorization framework [19,20]. Extensions of LDA, like its hierarchical [21] and online versions [22], have been developed, although they lack the integration of Dirichlet priors in modeling. Researchers have explored alternative models using conjugate priors and methods, like Gibbs sampling and Markov Chain Monte Carlo (MCMC) methods [23], which, despite their effectiveness, require longer convergence times compared to the variational Bayes approach.
In this paper, we introduce two novel models, GDMPCA and BLMPCA, that significantly improve text classification and sentiment analysis by combining generalized Dirichlet (GD) and Beta-Liouville (BL) distributions for a more in-depth understanding of text data complexities [16,24,25]. Both models employ variational Bayesian inference and collapsed Gibbs sampling for efficient and scalable computational performances, which is critical for handling large datasets.
The generalized Dirichlet (GD) distribution, introduced in [26], exhibits a more flexible covariance structure than its Dirichlet counterpart. Similarly, the Beta-Liouville (BL) distribution, enriched with additional parameters, offers improved adjustments for data spread and modeling efficiency. Our contribution was validated through a rigorous empirical evaluation on real-world datasets, which demonstrated our models’ superior accuracy and adaptability. This work represents a significant step forward in text analysis methodologies, bridging theoretical innovation with practical application, with experimental results demonstrating the relationships between these models.
The structure of the rest of this paper is as follows. In Section 2, we cover the related work. Section 3 introduces the extension of MPCA with generalized Dirichlet and Beta-Liouville distributions with all the details about the parameters estimation. Section 4 is devoted to the discussion of the experimental results. Finally, we conclude our work in Section 6.

2. Related Work

In this section, we delve into the vast array of the literature on topic modeling approaches. The foundation of this field is built upon traditional topic modeling techniques [10,11], with significant contributions from topic-class modeling [27,28,29] and the nuanced exploration of global and local document features [30,31].
Innovative strides have been made with the introduction of a two-stage topic extraction model for bibliometric data analysis, employing word embeddings and clustering for a more refined topic analysis [32]. This approach provides a nuanced lens with which to view the thematic undercurrents of scholarly communication.
The landscape of sentiment analysis is similarly evolving, with breakthroughs like a term-weighted neural language model paired with a stacked bidirectional LSTM (long short-term memory) framework, enhancing the detection of subtle sentiments like sarcasm in text [33]. Such advancements offer deeper insights into the complexities of language and its sentiments.
Cross-modal sentiment analysis also takes center stage with deep learning techniques, as seen in works that identify emotions from facial expressions [34]. These studies, which utilize convolutional neural networks and Inception-V3 transfer learning [35], pave the way for multimodal sentiment analysis, potentially influencing strategies for textual sentiment analysis.
A hybrid deep learning method has been introduced for analyzing sentiment polarities and knowledge graph representations, particularly focusing on health-related social media data, like tweets on monkeypox [36]. This underscores the importance of versatile and dynamic models in interpreting sentiment from real-time data streams.
Collectively, these contemporary works highlight the expansive applicability and dynamic nature of deep learning across various domains and data types. Their inclusion in our review underlines the potential for future cross-disciplinary research, expanding the scope of sentiment analysis to include both text and image data.
Alongside these emerging approaches, well-established techniques such as principal component analysis (PCA) and its text retrieval counterpart, latent semantic indexing [37], continue to be pivotal. Probabilistic latent semantic indexing (pLSI) [11] and latent Dirichlet allocation (LDA) [12] further enrich the discussion on discrete data and topic modeling. Non-negative matrix factorization (NMF) [17] has also demonstrated effectiveness, emphasizing the need for models that can simultaneously handle clustering and projection. In addressing a gap in the literature, a multinomial PCA model has been proposed to offer probabilistic interpretations of the relationships between documents, clusters, and factors [19].
Our focus on the MPCA model and its extensions aims to consolidate these disparate strands of research, presenting a comprehensive framework for topic modeling that accounts for both clustering and projection, reflecting the ongoing dialogue within the research community on these topics.

2.1. Multinomial PCA

Probabilistic approaches to reducing dimensions generally hypothesize that each observation x i corresponds to a hidden variable, referred to as a latent variable θ i . This latent variable exists within a subspace of dimension K. Typically, the relationship involves a linear mapping ( β ) within the latent space coupled with a probabilistic mechanism.
In the probabilistic PCA (pPCA) framework, as detailed in the work in [38], it is posited that each observation x i originates from a standard Gaussian distribution N K ( 0 K ; Z K ) . The assumption of a Gaussian distribution is also employed for the conditional distribution of the observations:
x i | θ i N v ( β θ i + μ , σ 2 Z V )
where Z is a “standard” normal distribution, ( β , μ ) are the model parameters, and σ 2 is the variance that is learned using maximum likelihood inference.
The Gaussian assumption is suitable for real-valued data, yet it is less applicable to non-negative count data. Addressing this, [19] introduced a variant of pPCA where the latent variables are modeled as a discrete probability distribution, specifically using a Dirichlet distribution, where as m D i r ( α ) ,
D ( m ; α ) = 1 Z ( α ) k = 1 K m k α k 1 ( m )
where α = ( α 1 , , α k ) 0 .
Then, the probabilistic function is assumed to be multinomial:
m D i r i c h l e t ( α ) C M u l t i n o m i a l ( m , L ) , w k M u l t i n o m i a l ( Ω k , c k )
The variables m and w are assumed to be hidden parameters for each document. For the parameter estimation of MPCA, first, the variable Ω is estimated using the Dirichlet prior on m using parameters α [19]. The likelihood model for the MPCA is given as follows [20]:
p ( m , w | α , Ω ) = Γ ( k α k ) k Γ ( α k ) C w 1 , 1 , w 1 , 2 , w k , 1 , w 1 , J w K , J L K m K a k 1 k , j m k w k , j Ω k , j w k , j
In the MPCA model, it is assumed that each observation x i can be broken down into a probabilistic mixture of K topics that represent the whole corpus. Then, m indicates the observation with mixture weights in the latent space, and Ω is a global parameter that encapsulates all the information at the corpus level.
As a result, the following equation is derived when the hidden variables have a Dirichlet prior [19]:
m D i r i c h l e t ( α ) Ω k D i r i c h l e t ( 2 f )
The following updated formula converges to the local maximum log p ( Ω , α m | r ) , where Γ ( k α k ) k Γ ( α k ) is a normalizing constant for the Dirichlet, and r is the total row-wise number of words in the document representation with the k component [19]:
γ j , k , [ i ] = Γ ( k α k ) k Γ ( α k ) 1 Ω k , j m k , [ i ]
m k , [ i ] = Γ ( k α k ) k Γ ( α k ) ( a k 1 + j r j , [ i ] γ j , k , [ i ] )
Equations (8) and (9) are the parameters for a multinomial and a Dirichlet, respectively.
Ω k , j = Γ ( k α k ) k Γ ( α k ) ( 2 f + i r j , [ i ] γ j , k , [ i ] )
Ψ 0 ( a k ) Ψ 0 ( k a k ) = log ( 1 / k ) + i log ( m k , [ i ] ) 1 + I
According to the exponential family definition (Appendix A), Equation (9) rewrites α in terms of its dual representation. Minka’s approach is used to derive α , where n k is the number of times that the outcome was k [39]:
n k = i δ ( x i k ) n i = k n i k
α k n e w = a k i Ψ ( n i k + a k ) Ψ ( a k ) i Ψ ( n i k + k a k ) Ψ ( k a k )

Connection between MPCA and LDA

The multinomial PCA model is closely connected to LDA [12] and forms the foundation over several topic models.
In text analysis, an observation typically refers to a document represented by a sequence of tokens or words, denoted as w i = w i n , where n = 1 L i . Each word w i n within a document i is initially linked to a topic, which is specified by a vector z i n that is derived from a M u l t i n o m i a l ( 1 , β k ) distribution. The model for any given document i can be described as follows:
θ k D i r i c h l e t ( α ) z i n | θ i M u l t i n o m i a l ( 1 , θ i ) w i n | z i n k M u l t i n o m i a l ( 1 , β k )
At the word level, marginalizing on z i n yields a distribution similar to Equation (3):
w i n | θ i M u l t i n o m i a l ( 1 , β θ i )
Furthermore, the distinction between LDA and MPCA is that LDA is a word-level model, whereas MPCA is a document-level model. Since GDMPCA and BLMPCA are new variations of MPCA, both new models are assumed to be document-level in the following proposed approaches.

3. Proposed Models

In this section, we present two pioneering models, generalized Dirichlet Multinomial Principal Component Analysis (GDMPCA) and Beta-Liouville Multinomial Principal Component Analysis (BLMPCA), which were designed to revolutionize text classification and sentiment analysis. At the core of our approaches is the integration of generalized Dirichlet and Beta-Liouville distributions, respectively, into the PCA framework. This integration is pivotal, as it allows for a more nuanced representation of text data, capturing the inherent sparsity and thematic structures more effectively than traditional methods.
The GDMPCA model leverages the flexibility of the generalized Dirichlet distribution to model the variability and co-occurrence of terms within documents, enhancing the model’s ability to discern subtle thematic differences. On the other hand, the BLMPCA model utilizes the Beta-Liouville distribution to precisely capture the polytopic nature of texts, facilitating a deeper understanding of sentiment and thematic distributions. Both models employ variational Bayesian inference, offering a robust mathematical framework that significantly improves computational efficiency and scalability. This approach not only aids in handling large datasets with ease but also ensures that the models remain computationally viable without sacrificing accuracy.
To elucidate the architecture of our proposed models, we delve into the algorithmic underpinnings, detailing the iterative processes that underlie the variational Bayesian inference technique. This includes a comprehensive discussion of the optimization strategies employed to enhance convergence rates and ensure the stability of the models across varied datasets. Moreover, we provide a comparative analysis, drawing parallels and highlighting distinctions between our models and existing text analysis methodologies. This comparison underscores the superior performances of GDMPCA and BLMPCA in terms of accuracy, adaptability, and computational efficiency, as evidenced by an extensive empirical evaluation on diverse real-world datasets.
Our exposition on the practical implications of these models reveals their broad applicability across numerous domains, from automated content categorization to nuanced sentiment analysis in social media texts. The innovative aspects of the GDMPCA and BLMPCA models, coupled with their empirical validation, underscore their potential to set a new standard in text analysis, offering researchers and practitioners alike powerful tools for uncovering insights from textual data.
Table 1 summarizes the relevant variables for the proposed models.

3.1. Generalized Dirichlet Multinomial PCA

Bouguila [40] demonstrated that when mixture models are used, the generalized Dirichlet (GD) distribution is a reasonable alternative to the Dirichlet distribution for clustering count data.
As we mentioned previously, the GD distribution, like the Dirichlet distribution, is a conjugate prior to the multinomial distribution. Furthermore, the GD has a more general covariance matrix [40].
Therefore, the variational Bayes approach will be utilized to develop an extension of the MPCA model incorporating the generalized Dirichlet assumption. GDMPCA is anticipated to perform effectively because the Dirichlet distribution is a specific instance of the GD [41]. Like MPCA, GDMPCA is a fully generative model applied to a corpus. It considers a collection of M documents represented as the corpus, denoted by D = { w 1 , w 2 , , w M } . Each document w m consists of a sequence of N m words, expressed as w m = ( w m 1 , , w m N m ) . Words within a document are represented by binary vectors from a vocabulary of V words, where if the j-th word is selected, w j n = 1 , and if not, w j n = 0 [42]. The GDMPCA model then describes the generation of each word in the document through a series of steps involving c, a d + 1 dimensional binary vector of topics:
m G D ( ξ ) z M u l t i n o m i a l ( m , L ) w k M u l t i n o m i a l ( Ω k , c k )
If the i-th topic is chosen, z i n = 1 , and in other cases, z i n = 0 . m = ( m 1 , , m d + 1 ) , where m d + 1 = 1 i = 1 d m i .
The multinomial probability p ( w n z n , Ω w ) is conditioned on the variable z n . The distribution G D ( ξ ) is a d-variate generalized Dirichlet distribution characterized by the parameter set ξ = ( a 1 , b 1 , , a d , b d ) , with its probability distribution function denoted by p, where γ i = b i a i + 1 b i + 1 [42]:
p ( m 1 , , m d | ξ ) = i = 1 d Γ ( a l + b l ) Γ ( a l ) Γ ( b l ) m i a l 1 ( 1 j = 1 i m j ) γ i
The GD distribution simplifies to a Dirichlet distribution when b i = a ( i + 1 ) + b ( i + 1 ) .
The mean and the variance matrix of the GD distribution are as follows [41]:
E ( m i ) = a l a l + b l k = 1 i 1 b k a k + b k
v a r ( m i ) = E ( m i ) a l + 1 a l + b l + 1 k = 1 i 1 b k + 1 a k + b k + 1 E ( θ i )
and the covariance between m i and m j is given by
c o v ( m i , m j ) = E ( m j ) a l a l + b l + 1 k = 1 i 1 b k + 1 a k + b k + 1 E ( m i )
The covariance matrix of the GD distribution offers greater flexibility compared to the Dirichlet distribution, due to its more general structure. This additional complexity allows for an extra set of parameters, providing d 1 additional degrees of freedom, which enables the GD distribution to more accurately model real-world data. Indeed, the GD distribution fits count data better than the commonly used Dirichlet distribution [43]. The Dirichlet and GD distributions are both members of the exponential family (Appendix A). Furthermore, they are also conjugate priors to the multinomial distribution. As a result, we can use the following method to learn the model.
The likelihood for the GDMPCA is given as follows:
p ( m , w | ξ , Ω ) = Γ ( a i + b i ) Γ ( a i ) Γ ( b i ) z w 1 , 1 , w 1 , 2 , w k , 1 , w 1 , J w K , J L m k b k 1 1 i = 1 k 1 [ m i a i 1 ( j = 1 k m j ) b i 1 + ( a i + b i ) ] k , j m k w k , j Ω k , j w k , j
Hence, when hidden variables are assigned GD priors, and given a defined universe of words, we use an empirical prior derived from the observed proportions of words in the universe, denoted by f, where k f k = 1 . The equation, then, is structured as follows:
m G D ( ξ ) Ω k G D ( 2 f )
where 2 shows the small size of the prior sample size.
First, we will calculate the parameters of GD utilizing the Hessian matrix as described in Appendix B.1.2, following Equations (19) and (20). To find the optimal variational parameters, we minimize the Kullback–Leibler (KL) divergence between the variational distribution and the posterior distributions p ( m , w | Ω , ξ ) . This is achieved through a repetitive fixed-point method. We specify the variational parameters as follows:
q ( m , c | γ , Φ ) = q ( m | γ ) k = 1 K q ( c k | Φ k )
As an alternative to the posterior distribution p ( m , c , w , ξ , Ω ) , we determine the variational parameters γ and Φ through a detailed optimization process outlined subsequently. To simplify, Jensen’s inequality is applied to establish a lower bound on the log likelihood, which allows us to disregard parameters γ and Φ [44]:
log p ( w | ξ , Ω ) = log z p ( m , c , w | ξ , Ω ) d m = log z p ( m , c , w | ξ , Ω ) q ( m , c ) q ( m , c ) d m z log p ( m , c , w | ξ , Ω ) q ( m , c ) d m z q ( m , c ) log q ( m , c ) d m = E [ log p ( m , c , w | ξ , Ω ) ] E [ log q ( m , c ) ]
Consequently, Jensen’s inequality provides a lower bound on the log likelihood for any given variational distribution q ( m , c | γ , Φ ) .
If the right-hand side of Equation (22) is denoted as L ( γ , Φ ; ξ , Ω ) , the discrepancy between the left and right sides of this equation represents the KL divergence between the variational distribution and the true posterior probabilities. This re-establishes the importance of the variational parameters, leading to the following expression:
log p ( w | ξ , Ω ) = L ( γ , Φ ; ξ , Ω ) + D ( q ( m , c | γ , Φ ) ) | p ( m , c | x , ξ , Ω )
As demonstrated in Equation (23), maximizing the lower bound L ( γ , Φ ; ξ , Ω ) with respect to γ and Φ is equivalent to minimizing the Kullback–Leibler (KL) divergence between the variational posterior probability. By factorizing the variational distributions, we can describe the lower bound as follows:
L ( γ , Φ ; ξ , Ω ) = E q [ log p ( m | ξ ) ] + E q [ log p ( c | m ) ] + E q [ log p ( w | c , Ω ) ] E q [ log q ( m ) ] E q [ log q ( c ) ]
After that, we can extend Equation (A7) in terms of the model parameters ( ξ , Ω ) and variational parameters ( γ , Φ ) (A13).
In order to find ϕ n l , we proceed to maximize with the respect to ϕ n l , so we have the following equations:
L [ m n l ] = m n l ( Ψ ( γ l ) Ψ ( γ l + Φ ) ) + m n l log Ω w ( l v ) m n l log m n l + λ n ( l l = 1 d + 1 m n ( l l ) 1 )
and therefore, we have
L ϕ n l = ( Ψ ( γ l ) Ψ ( γ l + Φ ) ) + log Ω l v log ϕ n l 1 + λ n
Setting the above equation to zero leads to
m n l = Ω l v e ( λ n 1 ) e ( Ψ ( γ l ) Ψ ( γ l + Φ ) )
Next, we maximize Equation (A13) with respect to γ i . The terms containing γ i are
L [ ξ q ] = l = 1 d [ a l ( Ψ ( γ l ) Ψ ( γ l + Φ ) ) + ( Ψ ( γ l ) Ψ ( γ l + Φ ) ) ( b l a l + 1 b l + 1 ) ] + n = 1 N m n l ( Ψ ( γ l ) Ψ ( γ l + Φ ) + n = 1 N m n ( d + 1 ) ( Ψ ( γ d ) Ψ ( γ d + Φ d ) ) l = 1 d ( log Γ ( γ l + Φ l ) log Γ ( γ l ) log Γ ( Φ l ) ) + l = 1 d ( Ψ ( γ l ) γ l ( Ψ ( γ l + Φ l ) ) + ( Ψ ( Φ ) Ψ ( Φ + γ l ) ) ( Φ γ l + 1 Φ l + 1 ) ) )
Setting the derivative of the above equation to zero leads to the following updated parameters:
γ l = a l + n = 1 N m n l
Φ l = b l + n = 1 N l = l + 1 d + 1 m n ( l )
The challenge of deriving empirical Bayes estimates for the model parameters ξ and Ω is tackled by utilizing the variational lower bound as a substitute for the marginal log probability, using variational parameters γ and Φ . The empirical Bayes estimates are then determined by maximizing this lower bound in relation to the model parameters. Until now, our discussion has centered on the log probability for a single document; the overall variational lower bound is computed as the sum of the individual lower bounds from each document. In the M-step, this bound is maximized with respect to model parameters ξ and Ω . Consequently, the entire process is akin to performing a coordinate ascent as outlined in Equation (31). We formulate the update equation for estimating Ω by isolating terms and incorporating Lagrange multipliers to maximize the bound with respect to Ω :
L [ Ω ] = d = 1 M n = 1 N s l = 1 K + 1 j = 1 V m d n l w d n j log Ω ( l j ) + l = 1 K + 1 λ l ( j = 1 V Ω w ( i j ) )
To derive the update equation for Ω ( l j ) , we take the derivative of the variational lower bound with respect to Ω ( l j ) and set this derivative to zero. This step ensures that we find the point where the lower bound is maximized with respect to the parameter Ω ( l j ) .
Ω ( l j ) d = 1 M n = 1 N d m d n l w d n j
The updates mentioned lead to convergence at a local maximum of the lower bound of log p ( Ω , ξ | r ) , which is optimal for all product approximations of the form q ( m ) q ( w ) for the joint probability p ( m , w | Ω , ξ , r ) . This approach ensures that the variational parameters are adjusted to optimally approximate the true posterior distributions within the constraints of the model.
Φ l = Γ ( a i + b i ) Γ ( a i ) Γ ( b i ) m n l ( Ψ ( γ l ) Ψ ( γ l + Φ ) )
γ l = a l + n = 1 N m n l
Ω ( l j ) = Γ ( a i + b i ) Γ ( a i ) Γ ( b i ) ( 2 f j d = 1 M n = 1 N d m d n l w d n j )

Collapsed Gibbs Sampling Method

Utilizing the fundamental procedure of the GD distribution as delineated in the all-encompassing generative formula p ( c , z , θ , φ , w | , Ω , ξ , μ ) within our innovative methodology, we can express it in the following manner:
p ( c , z , θ , φ , w | , Ω , ξ , μ ) = p ( w | μ ) p ( θ | Ω ) p ( φ | ξ ) × n = 1 N p ( z n | θ ) p ( x n | z n n , φ )
Here, p ( θ | Ω ) signifies the GD document prior distribution, where Ω = ( a 1 , b 1 , , a n , b n ) serves as a hyperparameter. Simultaneously, p ( φ | ξ ) , with ξ = ( α 1 , β 1 , , α d , β d ) as its hyperparameters, represents the GD corpus prior distribution. The process of Bayesian inference seeks to approximate the posterior distribution of hidden variables z by integrating out parameters, which can be mathematically depicted as follows:
p ( c , z | w , Ω , ξ ) = W θ φ p ( c , z , θ , φ , | Ω , ξ ) d φ d θ
Crucially, the joint distribution is expressed as a product of Gamma functions, as highlighted in prior research [12,45,46]. This expression facilitates the determination of the expectation value for the accurate posterior distribution:
p ( z i j = k | c , w , Ω , ξ ) = E p ( z i j | w , c , Ω , ξ ) [ p ( z i j = k | z i j , c , w , Ω , ξ ) ]
Employing the GD prior results in the posterior calculation as outlined below:
p ( z i j = k | z i j , c , w , Ω , ξ ) ( N j k i j + α w k ) ( β w k + l = k + 1 K + 1 N j l i j ) ( α w k β w k + l = k + 1 K + 1 N j l i j ) × ( N k v i j i j + a v ) ( b v + d = v V + 1 N k d i j i j ) a v + b v + d = v V + 1 N k d i j i j ) = A ( K )
This leads to a posterior probability normalization as follows:
p ( z i j = k | z i j , x , Ω , ξ ) = A ( k ) k = 1 K A ( k )
The sequence from Equation (38) to Equation (40) delineates the complete collapsed Gibbs sampling procedure, encapsulated as follows:
p ( z i j = k | c , w , Ω , ξ ) = E p ( z i j | w , c , Ω , ξ ) A ( k ) k = 1 K A ( k )
The implementation of collapsed Gibbs sampling in our GD-centric model facilitates sampling directly from the actual posterior distribution p, as indicated in Equation (41). This sampling technique is deemed more accurate than those employed in variational inference models, which typically approximate the distribution from which samples are drawn [46,47]. Hence, our model’s precision is ostensibly superior.
Upon the completion of the sampling phase, parameter estimation is conducted using the methodologies discussed.

3.2. Beta-Liouville Multinomial PCA

For the Beta-Liouville Multinomial PCA (BLMPCA) model, we define a corpus as a collection of documents with the same assumption described in the GDMPCA section. Hence, we have the following procedure for the model for every single word of the document. The BLMPCA model proceeds with generating every single word of the document with the following steps, where c is a d + 1 -dimensional binary vector of topics defined:
m B L ( Υ ) z M u l t i n o m i a l ( m , L ) , w k M u l t i n o m i a l ( Ω k , c k )
In the model described, each topic is represented by a binary variable, where z i n = 1 indicates that the i-th topic is chosen for the n-th word, and z i n = 0 indicates it is not chosen. The vector z n is a ( D + 1 ) -dimensional binary vector representing the topic assignments across all D + 1 topics for a given word. The vector m is defined as m = ( m 1 , m 2 , , m D + 1 ) , where m D + 1 = 1 i = 1 D m i captures the distribution of topic proportions across the document, ensuring that the sum of proportions across all topics equals 1.
A chosen topic is associated with a multinomial prior w over the vocabulary, where Ω w i j = p ( w j = 1 | z i = 1 ) describes the probability of the j-th word being selected given that the i-th topic is chosen. This formulation allows for each word in the document to be drawn randomly from the vocabulary conditioned on the assigned topic.
The probability p ( w n | z n , Ω w ) is a multinomial probability that conditions on the topic assignments z n and the topic-word distributions Ω w , effectively modeling the likelihood of each word in the document given the topic assignments.
Additionally, B L ( Υ ) represents a d-variate Beta-Liouville distribution with parameters Υ = ( α 1 , . . . , α D , α , β ) . The probability distribution function of this Beta-Liouville distribution encapsulates the prior beliefs about the distribution of topics across documents, accommodating complex dependencies among topics and allowing for flexibility in modeling topic prevalence and co-occurrence within the corpus.
P ( θ 1 , , θ D | Υ ) = Γ ( d = 1 D α d ) Γ ( α + β ) Γ ( α ) Γ ( β ) d = 1 D θ d α d 1 Γ ( α d ) × d = 1 D θ d α l = 1 D α l × 1 l = 1 D θ l β 1
A Dirichlet distribution is the special case of BL if β d = α d + 1 + β d + 1 [42,45]. The mean, the variance, and the covariance in the case of a BL distribution are as follows [45]:
E ( θ d ) = α α + β α d d = 1 D α d
v a r ( θ d ) = ( α α + β ) 2 α d ( α d + 1 ) ( m = 1 D α m ) ( m = 1 D α m + 1 ) E ( θ d ) 2 α d 2 ( m = 1 D α m ) 2
and the covariance between θ l and θ k is given by
C o v ( θ l , θ k ) = α l α k d = 1 D α d ( α + 1 ) ( α ) ( α + β + 1 ) ( α + β ) d = 1 D α d + 1 α α + β d = 1 D α d
The earlier equation illustrates that the covariance matrix of the Beta-Liouville distribution offers a broader scope compared to the covariance matrix of the Dirichlet distribution. For the parameter estimation of BLMPCA, first, the parameter Ω is estimated using the Beta-Liouville prior on m using parameter Υ [19]. The likelihood model for the BLMPCA is given as follows:
p ( m , w | Υ , Ω ) = Γ ( α ) Γ ( β ) Γ ( d = 1 D α d ) Γ ( α + β ) z w 1 , 1 , w 1 , 2 , w k , 1 , w 1 , J w K , J L [ 1 Γ ( α d ) m k α d 1 + k m k α d α d + ( 1 k m k ) β 1 ] k , j m k w k , j Ω k , j w k , j
For the Beta-Liouville priors, we have the following:
m B L ( Υ ) Ω k B L ( 2 f )
In the following step, we will estimate the parameters for Ω using the Beta-Liouville prior and the Hessian matrix (Appendix C). As we explained in the previous Section 3.1, we should estimate the model parameters ( Υ , Ω ) and the variational parameters ( γ , Φ ) according to Equations (21), (22) and (A7) to find m n l , and we proceed to maximize with respect to m n l ; so, we have the following equations:
L ( γ , Φ ; Υ , Ω ) = log ( Γ ( d = 1 D α d ) ) + log ( Γ ( α + β ) ) log ( Γ ( α ) ) log ( Γ ( β ) ) d = 1 D log Γ ( α d ) + d = 1 D α d ( Ψ ( γ d ) Ψ ( l = 1 D γ l ) + α ( Ψ ( α γ ) Ψ ( α γ + β γ ) ) + β ( Ψ ( β γ ) Ψ ( α γ + β γ ) ) + β ( Ψ ( β γ ) Ψ ( α γ + β γ ) ) + n = 1 N d = 1 D m n d ( Ψ ( γ d ) Ψ ( l = 1 D γ l ) + Ψ ( α γ ) Ψ ( α γ + β γ ) ) + n = 1 N m n ( D + 1 ) ( Ψ ( β γ ) Ψ ( α γ + β γ ) ) + n = 1 N l = 1 D + 1 j = 1 V m n l w n j log ( Ω l j ) ( log ( Γ ( l = 1 D α l ) ) + log ( Γ ( α + β ) ) log Γ ( α ) log Γ ( β ) i = 1 D log Γ ( α i ) + i = 1 D α i ( Ψ ( γ m i ) Ψ ( l = 1 D γ m ( l ) ) ) + α ( Ψ ( α m γ ) Ψ ( α m γ β m γ ) ) + β ( Ψ ( β m γ ) Ψ ( α m γ + β m γ ) ) ) ( n = 1 N l = 1 D + 1 m n l log ( m n l ) )
To find m n l , we proceed to maximize with respect to ϕ n l :
L [ m n l ] = m n l ( Ψ ( γ i ) Ψ ( l = 1 D γ l ) ) + m n l log β w ( i v ) m n l log ( m n l ) + λ n ( l = 1 D m n l 1 )
Therefore, we have
L ϕ n l = ( Ψ ( γ d ) Ψ ( l = 1 D γ l ) ) + log β w ( i v ) log ϕ n l 1 + λ n
The next step is to optimize Equation (49) to find the update equations for the variational; we separate the terms containing the variational Beta-Liouville parameters once more.
L [ ξ q ] = α d ( Ψ ( γ d ) ) Ψ ( l = 1 D γ l ) + α ( Ψ ( α γ ) Ψ ( α γ + β γ ) ) + β ( Ψ ( α γ ) Ψ ( α γ + β γ ) ) + n = 1 N ϕ n ( Ψ ( γ l ) Ψ ( l = 1 D γ l ) + Ψ ( α γ ) Ψ ( α γ + β γ ) ) + n = 1 N ϕ n ( D + 1 ) ( Ψ ( β γ ) Ψ ( α γ + β γ ) ) ( log ( Γ ( l = 1 D γ l ) ) + log ( γ ( α γ + β γ ) log ( Γ ( α γ ) ) log ( Γ ( β γ ) ) log ( Γ ( γ l ) ) + γ l ( Ψ ( γ l ) + Ψ ( α γ ) Ψ ( α γ + β γ ) ) Ψ ( l = 1 D γ l ) + α γ ( Ψ ( α γ ) Ψ ( α γ + β γ ) ) + β γ ( Ψ ( β γ ) Ψ ( α γ + β γ ) ) )
Selecting the terms containing variational Beta-Liouville variables γ i , α γ , and β γ , we have
L ( γ i ) = α i ( Ψ ( γ i ) ) ( l = 1 D α l ) ( Ψ ( l = 1 D γ l ) ) + n = 1 N ϕ n i ( Ψ ( γ i ) Ψ ( l = 1 D γ l ) ) ( log Γ ( l = 1 D ) log Γ ( γ i ) + γ i ( Ψ ( l = 1 D γ l ) d = 1 D γ d )
and
L [ α γ ] = α ( Ψ ( α γ ) Ψ ( α γ + β γ ) ) + β ( Ψ ( α γ + β γ ) ) + ( Ψ ( α γ ) Ψ ( α γ + β γ ) ) n = 1 N i = 1 D ϕ n i n = 1 N ϕ n ( D + 1 ) ( Ψ ( α γ + β γ ) ) ( log ( α γ + β γ ) log ( Γ ( α γ ) ) + α γ ( Ψ ( α γ ) Ψ ( α γ + β γ ) ) + β γ ( Ψ ( α γ + β γ ) ) )
Setting Equations (52)–(54) to zero, we have the following update parameters:
γ i = α + n = 1 N ϕ n i
α γ = α + n = 1 N d = 1 D ϕ n d
β γ = β + n = 1 N ϕ n ( D + 1 )
We address the challenge of deriving empirical Bayes estimates for the model parameters Υ and Ω by utilizing the variational lower bound as a substitute for the marginal log likelihood. This approach fixes the variational parameters γ and Φ at values determined through variational inference. We then optimize this lower bound to obtain the empirical Bayes estimates of the model parameters.
To estimate Ω w , we formulate necessary update equations. The process of maximizing Equation (52) with respect to Ω results in the following equation:
L [ Ω w ] = d = 1 M n = 1 N s l = 1 D + 1 j = 1 V ϕ d n l w d n j log ( Ω w ( l j ) ) + l = 1 D + 1 λ l ( j = 1 V Ω w ( l j ) 1 )
Taking the derivatives with the respect to β w ( l j ) and setting it to zero yields in Appendix C.1:
Ω w ( l j ) d = 1 M n = 1 N d m d n l w d n j

Beta-Liouville Parameter

The objective of this subsection is to determine the estimates of the model’s parameters using variational inference techniques [48].
L [ ξ ] = m = 1 M ( log ( Γ ( l = 1 D α l ) ) + log ( Γ ( α + β ) ) log Γ ( α ) log Γ ( β ) i = 1 D log Γ ( α i ) + i = 1 D α i ( Ψ ( γ m i ) Ψ ( l = 1 D γ m ( l ) ) ) + α ( Ψ ( α m γ ) Ψ ( α m γ β m γ ) ) + β ( Ψ ( β m γ ) Ψ ( α m γ + β m γ ) ) )
The derivative of the above equation with respect to the BL parameter is given by
L [ ξ ] α l = M ( Ψ ( l = 1 D ) Ψ ( α l ) ) + m = 1 M ( Ψ ( γ m l ) Ψ ( l = 1 D γ m ( l ) ) ) L [ ξ ] α = M [ Ψ ( α + β ) Ψ ( α ) ] + m = 1 M ( Ψ ( α m γ ) Ψ ( α m γ + β m γ ) ) L [ ξ ] β = M [ Ψ ( α + β ) Ψ ( β ) ] + m = 1 M ( Ψ ( β m γ ) Ψ ( α m γ + β m γ ) )
From the equations presented earlier, it is evident that the derivative in Equation (52) with respect to each of the BL parameters is influenced not only by their individual values but also by their interactions with one another. Consequently, we utilize the Newton–Raphson method to address this optimization problem. To implement the Newton–Raphson method effectively, it is essential to first calculate the Hessian matrix for the parameter space, as illustrated below [49]:
2 L [ ξ ] α l α j = M ( δ ( i , j ) Ψ ( α i ) + Ψ ( l = 1 D α l ) ) 2 L [ ξ ] α 2 = M ( Ψ ( α + β ) Ψ ( α ) ) 2 L [ ξ ] α β = M Ψ ( α + β ) 2 L [ ξ ] β 2 = M ( Ψ ( α + β ) Ψ ( β ) )
The Hessian matrix shown above is very similar to the Hessian matrix of the Dirichlet parameters in the MPCA model and generalized Dirichlet parameters in GDMPCA. In fact, the above matrix can be divided into two completely separate matrices using parameters α d , α , and β . Each of the two parts’ parameter derivation will be identical to the Newton–Raphson model provided by MPCA and GDMPCA.

3.3. Inference via Collapsed Gibbs Sampling

The collapsed Gibbs sampler (CGS) contributes to the inference by estimating posterior distributions through a Bayesian network of conditional probabilities, which are determined through a sampling process of hidden variables. Compared to the traditional Gibbs sampler that functions in the combined space of latent variables and model parameters, the CGS offers significantly faster estimation times. The CGS operates within the collapsed space of latent variables, where, in the joint distribution p ( X , z , θ , ϕ , w | Ω , Υ , μ ) , the model parameters θ and ϕ are marginalized out. This marginalization leads to the marginal joint distribution p ( X , z , w | Ω , Υ , μ ) , which is defined as follows:
p ( x , z , w | Ω , Υ ) = W θ φ p ( X , z , θ , φ , w | Ω , ξ ) d φ d θ
In using Equation (63), the method calculates the conditional probabilities of the latent variables z i j by considering the current state of all other variables while excluding the specific variable z i j itself [50]. Meanwhile, the collapsed Gibbs sampler (CGS) determines the topic assignments for the observed words by employing the conditional probability of the latent variables, where “ i j ” indicates counts or variables with z i j excluded [50]. This specific conditional probability is defined as follows [51]:
p ( z i j = k | z i j , X , w , Ω , Υ ) = p ( z i j , z i j , X , w | Ω , Υ ) p ( z i j , X , w | Ω , Υ )
The sampling mechanism of the collapsed Gibbs approach can be summarized as an expectation problem:
p ( z i j = k | X , w , Ω , Υ ) = E p ( z i j | w , X , Ω , Υ ) [ p ( z i j = k | z i j , X , w , Ω , Υ ) ]
The collapsed Gibbs sampling Beta-Liouville multinomial procedure consists of two phases for assigning documents to clusters. First, each document is assigned a random cluster for initialization. After that, each document is assigned a cluster based on the Beta-Liouville distribution after a specified number of iterations.
The goal is to use a network of conditional probabilities for individual classes to sample the latent variables from the joint distribution p ( X , z | w , Ω , Υ ) . The assumption of conjugacy allows the integral in Equation (63) to be estimated.
p ( X , z | w , υ ) = C j = 1 M Γ ( i = 1 k α i ) Γ ( α + β ) i = 1 k Γ ( α i ) Γ ( α ) Γ ( β ) × i = 1 k Γ ( α i ) Γ ( α ) Γ ( β ) Γ ( α + β ) Γ ( i = 1 K α i )
The likelihood of the multinomial distribution, defined by the parameter Υ , and the probability density function of the Beta-Liouville distribution can be expressed as follows:
p ( X | Υ ) = p ( X | θ ) p ( θ | α 1 , , β , α ) d θ = k = 1 k θ k m k Γ ( k = 1 k α k ) Γ ( α + β ) Γ ( α ) Γ ( β ) k = 1 K θ k α k 1 Γ ( α k ) × ( k = 1 K θ k ) α α k ( 1 k = 1 K θ k ) β 1 d θ
By integrating the probability density function of the Beta-Liouville distribution over the parameter θ and incorporating updated parameters derived from the remaining integral in Equation (69), we are able to express it as a fraction of Gamma functions. The following shows the updated parameters, where N j k represents counts corresponding to variables [45,51]:
α K = α k + j = 1 k N j k α = α + N j k β = β + N j k
Equation (67) is then equivalent to
p ( k | α 1 , , α k , β , α ) = Γ ( k = 1 K α k ) Γ ( α + β ) Γ ( α + k = 1 k 1 m k ) Γ ( β + m k ) Γ ( α ) Γ ( β ) k = 1 K Γ ( α k ) Γ ( k = 1 K ( α k + m k ) ) k = 1 K Γ ( α k + m k ) Γ ( α + k = 1 K 1 m k + β + m k )
The parameters α 1 , , α k , α , and β correspond to the Beta-Liouville distribution, while m k represents the number of documents in cluster k.
After the sampling process, parameter estimation is performed. Subsequently, the empirical likelihood method [47] is utilized to validate the results using a held-out dataset. Ultimately, this process leads to the estimation of the class conditional probability p ( X | w , Ω , Υ ) within the framework of collapsed Gibbs sampling:
p ( X | w , Ω , Υ ) = i j k = 1 K 1 S s = 1 S θ ˜ j k s φ ˜ k w s
The parameters are then computed as follows:
θ ˜ j k s = ( N j k + α k ) ( α j k + l = k + 1 K + 1 N j l ) ( N j k + β k ) ( a k b k + l = k + 1 K + 1 N j l ) ( α j + l = k + 1 K + 1 N j l )
φ ˜ k w s = ( N j k + α w ) ( α j w + l = k + 1 K + 1 N j l ) ( N j k + β w ) ( α w b w + l = k + 1 K + 1 N j l ) ( α w j + l = k + 1 K + 1 N j l )
where S is the size of a sample.

4. Experimental Results

In this section, we validate our proposed algorithms’ efficiency for two distinct and challenging applications, namely, topic modeling for medical text and sentiment analysis. Each model’s evaluation is based on the success rate for each dataset and the perplexity [3,9,52,53], which is a common measure used in language modeling and is defined as:
p r e p ( D t e s t ) = exp ln p ( D t e s t ) d | w d |
where | w d | is the length of document d. A lower perplexity score indicates better generalization performance. In addition to the perplexity metric, the success rate is employed as a key performance indicator to evaluate our models, reflecting the proportion of correctly identified topics within a corpus in topic modeling. The success rate serves as a straightforward measure of a model’s efficacy, capturing its ability to accurately classify documents into the correct topical categories, which is essential for effective information retrieval and knowledge discovery in the domain of text analysis. The main goal of both applications is to compare the GDMPCA, BLMPCA, and MPCA performances. The choice of these datasets is pivotal to our research as they offer a broad spectrum of analytical scenarios, from topic modeling for medical text to sentiment analysis, thus enabling a thorough investigation into the models’ adaptability and accuracy. By encompassing datasets with distinct characteristics, we are able to demonstrate the strengths of our proposed models in varied contexts, highlighting their potential as a versatile tool in the field of text analysis.

4.1. Topic Modeling

The goal of text classification is to assign documents to predefined subject categories, a problem extensively researched with various approaches [42,54,55]. Topic modeling, a common application in natural language processing, is used for analyzing texts from diverse sources and for document clustering [56]. It identifies key “topics” in a text corpus using unsupervised statistical methods, where topics are keyword mixtures with a probability distribution, and documents are composed of topic mixtures [12]. The “CMU Book Summary Dataset” was used to validate the model performance, containing plot summaries and metadata for 16,559 books [57]. The models’ accuracy was tested by training on various document numbers and observing the impact of latent topics on the classification accuracy. Using variational Bayes inference, the models showed similar performances, but BLMPCA excelled, particularly in classifying similar classes.
In Table 2, Table 3 and Table 4, we present the first three topics, the perplexity measurements, and time complexity for all models compared in this study. The success rates obtained using GDMPCA, BLMPCA, and MPCA are depicted in Figure 1. These examples demonstrate that our proposed models, which incorporate Generalized Dirichlet and Beta-Liouville distributions, yield more accurate classifications in scenarios where distinct classes exhibit similarities, in contrast to the traditional MPCA which is a Dirichlet-based model. Additionally, in Table 5 and Table 6, we show the results for the collapsed Gibbs sampling.

4.2. Topic Modeling for Medical Text

Topic modeling plays a crucial role in navigating the complexities of health and medical text mining, despite the inherent challenges of data volume and redundancy in this domain. The study by Onan et al. [58] marked a significant advancement, presenting an optimized topic modeling approach that utilizes ensemble pruning. This method significantly improves the categorization of biomedical texts by enhancing precision and managing the computational challenges posed by the extensive data typical of medical documents. With vast amounts of health-related data, specialists struggle to find pertinent information, exemplified by the millions of papers on PubMed and hospital discharge records in the United States in 2015. This study utilized the TMVAr corpus from PubMed and the TMVAr-Dataset containing health-related Twitter news to evaluate models [59,60,61,62,63,64].

TMVAr Dataset

The TMVar Corpus dataset, comprising 500 PubMed papers with manual annotations of various mutation mentions, was utilized to evaluate our models. Table 7 and Table 8 elucidate the perplexity comparison and time complexity for the TMVAR dataset, offering insight into the performances of our proposed methods. Moreover, Table 9 and Table 10 present the outcomes of the collapsed Gibbs sampling. As indicated in the tables, the time complexity of this method is higher, yet the perplexity is lower.
Furthermore, as shown in Table 11, the BLMPCA model successfully extracts pertinent topics, which is indicative of the model’s nuanced analytical capabilities. Figure 2 further illustrates the success rate of our proposed models in comparison to the traditional MPCA, highlighting the enhanced classification accuracy achieved by our methods.

4.3. Sentiment Analysis

Sentiment analysis, crucial for interpreting emotions in texts from various sources, benefits from advanced methodologies beyond mere word analysis [65]. Recent studies, such as [66,67], have demonstrated the effectiveness of deep learning and text mining in capturing nuanced sentiment expressions. Additionally, the authors of [68] highlighted the potential of ensemble classifiers in improving the sentiment classification accuracy. These innovations showcase the shift toward more complex analyses that consider semantics, context, and intensity for a more accurate sentiment understanding.
The “Multi-Domain Sentiment Dataset”, containing Amazon.com product reviews across various domains, was used for analysis [69]. This dataset, with extensive reviews on books and DVDs, provides data for basic analysis. The applied model, using K = 8 topics, assumed that each topic comprises a bag of words with specific probabilities, and each document is a mix of these topics. The model’s goal was to learn the distributions of words and topics in the corpus.
We demonstrated that the overall sentiment of the dataset tends to be positive, influenced by the presence of high-frequency words with positive connotations within the corpus. This observation is substantiated by the sentiment analysis framework we employed. Table 12 and Table 13 provides a detailed explanation of the perplexity measures and time complexity tested for sentiment analysis. Furthermore, the findings from the topic modeling of eight emotions and two sentiments are displayed in Table 14 and Table 15. Figure 3 shows that our proposed models outperform the previous model. Figure 3 shows the success rates for MPCA, GDMPCA, and BLMPCA on sentiment analysis, with GDMPCA and BLMPCA outperforming MPCA as the number of emotions analyzed increases. This indicates their better suitability for complex emotion detection tasks in practical applications.
Furthermore, Table 16 and Table 17 present the results for the collapsed Gibbs sampling. Additionally, Table 18 and Table 19 display the accuracy and recall of various classifiers utilized for emotion detection. Table 20 shows the F1-scores for various classifiers, indicating the balanced harmonic mean of the precision and recall for SVM, Naive Bayes, and MLP classifiers when applied with MPCA, GDMPCA, and BLMPCA models in sentiment analysis.

5. Discussion

We delved into the comparative advantages of the GDMPCA and BLMPCA models over existing methods in text classification and sentiment analysis. The superior performances of our proposed models can be attributed to several key factors. Firstly, the incorporation of Generalized Dirichlet and Beta-Liouville distributions allows for a more nuanced modeling of text data, which captures the intricacies of word distributions more effectively than traditional methods. This results in a more accurate representation of the underlying thematic structures in the data. For instance, in the CMU Book Summary Dataset, the intricacies of literary themes were better represented, showcasing the models’ aptitude for multifaceted textual analysis. This was attributed to the models’ ability to account for the co-occurrence and complex interrelationships of terms within documents, a feature less emphasized in MPCA due to its assumption of component independence.
In the TMVAR Corpus from PubMed, the medical text presented a challenge due to its specialized lexicon and the density of information. The BLMPCA model excelled by exploiting its additional parameters, optimizing data representation in this high-dimensional space, thus underscoring the importance of model selection aligned with dataset characteristics. The sentiment analysis on the Multi-Domain Sentiment Dataset further emphasized the adaptability of our models. Here, BLMPCA demonstrated its finesse in discerning subtle sentiments from Amazon.com reviews, outperforming traditional approaches that may not have captured the emotional granularity present in user-generated content.
However, the sophistication of GDMPCA and BLMPCA comes with greater computational demands, as reflected in longer convergence times. This trade-off between accuracy and computational efficiency underscores the necessity of careful model selection in practice, considering the scale of the data and available computational resources. Although our proposed models signify a leap forward in text analysis methodologies, they are not without limitations. The reliance on variational inference and assumptions specific to the models may not be universally applicable to all types of textual data, suggesting room for future refinement and the exploration of alternative distributions or learning strategies.
The findings of this study illuminate the potential of integrating advanced probabilistic distributions into PCA to uncover deeper insights within text data. It is a testament to the evolution of statistical models in text analysis, pointing toward an exciting trajectory for future research in the field. The ongoing dialogue within the academic community on these topics is reflective of the dynamic nature of machine learning and its applications to natural language processing. As we continue to push the boundaries, it is imperative to balance innovation with practicality, ensuring that our models are not only theoretically robust but also computationally viable and accessible for varied applications.

6. Conclusions

In this paper, two novel models, generalized Dirichlet Multinomial PCA and Beta-Liouville Multinomial PCA, were proposed to improve the accuracy of the MPCA model for multi-topic modeling and text classification. We followed a Bayesian analysis that considers the generalized Dirichlet and Beta-Liouville assumptions. We demonstrated that our two proposed models have more flexibility. The models were used in two separate applications: text classification and sentiment analysis. The results show that the two proposed models, in all applications, achieved superior performances, as represented by the high prediction accuracy in comparison to that of the MPCA. It could be claimed that the proposed models, using different prior assumptions, yield better results than the standard methods. Specially, the BLMPCA provides the best improvement compared to the GDMPCA and MPCA for all the tested data. Crucially, the employment of collapsed Gibbs sampling for parameter inference was proven efficient and effective, despite its time-consuming nature. This method substantially boosts our models’ computational capabilities, allowing for the superior discovery of latent topics in text corpora and marking a noteworthy advancement over the MPCA model. Future approaches for research will concentrate on model modifications and improvements to achieve greater precision in topic modeling. In addition, future works could be devoted to extending the proposed models to other applications and significantly extending the proposed model to fit a variety of data as well as real-time streaming data.

Author Contributions

P.K. was involved in the model’s design methodology and implementation and writing the initial draft. E.I.K. helped with critical review, commentary, and revision. N.B. helped in critical review and revision, as well as with oversight and leadership responsibilities for the research activity planning. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

Acknowledgments

We would like to express our gratitude for the invaluable academic support received during the course of this research. Notably, we extend our thanks to the work of Shojaee Bakhtiari, Ali, whose dissertation “Count Data Modeling and Classification Using Statistical Hierarchical Approaches and Multi-topic Models”, completed at Concordia University in 2014, provided foundational insights and methodologies that significantly guided our analysis and conclusions.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
GDGeneralized Dirichlet;
BLBeta-Liouville;
PCAPrincipal Component Analysis;
LDALatent Dirichlet Allocation;
MPCAMultinomial Principal Component Analysis;
PLSAProbabilistic Latent Semantic Analysis;
pLSIProbabilistic Latent Semantic Indexing;
SVDSingular Value Decomposition;
NMFNon-negative Matrix Factorization;
EMExpectation-Maximization;
CGSCollapsed Gibbs Sampling;
MCMCMarkov Chain Monte Carlo;
BLMPCABeta-Liouville Multinomial Principal Component Analysis;
GDMPCAGeneralized Dirichlet Multinomial Principal Component Analysis;
SVMSupport Vector Machine;
NLPNatural Language Processing.

Appendix A. Exponential Family Distribution

The following introduces the general exponential family of distributions:
We have a vector of T functions t ( x ) and d parameters θ for each individual sample point, which is a vector of measurements x, both of dimension T and likely subject to some additional constraints. The following is the likelihood q ( x | θ ) [70]:
q ( x | θ ) = 1 Y t ( x ) Z t ( θ ) e x p ( t ( x ) θ )
Z t ( θ ) is modified to Z, or a distinguishing subscript is inserted. When y is distributed as q ( y | ϕ ) , the notation E q ( y | ϕ ) is used to describe the expected value of quantity A. There are two main concepts that must be given [71]:
μ t E q ( y | ϕ ) { t ( x ) } = log Z t θ Σ t E q ( y | ϕ ) { ( t ( x ) μ t ) ( t ( x ) μ t ) ) } = 2 log Z t θ θ = μ t θ
The mean vector μ t shares the same dimensionality as θ , and the matrix Σ t encapsulates the covariance of t ( x ) , as noted in [20]. Notably, μ t serves as a counterpart to the parameter set θ . Specifically, when μ t is fully ranked, it functions as the Hessian for changes in the basis. Moreover, μ t represents the expected Fisher Information of the distribution. Both t and Σ t can be directly derived from Z t , indicating a unique relationship where μ t acts as a complementary parameter set to θ . In situations where μ t possesses maximum rank, it is instrumental in basis transformations and also signifies the intended Fisher Information for the distribution.
We further detail the characteristics of the exponential family for the Dirichlet, generalized Dirichlet, and Beta-Liouville distributions in Table A1. Another essential feature of the exponential family is the computation of the maximum a posteriori (MAP) estimates for parameters, derived from a dataset consisting of I data points. This setup often reflects the structure of a conjugate prior, facilitating the estimation process. One common approach involves the use of an “effective” prior sample size, characterized by relevant statistics ν t and a prior sample size of S t . This special method for calculating MAP for parameters within the exponential family provides an approximation for their dual aspects, as explored in [20].
μ t ^ = ν t + Σ i t ( x i ) S t + I
Table A1. Exponential family characterizations for Dirichlet, GD and BL distributions.
Table A1. Exponential family characterizations for Dirichlet, GD and BL distributions.
MODEL Z t t k ( x ) θ k μ t , k
Dirichlet k Γ ( α k ) Γ ( k α k ) log ( x 1 ) , , log ( x k + 1 ) α k Ψ 0 ( α k ) Ψ ( k ( α k ) )
GD Γ ( a i ) Γ ( b i ) Γ ( a i + b i ) log ( x 1 ) , , log ( 1 t = 1 D x t ) log ( 1 t = 1 D 1 x t ) a k , b k Ψ ( a i ) Ψ ( a i + b i ) + m = 1 i 1 ( Ψ ( b m ) Ψ ( a m + b m )
BL Γ ( d = 1 D α d ) Γ ( α + β ) Γ ( α ) Γ ( β ) log ( x 1 ) log ( d = 1 D x d ) log x D log ( d = 1 D x d ) α k , α , β Ψ ( α ) Ψ ( α + β ) + Ψ ( α d ) Ψ ( d α d )

Appendix A.1. The Generalized Dirichlet Distribution Exponential Form

Since the GD distribution belongs to the exponential family of distributions, it can be represented in general as follows:
P ( θ | ξ ) = Z t ( θ ) × e x p [ l = 1 2 d G l ( θ ) T l ( θ ) ]
where
Z t ( θ ) = l = 1 d Γ ( α l + β l ) Γ ( α l ) + Γ ( β l ) G l ( ξ ) = α l , ( l : 1 , , d ) G l ( ξ ) = β l d α l d + 1 β l d + 1 , ( l : d + 1 , 2 d 1 ) G l ( ξ ) = β l ( l : 2 d ) T l ( θ ) = log ( θ l ) , ( l : 1 , d ) T l ( θ ) = log ( 1 t = 1 d 1 θ t ) , ( l : d + 1 , 2 d )
In the formulation provided, Z ( θ ) represents the normalization factor, G ( θ ) is the natural parameter, and T ( θ ) denotes the sufficient statistics of the distribution. It is established that within the framework of the exponential family of distributions, the derivative of the logarithm of the normalization factor Z ( θ ) with respect to the natural parameters G ( θ ) is equal to the expected value of the sufficient statistics T ( θ ) . This relationship underscores the fundamental connection between these components in statistical modeling within the exponential family. Therefore, we have
E [ log ( θ l ) ] = ψ ( α l + β l ) ψ ( α l ) ψ ( β l ) , l = 1 , , d E [ log ( 1 t = 1 l θ t ) ] = ψ ( β l ) ψ ( α l + β l ) , l = 1 , , d

Appendix B. Parameters for GDMPCA

In breaking down the L parameter for GDMPCA, we have the following.
By factorizing log p ( w | ξ , Ω ) E q [ ( θ , z , w ) | ξ , Ω ] E q [ log q ( z , θ ) ] , we have
L ( γ , Φ ; ξ , Ω ) = E q [ log p ( m | ξ ) ] + E q [ log p ( c | m ) ] + E q [ log p ( w | c , Ω ) ] E q [ log q ( m ) ] E q [ log q ( c ) ]
In the following, we derive each of the five factors of the above equation:
E q [ log p ( θ | ξ ) ] = l = 1 d [ log Γ ( α l + β l ) log Γ ( α l ) log Γ ( β l ) ] + l = 1 d [ α l ( Ψ ( γ l ) Ψ ( γ l + δ l ) ) + ( Ψ ( δ l ) Ψ ( γ l + δ l ) ) ( β l α l + 1 β l + 1 ) ]
E q [ log p ( z | θ ) ] = n = 1 N l = 1 d ϕ n l ( Ψ ( γ l ) Ψ ( γ l + δ l ) ) + n = 1 N ϕ n ( d + 1 ) ( Ψ ( δ d ) Ψ ( δ d + γ d ) )
E q [ log p ( w | z , Ω ) ] = n = 1 N l = 1 d + 1 j = 1 v ϕ n l w n j log ( Ω ( l j ) )
We should mention that Ω ( l j ) = p ( w n j = 1 | z l = 1 ) :
E q [ log q ( θ ) ] = l = 1 d ( log Γ ( γ l + δ l ) log Γ ( γ l ) log Γ ( δ l ) ) + l = 1 d [ γ l ( Ψ ( γ l ) Ψ ( γ l + δ l ) ) + ( Ψ ( δ l ) Ψ ( δ l + γ l ) ) ( δ l γ l + 1 δ l + 1 ) ]
E q [ log q ( z ) ] = n = 1 N l = 1 D + 1 ϕ n l log ( ϕ n l )
Subsequently, we will elaborate on Equation (A7) by expanding it with respect to both the model parameters and the variational parameters.
L ( γ , Φ ; ξ , Ω ) = l = 1 d [ log Γ ( a l + b l ) log Γ ( a l ) log Γ ( b l ) ] + l = 1 d [ a l ( Ψ ( γ l ) Ψ ( γ l + Φ ) ) + ( Ψ ( Φ ) Ψ ( γ l + Φ ) ) ( a l a l + 1 b l + 1 ) ] + n = 1 N l = 1 d m n l ( Ψ ( γ l ) Ψ ( γ l + Φ ) ) + n = 1 N m n ( d + 1 ) ( Ψ ( Φ ) Ψ ( Φ + γ d ) ) + n = 1 N l = 1 d + 1 j = 1 v m n l w n j log ( Ω i j ) l = 1 d ( log Γ ( γ l + Φ ) log Γ ( γ l ) log Γ ( Φ ) ) l = 1 d [ γ l ( Ψ ( γ l ) Ψ ( γ l + Φ ) ) + ( Ψ ( Φ ) Ψ ( Φ + γ l ) ) ( Φ γ l + 1 Φ l + 1 ) ]

Appendix B.1. Variational Generalized Dirichlet

To derive the update equations for the variational parameters in the generalized Dirichlet model, you start by isolating the terms in Equation (A7) that contain the variational parameters of the generalized Dirichlet. This involves examining the equation to identify which parts specifically involve these parameters and then focus on manipulating these parts to derive expressions for updating the parameters during the variational inference process. This method allows for the iterative refinement of the parameters, enhancing model accuracy with respect to the data being analyzed.
L [ ξ q ] = l = 1 d [ α l ( Ψ ( γ l ) Ψ ( γ l + δ l ) ) + ( Ψ ( γ l ) Ψ ( γ l + δ l ) ) ( β l α l + 1 β l + 1 ) ] + n = 1 N ϕ n l ( Ψ ( γ l ) Ψ ( γ l + δ l ) + n = 1 N ϕ n ( d + 1 ) ( Ψ ( γ d ) Ψ ( γ d + δ d ) ) l = 1 d ( log Γ ( γ l + δ l ) log Γ ( γ l ) log Γ ( δ l ) ) + l = 1 d ( Ψ ( γ l ) γ l ( Ψ ( γ l + δ l ) ) + ( Ψ ( δ l ) Ψ ( δ l + γ l ) ) ( δ l γ l + 1 δ l + 1 ) ) )
Setting the derivative of the above equation to zero leads to the following updated parameters:
γ l = α l + n = 1 N ϕ n l
γ l = β l + n = 1 N l l = l + 1 d + 1 ϕ n ( l l )

Appendix B.1.1. Topic-Based Model

To derive the update equations for β w , maximize Equation (A7) with respect to β w . This involves setting the derivatives to zero, mirroring the optimization process used in MPCA, resulting in similar equations.
L [ β w ] = d = 1 M n = 1 N s l = 1 K + 1 j = 1 V ϕ d n l w d n j log β w ( l j ) + l = 1 K + 1 λ l ( j = 1 V β w ( i j ) )
Taking the derivative with respect to β w ( l j ) and setting it to zero yields
β w ( l j ) d = 1 M n = 1 N d ϕ d n l w d n j
In this scenario, because there are hidden variables present in the primary objective function, the situation is not fully addressed by Equations (33) and (34). However, the probability distribution q ( w | γ , r , m ) can be accurately modeled using multinomials, which ensures that the minimum Kullback–Leibler (KL) divergence reaches zero. Consequently, the iterative updates will converge towars a local extremum of the log probability log p ( Ω , m | r ) .
γ l = Γ ( a i + b i ) Γ ( a i ) Γ ( b i ) Ω m n l
m n l = Γ ( a i + b i ) Γ ( a i ) Γ ( b i ) Ω l v e ( λ n 1 ) e ( Ψ ( γ l ) Ψ ( γ l + Φ ) )
Ω i j = Γ ( a i + b i ) Γ ( a i ) Γ ( b i ) ( 2 f j + ( n e ( λ n 1 ) e ( Ψ ( γ l ) Ψ ( γ l + Φ ) ) )
e λ n 1 = 1 l = 1 d m n l e ( Ψ ( γ l ) Ψ ( γ l + Φ l ) ) + m ( d + 1 ) n e ( Ψ ( Φ d ) Ψ ( Φ d + γ d ) )

Appendix B.1.2. Generalized Dirichlet Parameter

We select the components of Equation (A7) that involve the generalized Dirichlet parameters ξ .
L [ ξ ] = m = 1 M ( log ( Γ ( α l + β l ) ) log Γ ( α l ) ) log ( Γ ( β l ) ) ) + m = 1 M ( α l ( Ψ ( γ m l Ψ ( γ m l + δ m l ) ) + β l ( Ψ ( δ m l ) Ψ ( δ m l γ m l ) ) )
Taking the derivative of the mentioned equation with respect to the generalized Dirichlet parameters yields
L [ ξ ] α l = M ( Ψ ( α l + β l ) Ψ ( α l ) ) + m = 1 M ( Ψ ( γ m l ) Ψ ( γ m l + δ m l ) )
and
L [ ξ ] β l = M ( Ψ ( α l + β l ) Ψ ( β l ) ) + m = 1 M ( Ψ ( δ m l ) Ψ ( γ m l + δ m l ) )
When applying the Newton–Raphson method to solve for the parameters, it is crucial to obtain the Hessian matrix with respect to the parameter space. The Hessian matrix of the likelihood function in this case assumes a particularly interesting form, as detailed below:
2 L [ ξ ] α l 2 = M [ Ψ ( α l + β l ) Ψ ( α l ) ]
2 L [ ξ ] β l 2 = M [ Ψ ( α l + β l ) Ψ ( β l ) ]
2 L [ ξ ] α l β l = M [ Ψ ( α l + β l ) ]
The non-diagonal entries of the Hessian matrix are zero, which imparts a block diagonal structure to the matrix. This configuration simplifies the calculation of the inverse Hessian matrix, as it reduces to inverting the matrices along the diagonal. This simplification allows for an easier derivation of the inverse.

Appendix C. Variational BLMPCA

To derive the parameter ϕ , which represents the probability that the n-th word is generated by the l-th hidden topic, we maximize the relevant function with respect to ϕ . This involves adjusting ϕ to optimize the likelihood of the observed data given the model’s assumptions about topic distributions:
L [ ϕ n l ] = ϕ n i ( Ψ ( γ i ) Ψ ( l = 1 D γ l ) ) + ϕ n i log β w ( i v ) ϕ n i log ϕ n i + λ n ( l = 1 D ϕ n ( l ) 1 )
and
L [ ϕ n ( D + 1 ) ] = ϕ n ( D + 1 ) ( Ψ ( β γ Ψ ( α γ + β γ ) ) ) + ϕ n ( D + 1 ) log β ( D + 1 ) v ϕ n ( D + 1 ) log ϕ n ( D + 1 ) + λ n ( i = 1 D ϕ n ( i ) 1 )
and therefore we have
L ϕ n l = ( Ψ ( γ d ) Ψ ( l = 1 D γ l ) ) + log β w ( i v ) log ϕ n i 1 + λ n
and
L ϕ n ( D + 1 ) = ( Ψ ( β γ ) Ψ ( α γ + β γ ) )
Setting the above equation to zero leads to
ϕ n l = β l v e ( λ n 1 ) e ( Ψ ( γ i ) Ψ ( i i = 1 D γ i i ) )
ϕ n ( D + 1 ) = β ( D + 1 ) v e ( λ n 1 ) e ( Ψ ( β γ ) Ψ ( α γ + β γ ) )
Considering that d = 1 D + 1 ϕ n ( d ) = 1 for the normalization factor, we have
e λ n 1 = 1 β ( D + 1 ) v e ( λ n 1 ) e ( Ψ ( β γ ) Ψ ( α γ + β γ ) ) + β l v e ( λ n 1 ) e ( Ψ ( γ i ) Ψ ( i i = 1 D γ i i ) )

Appendix C.1. Variational Beta-Liouville

The updates mentioned were designed to converge to a local maximum of a lower bound of log p ( Ω , Υ | r ) , which is optimal for all product approximations such as q ( m ) q ( w ) for the joint probability p ( m , w | Ω , Υ , r ) . This approach ensures that the variational parameters are fine-tuned to best approximate the true posterior distributions within the constraints of the model.
Φ l = Γ ( α ) Γ ( β ) Γ ( d = 1 D α d ) Γ ( α + β ) m n l ( λ n 1 ) ( Ψ ( γ l ) Ψ ( l = 1 D γ l )
γ l = α l + n = 1 N m n l
Ω ( l j ) = Γ ( α ) Γ ( β ) Γ ( d = 1 D α d ) Γ ( α + β ) ( 2 f d = 1 M n = 1 N d m d n l w d n j )
In this case, variable Ω vanishes because m is defined in terms of the KL approximation. In the second step, the algorithm now optimizes for m. Since q ( w | γ , r , m ) can be precisely modeled with multinomials, the minimum KL divergence is zero. As a result, the updates that follow converge to a local threshold of log p ( Ω , m | r ) :
γ l = Γ ( α ) Γ ( β ) Γ ( d = 1 D α d ) Γ ( α + β ) Ω m n l
m n l = Γ ( α ) Γ ( β ) Γ ( d = 1 D α d ) Γ ( α + β ) Ω l v e ( λ n 1 ) e ( Ψ ( γ i ) Ψ ( i i = 1 D γ i i )
Ω i j = Γ ( α ) Γ ( β ) Γ ( d = 1 D α d ) Γ ( α + β ) ( 2 f + ( n e ( λ n 1 ) e ( Ψ ( γ i ) Ψ ( i i = 1 D γ i i ) )
Considering that d = 1 D + 1 ϕ n ( d ) = 1 for the normalization factor, we have
e λ n 1 = 1 m ( D + 1 ) v e ( λ n 1 ) e ( Ψ ( β γ ) Ψ ( α γ + β γ ) ) + m l v e ( λ n 1 ) e ( Ψ ( γ i ) Ψ ( i i = 1 D γ i i ) )

References

  1. Aggarwal, C.C. An Introduction to Cluster Analysis. In Data Clustering: Algorithms and Applications; Aggarwal, C.C., Reddy, C.K., Eds.; CRC: Boca Raton, FL, USA, 2013; pp. 1–28. [Google Scholar]
  2. Mao, J.; Jain, A.K. Artificial neural networks for feature extraction and multivariate data projection. IEEE Trans. Neural Netw. 1995, 6, 296–317. [Google Scholar] [PubMed]
  3. Yu, S.; Yu, K.; Tresp, V.; Kriegel, H.P. A probabilistic clustering-projection model for discrete data. In Proceedings of the European Conference on Principles of Data Mining and Knowledge Discovery; Springer: Berlin/Heidelberg, Germany, 2005; pp. 417–428. [Google Scholar]
  4. Lai, S.; Xu, L.; Liu, K.; Zhao, J. Recurrent convolutional neural networks for text classification. In Proceedings of the Twenty-Ninth AAAI Conference on Artificial Intelligence, Austin, TX, USA, 25–30 January 2015; pp. 2267–2273. [Google Scholar]
  5. Siddharthan, A.; Mani, I.; Maybury, M.T. (Eds.) Advances in Automatic Text Summarization; MIT Press: Cambridge, MA, USA, 1999. [Google Scholar]
  6. Beeferman, D.; Berger, A.; Lafferty, J. Statistical models for text segmentation. Mach. Learn. 1999, 34, 177–210. [Google Scholar] [CrossRef]
  7. Jelodar, H.; Wang, Y.; Yuan, C.; Feng, X.; Jiang, X.; Li, Y.; Zhao, L. Latent Dirichlet allocation (LDA) and topic modeling: Models, applications, a survey. Multimed. Tools Appl. 2019, 78, 15169–15211. [Google Scholar] [CrossRef]
  8. Feldman, R. Techniques and applications for sentiment analysis. Commun. ACM 2013, 56, 82–89. [Google Scholar] [CrossRef]
  9. Hua, T.; Lu, C.T.; Choo, J.; Reddy, C.K. Probabilistic topic modeling for comparative analysis of document collections. ACM Trans. Knowl. Discov. Data (TKDD) 2020, 14, 1–27. [Google Scholar] [CrossRef]
  10. Cohn, D.A.; Hofmann, T. The missing link-a probabilistic model of document content and hypertext connectivity. In Proceedings of the Advances in Neural Information Processing Systems, Vancouver, BC, Canada, 3–8 December 2001; pp. 430–436. [Google Scholar]
  11. Hofmann, T. Probabilistic latent semantic indexing. In Proceedings of the 22nd Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, Berkeley, CA, USA, 15–19 August 1999; pp. 50–57. [Google Scholar]
  12. Blei, D.M.; Ng, A.Y.; Jordan, M.I. Latent dirichlet allocation. J. Mach. Learn. Res. 2003, 3, 993–1022. [Google Scholar]
  13. Ding, C.; He, X.; Zha, H.; Simon, H.D. Adaptive dimension reduction for clustering high dimensional data. In Proceedings of the 2002 IEEE International Conference on Data Mining, Maebashi City, Japan, 9–12 December 2002; pp. 147–154. [Google Scholar]
  14. Li, T.; Ma, S.; Ogihara, M. Document clustering via adaptive subspace iteration. In Proceedings of the 27th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, Sheffield, UK, 25–29 July 2004; pp. 218–225. [Google Scholar]
  15. Syed, S.; Spruit, M. Full-text or abstract examining topic coherence scores using latent dirichlet allocation. In Proceedings of the 2017 IEEE International Conference on Data Science and Advanced Analytics (DSAA), Tokyo, Japan, 19–21 October 2017; pp. 165–174. [Google Scholar]
  16. Edison, H.; Carcel, H. Text data analysis using Latent Dirichlet Allocation: An application to FOMC transcripts. Appl. Econ. Lett. 2021, 28, 38–42. [Google Scholar] [CrossRef]
  17. Lee, D.D.; Seung, H.S. Learning the parts of objects by non-negative matrix factorization. Nature 1999, 401, 788–791. [Google Scholar] [CrossRef]
  18. Collins, M.; Dasgupta, S.; Schapire, R.E. A Generalization of Principal Components Analysis to the Exponential Family. In Proceedings of the Advances in Neural Information Processing Systems 14: Natural and Synthetic, NIPS 2001, Vancouver, BC, Canada, 3–8 December 2001; pp. 617–624. [Google Scholar]
  19. Buntine, W. Variational extensions to EM and multinomial PCA. In Proceedings of the European Conference on Machine Learning, Helsinki, Finland, 19–23 August 2002; pp. 23–34. [Google Scholar]
  20. Jouvin, N.; Latouche, P.; Bouveyron, C.; Bataillon, G.; Livartowski, A. Clustering of count data through a mixture of multinomial PCA. arXiv 2019, arXiv:1909.00721. [Google Scholar] [CrossRef]
  21. Griffiths, T.L.; Jordan, M.I.; Tenenbaum, J.B.; Blei, D.M. Hierarchical topic models and the nested chinese restaurant process. In Proceedings of the Advances in Neural Information Processing Systems, Vancouver, BC, Canada, 13–18 December 2004; pp. 17–24. [Google Scholar]
  22. Hoffman, M.; Bach, F.R.; Blei, D.M. Online learning for latent dirichlet allocation. In Proceedings of the Advances in Neural Information Processing Systems, Vancouver, BC, Canada, 6–19 December 2010; pp. 856–864. [Google Scholar]
  23. Fitzgerald, W.J. Markov chain Monte Carlo methods with applications to signal processing. Signal Process. 2001, 81, 3–18. [Google Scholar] [CrossRef]
  24. Luo, Z.; Amayri, M.; Fan, W.; Bouguila, N. Cross-collection latent Beta-Liouville allocation model training with privacy protection and applications. Appl. Intell. 2023, 53, 17824–17848. [Google Scholar] [CrossRef] [PubMed]
  25. Najar, F.; Bouguila, N. Sparse document analysis using beta-liouville naive bayes with vocabulary knowledge. In Proceedings of the Document Analysis and Recognition–ICDAR 2021: 16th International Conference, Lausanne, Switzerland, 5–10 September 2021; Proceedings, Part II 16. Springer: Berlin/Heidelberg, Germany, 2021; pp. 351–363. [Google Scholar]
  26. Connor, R.J.; Mosimann, J.E. Concepts of independence for proportions with a generalization of the Dirichlet distribution. J. Am. Stat. Assoc. 1969, 64, 194–206. [Google Scholar] [CrossRef]
  27. Lacoste-Julien, S.; Sha, F.; Jordan, M.I. DiscLDA: Discriminative learning for dimensionality reduction and classification. In Proceedings of the Advances in Neural Information Processing Systems, Vancouver, BC, Canada, 8–10 December 2008; pp. 897–904. [Google Scholar]
  28. Rabinovich, M.; Blei, D. The inverse regression topic model. In Proceedings of the International Conference on Machine Learning. PMLR, 2014, Beijing, China, 21–26 June 2014; pp. 199–207. [Google Scholar]
  29. Ramage, D.; Hall, D.; Nallapati, R.; Manning, C.D. Labeled LDA: A supervised topic model for credit attribution in multi-labeled corpora. In Proceedings of the 2009 Conference on Empirical Methods in Natural Language Processing, Singapore, 6–7 August 2009; pp. 248–256. [Google Scholar]
  30. Chemudugunta, C.; Smyth, P.; Steyvers, M. Modeling general and specific aspects of documents with a probabilistic topic model. Adv. Neural Inf. Process. Syst. 2006, 19, 241–248. [Google Scholar]
  31. Ge, T.; Pei, W.; Ji, H.; Li, S.; Chang, B.; Sui, Z. Bring you to the past: Automatic generation of topically relevant event chronicles. In Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), Beijing, China, 26–31 July 2015; pp. 575–585. [Google Scholar]
  32. Onan, A. Two-stage topic extraction model for bibliometric data analysis based on word embeddings and clustering. IEEE Access 2019, 7, 145614–145633. [Google Scholar] [CrossRef]
  33. Onan, A.; Toçoğlu, M.A. A term weighted neural language model and stacked bidirectional LSTM based framework for sarcasm identification. IEEE Access 2021, 9, 7701–7722. [Google Scholar] [CrossRef]
  34. Meena, G.; Mohbey, K.K.; Indian, A.; Khan, M.Z.; Kumar, S. Identifying emotions from facial expressions using a deep convolutional neural network-based approach. Multimed. Tools Appl. 2023, 83, 15711–15732. [Google Scholar] [CrossRef]
  35. Meena, G.; Mohbey, K.K.; Kumar, S. Sentiment analysis on images using convolutional neural networks based Inception-V3 transfer learning approach. Int. J. Inf. Manag. Data Insights 2023, 3, 100174. [Google Scholar] [CrossRef]
  36. Meena, G.; Mohbey, K.K.; Kumar, S.; Lokesh, K. A hybrid deep learning approach for detecting sentiment polarities and knowledge graph representation on monkeypox tweets. Decis. Anal. J. 2023, 7, 100243. [Google Scholar] [CrossRef]
  37. Deerwester, S.; Dumais, S.T.; Furnas, G.W.; Landauer, T.K.; Harshman, R. Indexing by latent semantic analysis. J. Am. Soc. Inf. Sci. 1990, 41, 391–407. [Google Scholar] [CrossRef]
  38. Tipping, M.E.; Bishop, C.M. Probabilistic principal component analysis. J. R. Stat. Soc. Ser. B Stat. Methodol. 1999, 61, 611–622. [Google Scholar] [CrossRef]
  39. Minka, T. Estimating a Dirichlet Distribution; Technical Report; MIT: Cambridge, MA, USA, 2003; Volume 1, p. 1. [Google Scholar]
  40. Bouguila, N. Clustering of count data using generalized Dirichlet multinomial distributions. IEEE Trans. Knowl. Data Eng. 2008, 20, 462–474. [Google Scholar] [CrossRef]
  41. Bouguila, N.; Ziou, D. High-dimensional unsupervised selection and estimation of a finite generalized Dirichlet mixture model based on minimum message length. IEEE Trans. Pattern Anal. Mach. Intell. 2007, 29, 1716–1731. [Google Scholar] [CrossRef] [PubMed]
  42. Bakhtiari, A.S.; Bouguila, N. A variational bayes model for count data learning and classification. Eng. Appl. Artif. Intell. 2014, 35, 176–186. [Google Scholar] [CrossRef]
  43. Koochemeshkian, P.; Zamzami, N.; Bouguila, N. Flexible Distribution-Based Regression Models for Count Data: Application to Medical Diagnosis. Cybern. Syst. 2020, 51, 442–466. [Google Scholar] [CrossRef]
  44. Jordan, M.I.; Ghahramani, Z.; Jaakkola, T.S.; Saul, L.K. An introduction to variational methods for graphical models. Mach. Learn. 1999, 37, 183–233. [Google Scholar] [CrossRef]
  45. Bouguila, N. Count Data Modeling and Classification Using Finite Mixtures of Distributions. IEEE Trans. Neural Netw. 2011, 22, 186–198. [Google Scholar] [CrossRef]
  46. Ihou, K.E.; Bouguila, N.; Bouachir, W. Efficient integration of generative topic models into discriminative classifiers using robust probabilistic kernels. Pattern Anal. Appl. 2021, 24, 217–241. [Google Scholar] [CrossRef]
  47. Espinosa, K.L.C.; Barajas, J.; Akella, R. The generalized dirichlet distribution in enhanced topic detection. In Proceedings of the 21st ACM International Conference on Information and Knowledge Management, CIKM’12, Maui, HI, USA, 29 October–2 November 2012; pp. 773–782. [Google Scholar]
  48. Shojaee Bakhtiari, A. Count Data Modeling and Classification Using Statistical Hierarchical Approaches and Multi-topic Models. Ph.D. Thesis, Concordia University, Montreal, QC, Canada, 2014. [Google Scholar]
  49. Bakhtiari, A.S.; Bouguila, N. A latent Beta-Liouville allocation model. Expert Syst. Appl. 2016, 45, 260–272. [Google Scholar] [CrossRef]
  50. Teh, Y.; Newman, D.; Welling, M. A collapsed variational Bayesian inference algorithm for latent Dirichlet allocation. Adv. Neural Inf. Process. Syst. 2006, 19, 1353–1360. [Google Scholar]
  51. Ihou, K.E.; Bouguila, N. Stochastic topic models for large scale and nonstationary data. Eng. Appl. Artif. Intell. 2020, 88, 103364. [Google Scholar] [CrossRef]
  52. Li, S.; Zhang, Y.; Pan, R. Bi-directional recurrent attentional topic model. ACM Trans. Knowl. Discov. Data (TKDD) 2020, 14, 1–30. [Google Scholar] [CrossRef]
  53. Horgan, J. From complexity to perplexity. Sci. Am. 1995, 272, 104–109. [Google Scholar] [CrossRef]
  54. Sebastiani, F. Machine learning in automated text categorization. ACM Comput. Surv. 2002, 34, 1–47. [Google Scholar] [CrossRef]
  55. Riloff, E.; Lehnert, W. Information extraction as a basis for high-precision text classification. ACM Trans. Inf. Syst. 1994, 12, 296–333. [Google Scholar] [CrossRef]
  56. Wallach, H.M. Topic modeling: Beyond bag-of-words. In Proceedings of the 23rd International Conference on Machine Learning, Pittsburgh, PA, USA, 25–29 June 2006; pp. 977–984. [Google Scholar]
  57. Bamman, D.; Smith, N.A. New alignment methods for discriminative book summarization. arXiv 2013, arXiv:1305.1319. [Google Scholar]
  58. Onan, A. Biomedical text categorization based on ensemble pruning and optimized topic modelling. Comput. Math. Methods Med. 2018, 2018, 2497471. [Google Scholar] [CrossRef]
  59. Cohen, R.; Elhadad, M.; Elhadad, N. Redundancy in electronic health record corpora: Analysis, impact on text mining performance and mitigation strategies. BMC Bioinform. 2013, 14, 10. [Google Scholar] [CrossRef]
  60. Wrenn, J.O.; Stein, D.M.; Bakken, S.; Stetson, P.D. Quantifying clinical narrative redundancy in an electronic health record. J. Am. Med. Inform. Assoc. 2010, 17, 49–53. [Google Scholar] [CrossRef] [PubMed]
  61. Karami, A.; Gangopadhyay, A.; Zhou, B.; Kharrazi, H. Flatm: A fuzzy logic approach topic model for medical documents. In Proceedings of the 2015 Annual Conference of the North American Fuzzy Information Processing Society (NAFIPS) Held Jointly with 2015 5th World Conference on Soft Computing (WConSC), Redmond, WA, USA, 17–19 August 2015; IEEE: Piscataway, NJ, USA; pp. 1–6. [Google Scholar]
  62. Karami, A.; Gangopadhyay, A.; Zhou, B.; Kharrazi, H. A fuzzy approach model for uncovering hidden latent semantic structure in medical text collections. In Proceedings of the iConference 2015, Newport Beach, CA, USA, 24–27 March 2015. [Google Scholar]
  63. BIONLP. Available online: https://www.ncbi.nlm.nih.gov/research/bionlp/ (accessed on 3 October 2021).
  64. Karami, A.; Gangopadhyay, A.; Zhou, B.; Kharrazi, H. Fuzzy approach topic discovery in health and medical corpora. Int. J. Fuzzy Syst. 2018, 20, 1334–1345. [Google Scholar] [CrossRef]
  65. Agarwal, A.; Xie, B.; Vovsha, I.; Rambow, O.; Passonneau, R.J. Sentiment analysis of twitter data. In Proceedings of the Workshop on Language in Social Media (LSM 2011), Portland, OR, USA, 23 June 2011; pp. 30–38. [Google Scholar]
  66. Onan, A. Sentiment analysis on product reviews based on weighted word embeddings and deep neural networks. Concurr. Comput. Pract. Exp. 2021, 33, e5909. [Google Scholar] [CrossRef]
  67. Yan, X.; Li, G.; Li, Q.; Chen, J.; Chen, W.; Xia, F. Sentiment analysis on massive open online course evaluation. In Proceedings of the 2021 International Conference on Neuromorphic Computing (ICNC), Wuhan, China, 11–14 October 2021; pp. 245–249. [Google Scholar]
  68. Onan, A.; Korukoğlu, S.; Bulut, H. A multiobjective weighted voting ensemble classifier based on differential evolution algorithm for text sentiment classification. Expert Syst. Appl. 2016, 62, 1–16. [Google Scholar] [CrossRef]
  69. Blitzer, J.; Dredze, M.; Pereira, F. Biographies, bollywood, boom-boxes and blenders: Domain adaptation for sentiment classification. In Proceedings of the 45th Annual Meeting of the Association of Computational Linguistics, Prague, Czechia, 25–27 June 2007; pp. 440–447. [Google Scholar]
  70. Gupta, R.D.; Kundu, D. Exponentiated exponential family: An alternative to gamma and Weibull distributions. Biom. J. J. Math. Methods Biosci. 2001, 43, 117–130. [Google Scholar] [CrossRef]
  71. Buntine, W.L. Operations for learning with graphical models. J. Artif. Intell. Res. 1994, 2, 159–225. [Google Scholar] [CrossRef]
Figure 1. Success rate for CMU Book data.
Figure 1. Success rate for CMU Book data.
Electronics 13 01859 g001
Figure 2. Success rate for Tmvar corpus data.
Figure 2. Success rate for Tmvar corpus data.
Electronics 13 01859 g002
Figure 3. Success rate for sentiment dataset.
Figure 3. Success rate for sentiment dataset.
Electronics 13 01859 g003
Table 1. Parameters of generalized Dirichlet and Beta-Liouville distributions.
Table 1. Parameters of generalized Dirichlet and Beta-Liouville distributions.
ParameterGeneralized Dirichlet (GDMPCA)Beta-Liouville (BLMPCA)
ξ Parameters of GD distributionNot applicable
Υ Not applicableParameters of BL distribution
mMixture weights (GD)Mixture weights (BL)
zTopic assignmentsTopic assignments
wWords in documentsWords in documents
Ω Multinomial parameters (words)Multinomial parameters (words)
LNumber of words per documentNumber of words per document
C , Ω k , c k Multinomial parameters for topicsMultinomial parameters for topics
Table 2. Common topics identified with the BLMPCA model on the CMU Book dataset, each defined by a set of keywords.
Table 2. Common topics identified with the BLMPCA model on the CMU Book dataset, each defined by a set of keywords.
Topic NumberTopics
Topic 1girl, tells, find, two, man, when, return, after, also, finds, time, kill, later, help, killed
Topic 2he, one, back, man, time, house, father, police, story, mother, young, school, love, time, first
Topic 3tells, they, return, find, girl, back, one, house, story, after, dragon, find, schools, boy, jack
Topic 4earth, world, one, human, ship, book, planet, space, human, systems, time, years, in, people, would
Topic 5war, novel, new, world, army, story, one, group, book, states, general, british, president, first, american
Table 3. A comparison of the perplexity of the MPCA, GDMPCA, and BLMPCA models, indicating the model fit quality across different topic numbers (K) on the CMU Book dataset.
Table 3. A comparison of the perplexity of the MPCA, GDMPCA, and BLMPCA models, indicating the model fit quality across different topic numbers (K) on the CMU Book dataset.
K5101520
MPCA1455142213201215
GDMPCA1326143011901178
BLMPCA1319120311981177
Table 4. Time complexity comparison for MPCA, GDMPCA, and BLMPCA at varying topic levels (K) on the CMU Book dataset.
Table 4. Time complexity comparison for MPCA, GDMPCA, and BLMPCA at varying topic levels (K) on the CMU Book dataset.
K5101520
MPCA107.803140.1439150.9242161.7045
GDMPCA225.04230.544347.056408.064
BLMPCA251.64327.132352.296377.46
Table 5. Comparison perplexity scores of MPCA, GDMPCA, and BLMPCA, reflecting the model fit as the topic count (K) increases on the CMU Book dataset with CGS inference.
Table 5. Comparison perplexity scores of MPCA, GDMPCA, and BLMPCA, reflecting the model fit as the topic count (K) increases on the CMU Book dataset with CGS inference.
K5101520
MPCA1391.51448.615161580
GDMPCA1291.2131614281413
BLMPCA1310.41324.814161483.2
Table 6. Time complexity comparison for MPCA, GDMPCA, and BLMPCA with increasing topics (K) using CGS inference on the CMU Book dataset.
Table 6. Time complexity comparison for MPCA, GDMPCA, and BLMPCA with increasing topics (K) using CGS inference on the CMU Book dataset.
K5101520
MPCA431.212536.57634.69687.818
GDMPCA19125.21138.2642429.3922964.51
BLMPCA1998.842289.9243018.3683497.14
Table 7. Comparison of the perplexity of the MPCA, GDMPCA, and BLMPCA models, indicating the model fit quality across different topic numbers (K) on the TMVAR dataset with variation EM inference.
Table 7. Comparison of the perplexity of the MPCA, GDMPCA, and BLMPCA models, indicating the model fit quality across different topic numbers (K) on the TMVAR dataset with variation EM inference.
K5101520
MPCA2115208319841977
GDMPCA1996198919681959
BLMPCA1983196519541949
Table 8. Time complexity comparison for MPCA, GDMPCA, and BLMPCA with increasing topics (K) using variation EM inference on the TMVAR dataset.
Table 8. Time complexity comparison for MPCA, GDMPCA, and BLMPCA with increasing topics (K) using variation EM inference on the TMVAR dataset.
K5101520
MPCA9.5322.54326.09228.458
GDMPCA11.8324.84328.39230.758
BLMPCA18.5738.99744.56848.282
Table 9. Comparison of the perplexity of the MPCA, GDMPCA, and BLMPCA models, indicating the model fit quality across different topic numbers (K) on the TMVAR dataset with CGS inference.
Table 9. Comparison of the perplexity of the MPCA, GDMPCA, and BLMPCA models, indicating the model fit quality across different topic numbers (K) on the TMVAR dataset with CGS inference.
K5101520
MPCA2132.52232.82376.02460
GDMPCA1360.91182.41345.61938
BLMPCA1938.511350.51340.51440
Table 10. Time complexity comparison for MPCA, GDMPCA, and BLMPCA with increasing topics (K) using CGS inference on the TMVAR dataset.
Table 10. Time complexity comparison for MPCA, GDMPCA, and BLMPCA with increasing topics (K) using CGS inference on the TMVAR dataset.
K5101520
MPCA45.7462.89108.20200.63
GDMPCA56.74163.95252.35273.70
BLMPCA165.57336.93376.45392.71
Table 11. Common topics identified with the BLMPCA model in the TMVAR dataset, each defined by a set of keywords.
Table 11. Common topics identified with the BLMPCA model in the TMVAR dataset, each defined by a set of keywords.
Topic NumberTopics
Topic 1mutations, mutation, gene, family, patients, iron, exon, novel, autosomal, associated
Topic 2gene, p, cancer, polymorphism, expression, patients, associated, deletion, study, region
Topic 3gene, patients, dna mutation, polymorphism, detected, samples, family, study, results, dna
Topic 4dna mutation, mutations, homozygous, variants, family, ct, position, methods, associated, substitution
Topic 5gene, patients, protein mutation, dna, exon, study, genetic, cancer, substitution, genotype
Table 12. Comparison of the perplexity of the MPCA, GDMPCA, and BLMPCA models, indicating the model fit quality across different topic numbers (K) on sentiment data with variation EM inference.
Table 12. Comparison of the perplexity of the MPCA, GDMPCA, and BLMPCA models, indicating the model fit quality across different topic numbers (K) on sentiment data with variation EM inference.
K2358
MPCA1551153115421529
GDMPCA1549153915241521
BLMPCA1448154015311518
Table 13. Time complexity comparison for MPCA, GDMPCA, and BLMPCA with increasing topics (K) using variational EM inference on the sentiment analysis application.
Table 13. Time complexity comparison for MPCA, GDMPCA, and BLMPCA with increasing topics (K) using variational EM inference on the sentiment analysis application.
K5101520
MPCA130.54169.702182.756195.81
GDMPCA142.876185.7388200.0264214.314
BLMPCA158.23205.699221.522237.345
Table 14. Frequency of emotions identified in text data via topic modeling.
Table 14. Frequency of emotions identified in text data via topic modeling.
EmotionsCount
satisfied78,901
angry21,345
happy6521
joy82,345
disgust7125
Perfect45,459
Tearful3451
sad4387
Table 15. The counts of positive, negative, and unlabeled sentiments identified through sentiment analysis.
Table 15. The counts of positive, negative, and unlabeled sentiments identified through sentiment analysis.
SentimentCount
Positive213,232
Negative36,308
Unlabeled23,451
Table 16. Comparison of the perplexity of the MPCA, GDMPCA, and BLMPCA models, indicating the model fit quality across different topic numbers (K) on sentiment data with CGS inference.
Table 16. Comparison of the perplexity of the MPCA, GDMPCA, and BLMPCA models, indicating the model fit quality across different topic numbers (K) on sentiment data with CGS inference.
K2358
MPCA1451151115891639
GDMPCA1332139314221502
BLMPCA1316140114131498
Table 17. Time complexity comparison for MPCA, GDMPCA, and BLMPCA with increasing topics (K) using CGS inference on the sentiment analysis application.
Table 17. Time complexity comparison for MPCA, GDMPCA, and BLMPCA with increasing topics (K) using CGS inference on the sentiment analysis application.
K5101520
MPCA830.541069.7021282.7561495.81
GDMPCA924.4511258.781319.461383.17
BLMPCA1085.421264.241390.121473.623
Table 18. Accuracy comparisons for sentiment analysis classifiers.
Table 18. Accuracy comparisons for sentiment analysis classifiers.
ClassifierSVMNaiveBayesMLP
MPCA0.620.680.67
GDMPCA0.800.850.87
BLMPCA0.830.880.88
Table 19. Recall metrics for SVM, Naive Bayes, and MLP classifiers using MPCA, GDMPCA, and BLMPCA in sentiment analysis.
Table 19. Recall metrics for SVM, Naive Bayes, and MLP classifiers using MPCA, GDMPCA, and BLMPCA in sentiment analysis.
ClassifierSVMNaiveBayesMLP
MPCA0.610.590.66
GDMPCA0.790.760.85
BLMPCA0.850.820.89
Table 20. F1-score metrics for SVM, Naive Bayes, and MLP classifiers using MPCA, GDMPCA, and BLMPCA in sentiment analysis.
Table 20. F1-score metrics for SVM, Naive Bayes, and MLP classifiers using MPCA, GDMPCA, and BLMPCA in sentiment analysis.
ClassifierSVMNaive BayesMLP
MPCA0.61950.60410.6697
GDMPCA0.79990.77010.8593
BLMPCA0.85930.83130.8999
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Koochemeshkian, P.; Ihou Koffi, E.; Bouguila, N. Hidden Variable Models in Text Classification and Sentiment Analysis. Electronics 2024, 13, 1859. https://doi.org/10.3390/electronics13101859

AMA Style

Koochemeshkian P, Ihou Koffi E, Bouguila N. Hidden Variable Models in Text Classification and Sentiment Analysis. Electronics. 2024; 13(10):1859. https://doi.org/10.3390/electronics13101859

Chicago/Turabian Style

Koochemeshkian, Pantea, Eddy Ihou Koffi, and Nizar Bouguila. 2024. "Hidden Variable Models in Text Classification and Sentiment Analysis" Electronics 13, no. 10: 1859. https://doi.org/10.3390/electronics13101859

APA Style

Koochemeshkian, P., Ihou Koffi, E., & Bouguila, N. (2024). Hidden Variable Models in Text Classification and Sentiment Analysis. Electronics, 13(10), 1859. https://doi.org/10.3390/electronics13101859

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop