Next Article in Journal
Ultrasound-Assisted Extraction of an Extract with Dermatological and Nosocomial Activity from Agave nuusaviorum, a Mexican Endemic Plant
Next Article in Special Issue
Vehicle-Following Control Based on Deep Reinforcement Learning
Previous Article in Journal
AWMC: Abnormal-Weather Monitoring and Curation Service Based on Dynamic Graph Embedding
Previous Article in Special Issue
A Distributed Bi-Behaviors Crow Search Algorithm for Dynamic Multi-Objective Optimization and Many-Objective Optimization Problems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel Mixed-Attribute Fusion-Based Naive Bayesian Classifier

1
Big Data Institute, College of Computer Science and Software Engineering, Shenzhen University, Shenzhen 518060, China
2
Guangdong Laboratory of Artificial Intelligence and Digital Economy (SZ), Shenzhen 518107, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2022, 12(20), 10443; https://doi.org/10.3390/app122010443
Submission received: 26 September 2022 / Revised: 8 October 2022 / Accepted: 10 October 2022 / Published: 17 October 2022
(This article belongs to the Special Issue Recent Advances in Machine Learning and Computational Intelligence)

Abstract

:
The Naive Bayesian classifier (NBC) is a well-known classification model that has a simple structure, low training complexity, excellent scalability, and good classification performances. However, the NBC has two key limitations: (1) it is built upon the strong assumption that condition attributes are independent, which often does not hold in real-life, and (2) the NBC does not handle continuous attributes well. To overcome these limitations, this paper presents a novel approach for NBC construction, called mixed-attribute fusion-based NBC (MAF-NBC). It alleviates the two aforementioned limitations by relying on a mixed-attribute fusion mechanism with an improved autoencoder neural network for NBC construction. MAF-NBC transforms the original mixed attributes of a data set into a series of encoded attributes with maximum independence as a pre-processing step. To guarantee the generation of useful encoded attributes, an efficient objective function is designed to optimize the weights of the autoencoder neural network by considering both the encoding error and the attribute’s dependence. A series of persuasive experiments was conducted to validate the feasibility, rationality, and effectiveness of the designed MAF-NBC approach. Results demonstrate that MAF-NBC has superior classification performance than eight state-of-the-art Bayesian algorithms, namely the discretization-based NBC (Dis-NBC), flexible naive Bayes (FNB), tree-augmented naive (TAN) Bayes, averaged one-dependent estimator (AODE), hidden naive Bayes (HNB), deep feature weighting for NBC (DFW-NBC), correlation-based feature weighting filter for NBC (CFW-NBC), and independent component analysis-based NBC (ICA-NBC).

1. Introduction

As one of the top 10 algorithms in the fields of data mining and machine learning [1], the naive Bayesian classifier (NBC) has been used in numerous domains. The main advantage of the NBC is its simple model structure that makes it easy to implement and its good theoretical interpretability. In recent years, the NBC also received much attention from the industry and academia since it can be easily deployed in distributed environments to process big data. Despite possessing several desirable properties, the NBC is built upon a strong assumption, called the attribute independence assumption, which states that condition attributes must be mutually independent with respect to the decision attribute. This assumption simplifies the calculation of posterior probabilities (the probabilities that attribute values of a sample are all observed for a given class). Rather than computing a posterior probability as a joint probability of condition attributes, the NBC calculates it as the product of multiple marginal probabilities. This makes the computation of NBC very efficient and allows the estimation of probabilities even with small data sets. However, the attribute independence assumption does not hold in many real-life data sets, which substantially limits the prediction performance of the NBC. Recent studies on the NBC mainly focused on finding ways to relax the independence assumption so as to further improve its generalization performance. There are two main approaches for improving the testing ability of the NBC [2]: (1) improving the structure of models and (2) performing data transformations. The former approach introduces complex Bayesian network structures to model attribute dependencies, while the latter applies feature selection or feature extraction methods to select independent attributes. Some representative studies of these two approaches are described next.
Model structure-oriented improvement methods use different estimation strategies (e.g., density estimation, Bayesian network, and attribute weighting) to estimate the conditional probability that a sample’s attribute values are observed given its class. The flexible naive Bayes (FNB) [3] relies on kernel density estimations to determine a class’s conditional probabilities. However, FNB can only handle continuous attributes and does not offer a solution to relax the attribute independence assumption. The tree-augmented naive (TAN) Bayes [4] algorithm is a semi-naïve Bayesian learning method that relaxes the attribute independence assumption by employing a tree structure, where each condition attribute only depends on the decision attribute and at most one condition attribute. The averaged one-dependent estimator (AODE) [5] is an ensemble classifier that applies an attribute selection algorithm to construct a series of one-dependent classifiers. Each classifier is a simple Bayesian network that is obtained by averaging all one-dependence classifiers. The hidden naive Bayes (HNB) classifier [6] uses the weighted sum of two-attribute dependencies to represent multiple-attribute dependence, where the weights are determined based on the mutual information between two condition attributes. The deep feature weighting for NBC (DFW-NBC) [7] technique and correlation-based feature weighting filter for NBC (CFW-NBC) [8] are two attribute weighting-based NBCs. DFW-NBC estimates each conditional probability of the NBC by deeply computing attribute weighted frequencies from training data while CFW-NBC weights the condition attributes by considering both the mutual relevance and the average mutual redundancy.
Data transformation-oriented improvement methods focus on the deep exploration of the training data to avoid complex and time-consuming structure learning. The NBCs obtained by this approach still rely on the independence assumption; thus, the performance may be degraded if the training data do not satisfy that assumption. Feature extraction techniques are the most commonly used to transform data with attribute dependencies into data that have no dependencies. Bressan and Vitria [9] applied class-conditional independent component analysis (CC-ICA) to improve the classification performance of the NBC. Qin et al. [10] proposed an ICA-based NBC (ICA-NBC) to improve the classification performance of the NBC. Fan and Poh [11] evaluated the classification performance of the NBC built using three feature extraction methods, namely principal component analysis (PCA), ICA, and CC-ICA. Experimental results have shown that PCA, ICA, and CC-ICA can slightly increase the testing accuracies of the NBC on the selected data sets. Jayanthi and Sasikala [12] trained an NBC based on website attributes extracted by PCA to perform web link spam detection. Zhang et al. [13] applied PCA to extract key attributes in network data and then trained the NBC to conduct network intrusion detection.
Although the aforementioned methods can improve the classification performance of the NBC for specific application scenarios, these methods do not provide a good solution for handling mixed attributes (continuous and categorical attributes). Some studies [14,15] revealed that the discretization of continuous attributes can lead to losing precious information about the original data set, while the determination of an optimal Bayesian network structure is an NP-hard problem [16]. To address the above limitations, this paper proposes a new strategy to enhance the generalization capability of the NBC for the mixed-attribute classification problem. The proposed approach not only retains as much information as possible about the original data but also keeps the model structure as simple as possible. The contributions of this paper are summarized as follows. A new mixed attribute fusion-based NBC (MAF-NBC) is proposed for mixed-attribute classification problems. To relax the attribute independence assumption, an autoencoder neural network (ANN) based on the minimization of both encoding error and attribute dependence is iteratively trained to transform the original mixed attributes into a series of independent and encoded attributes. Extensive experiments conducted on multiple data sets demonstrate the feasibility, rationality, and effectiveness of the MAF-NBC.
The remainder of this paper is organized as follows. Section 2 reviews the principles of the NBC. Section 3 presents the proposed MAF-NBC. Section 4 reports results from experiments to assess the performance of MAF-NBC. Finally, Section 5 concludes this paper and describes future studies.

2. Naive Bayesian Classifier for Mixed-Attribute Classification

Let there be a classification data set D = m = 1 M D m containing N samples. Each sample has D condition attributes including D 1 discrete attributes and D 2 continuous attributes. Moreover, all samples are divided into M different classes, where N m denotes the number of samples that belong to the m-th ( m = 1 , 2 , , M ) class.
D m = x n m , y n m x n m = a n 1 m , , a n D 1 m , b n 1 m , , b n D 2 m , y n m = w m , n = 1 , 2 , , N m ,
m = 1 M N m = N , D 1 + D 2 = D , a n i m , i 1 , 2 , , D 1 is the i-th discrete attribute value, b n j m , j 1 , 2 , , D 2 is the j-th continuous attribute value, and w 1 , w 2 , , w M is the class label set. The next paragraphs explain the principles of the naive Bayesian classifier (NBC) for the mixed-attribute classification problem on data set D .
Assume that there is a new sample x = a 1 , , a D 1 , b 1 , , b D 2 . The NBC determines its class label by using this equation:
y = arg max m = 1 , 2 , , M P w m x = arg max m = 1 , 2 , , M P x w m P w m P x arg max m = 1 , 2 , , M P x w m P w m ,
where P w m x is the posterior probability, P w m is the prior probability, and P x w m is the conditional probability. Generally, the prior probability in Equation (1) can be calculated as follows.
P w m = N m N .
The key of training an NBC for the mixed-attribute classification problem is to calculate the conditional probability based on the independent attribute assumption as follows.
P x w m = P a 1 , , a D 1 , b 1 , , b D 2 w m = i = 1 D 1 P a i w m j = 1 D 2 P b j w m .
Conditional probability P a i w m corresponding to a discrete attribute value a i is calculated as follows:
P a i w m = n = 1 N m I a i , a n i m N m ,
where
I u , v = 1 , if u = v 0 , if u v
is an indicator function that is used to count the frequency of a i in the i-th continuous attribute values a 1 i m , a 2 i m , , a N m , j m corresponding to samples belonging to the m-th class. John and Langley [3] constructed a flexible NBC (FNBC), which used the kernel density estimation technique [17] to calculate the term of P b j w m for the continuous attribute value in Equation (3) as follows:
P b j w m p b j w m = 1 N m n = 1 N m 1 2 π h j exp 1 2 b j b n j m h j 2
where p b j w m is the estimated probability density function (p.d.f.) value of b j based on the j-th continuous attribute values b 1 j m , b 2 j m , , b N m , j m corresponding to samples from the m-th class, and h j > 0 ( h j = 1 N m in [3]) is the bandwidth parameter. In addition, continuous attribute discretization can also be used to determine p b j w m as follows:
P b j w m = P c j w m = n = 1 N m I c j , c n j m N m
by transforming the continuous attribute values b j , b 1 j m , b 2 j m , , b N m , j m into the discrete attribute values c j , c 1 j m , c 2 j m , , c N m , j m . This form of discretization-based NBC is called dis-NBC in this study.

3. Mixed-Attribute Fusion-Based Naive Bayesian Classifier

As stated in the introduction, continuous attribute discretization and the attribute independence assumption limit the generalization performance of the NBC for the continuous attribute classification problem. To cope with these issues, recent studies either introduced complex structures to represent attribute dependencies or applied discretization techniques to transform mixed-attribute values into discrete attribute values [18,19,20]. This section presents a novel solution to the NBC-based mixed-attribute classification problem by considering attribute dependence and mixed-attribute transformation simultaneously. A mixed-attribute fusion strategy is designed to construct an NBC that can be trained based on the transformed continuous attributes with the minimum attribute dependence.
For discrete attributes of an original mixed-attribute (OMA) data set D , the one-hot encoding technique [21] is applied to transform them into 0-1 numerical attributes; i.e., a n i m , i = 1 , 2 , , D 1 is encoded as e n 1 m i , , e n K i m i for the discrete attribute value of the n-th sample x n m i , n = 1 , 2 , , N m , where the following is the case:
e n k m = 1 , if a n i m = A k i 0 , if a n i m A k i
and A 1 i , A 2 i , , A K i i is K i categorical values corresponding to the i-th discrete attribute. Then, the one-hot encoded form of the original sample x n m can be expressed as follows.
x ¯ n m = e n 1 m 1 , , e n K 1 m 1 , , e n 1 m D 1 , , e n K D 1 m D 1 , b n 1 m , , b n D 2 m .
In fact, the one-hot encoding technique causes attribute redundancy when extending a discrete attribute into multiple 0-1 numerical attributes. For example, the essence of the following transformation
A 1 A 1 A 2 1 0 1 0 0 1
is to represent a discrete attribute using two “repetitive” numerical attributes. The main role of one-hot encoding is to transform the discrete attribute values into numbers. The attribute independence assumption is not alleviated when constructing an NBC based on a one-hot encoded attribute (OHEA) data set.
X ¯ = x ¯ 1 1 x ¯ N 1 1 x ¯ 1 2 x ¯ N 2 2 x ¯ 1 M x ¯ N M M T .
Thus, an autoencoder neural network (ANN) is meticulously designed to solve the problems of attribute redundancy and attribute dependence mentioned above.
An ANN is a special single hidden-layer feed-forward neural network that has the same input matrix and output matrix. The main purpose of training an ANN is to find the optimal input layer weight matrix α = α r l R × L = α 1 α 2 α L and output layer matrix β = β l r L × R so that the practical output matrix X ¯ approximates the true output matrix X ¯ as closely as possible, where R = i = 1 D 1 K i + D 2 is the number of input layer nodes of the ANN and L is the number of hidden layer nodes of the ANN.
The following objective function for ANN is designed to transform one-hot encoded attributes into independent encoded attributes:
L X ¯ , α , β = λ L 1 X ¯ , α , β + 1 λ L 2 X ¯ , α , β ,
where λ ( 0 , 1 ) is a balance factor. Optimal weight matrices α ¯ and β ¯ are determined by minimizing the objective function as follows:
α ¯ , β ¯ = argmax α r l , β l r r = 1 , 2 , , R ; l = 1 , 2 , , L L X ¯ , α , β .
Here, it is worthwhile to note that the original intention of using an autoencoder neural network is to fuse the mixed attributes and to relax the attribute independence assumption rather than exploring the usage of deep learning technology. When a shallow learning can already meet the requirement of good NBC construction, it is unnecessary to resort to complex and time-consuming deep learning. The excessive attention on attribute transformation with deep learning is out of the scope of this study. Interested readers can refer to specialized studies on the combination of deep learning and supervised learning for more details on such approaches, e.g., deep support vector machine [22], deep decision tree [23], and deep nearest neighbor [24].
The first term of the objective function is the encoding error, which is used to measure the error between the practical output matrix X ¯ and the true output matrix X ¯ . It is defined as follows:
L 1 X ¯ , α , β = X ¯ X ¯ 2 2 = sigmoid X ¯ α β X ¯ 2 2
where
X ¯ ¯ = sigmoid X ¯ α = x ¯ ¯ 1 1 x ¯ ¯ N 1 1 x ¯ ¯ 1 M x ¯ ¯ N M M T
is a N × L hidden-layer output matrix that is a class-independent encoded-attribute (IEA) data set, such that x ¯ ¯ n m = x ¯ ¯ n 1 m , , x ¯ ¯ n L m = sigmoid x ¯ n m α 1 , , sigmoid x ¯ n m α L is the independent encoded form of the original sample x n m and
sigmoid s = 1 1 + exp s , s , +
is the sigmoid activation function. To minimize attribute dependence, the attribute dependence term in the objective function is defined as follows:
L 2 X ¯ , α , β = 2 L ( L 1 ) i = 1 L j = 1 j i L I ( h i , h j ) ,
where
h i = x ¯ ¯ 1 i 1 , , x ¯ ¯ N 1 , i 1 , , x ¯ ¯ 1 i M , , x ¯ ¯ N M , i M T
is the i-th independent encoded attribute, and I ( h i , h j ) is the mutual information between independent encoded attributes h i and h j and i , j 1 , 2 , , L and i j .
The updating rules of α r l and β l r are derived as follows. The second term of the objective function is unrelated to the output-layer weights. Thus, the updating rule mainly depends on the encoding error. The gradient descent method is used to determine the updating rule of β l r . The partial derivative of L X ¯ , α , β with respect to β l r is calculated as follows:
Δ β l r = L X ¯ , α , β β l r = λ L 1 X ¯ , α , β β l r
and then the updating rule of β l r is given as follows
β l r β l r ξ × Δ β l r ,
where ξ > 0 is the learning rate. For the input-layer weight α r l , the updating rule cannot be derived by using the gradient descent method because of the existence of mutual information terms. Here, a new updating strategy based on the Monte Carlo method [25] is designed as follows:
α r l α r l ζ × Δ α r l ,
where ζ > 0 is the learning rate and
Δ α r l = 1 N m = 1 M n = 1 N m L x ¯ n m , α , β
is the approximation of the gradient L X ¯ , α , β α r l .
Based on the IEA data set, mixed-attribute fusion-based NBC (MAF-NBC) determines the class label for a given new sample x = a 1 , , a D 1 , b 1 , , b D 2 as follows. First, the one-hot encoded form of x is expressed as follows:
x ¯ = e 1 1 , , e K 1 1 , , e 1 D 1 , , e K D 1 D 1 , b 1 , , b D 2 ,
where the following.
e k i = 1 , if a i = A k i 0 , if a i A k i , k = 1 , 2 , , K i , i = 1 , 2 , , D 1 .
Second, x ¯ is transformed into the following independent encoded expression:
x ¯ ¯ = x ¯ ¯ 1 , x ¯ ¯ 2 , , x ¯ ¯ L = sigmoid x ¯ α 1 , sigmoid x ¯ α 2 , , sigmoid x ¯ α L
based on the trained ANN with input weight matrix α = α r l R × L . Third, the class label of x is determined according to Equation (1), where the conditional probability is calculated as follows:
P x w m = P x ¯ w m = P x ¯ ¯ w m = l = 1 L P x ¯ ¯ l w m = l = 1 L x ¯ ¯ l p s w m ,
where p s w m is approximated with a normal p.d.f.
f l m s = 1 2 π σ l m exp 1 2 s μ l m σ l m 2 , s , + , m = 1 , 2 , , M , l = 1 , 2 , , L
with the mean value
μ l m = 1 N m n = 1 N m ¯ ¯ x n l m
and standard deviation (std).
σ l m = 1 N m 1 n = 1 N m x ¯ ¯ n l m μ l m 2 .
Here, an in-depth discussion regarding the normal p.d.f.  f l m is given. The sigmoid activation function is used in the designed ANN; thus, the outputs corresponding to each hidden layer’s nodes obey a quasi-normal probability distribution. A visual comparison of the sigmoid activation function and normal probability distribution functions with standard deviations 0.1, 0.5, 1.0, and 2.0, as shown in Figure 1, clearly demonstrates this empirical conclusion.
For independent attributes, the joint probability distribution is the normal probability distribution if marginal probability distributions are normal probability distributions. Thus, the joint probability P x ¯ ¯ w m can be modeled as the product of multiple marginal probabilities P x ¯ ¯ l w m , l = 1 , 2 , , L . For the sake of simplicity, hidden-layer biases are not used in the constructed ANN. The role of hidden-layer biases is to control the bias between the hidden layer’s input and the original point of the sigmoid function. When conducting classification or regression tasks, hidden-layer biases are helpful for the generation of high-performance learners. In MAF-NBC, the ANN is designed to transform one-hot encoded attributes into independent encoded attributes that are expected to have minimum dependence. The reasonable objective function shown in Equation (11) is able to guarantee the generation of independent encoded attributes even though hidden-layer biases are not used in ANN. The following experiments will support the aforementioned discussion and conclusion.

4. Experimental Settings and Results

A series of experiments are conducted in this section to validate the feasibility, rationality, and effectiveness of the proposed mixed-attribute fusion-based naive Bayesian classifier (MAF-NBC). Experiments were conducted using 20 KEEL [26] mixed-attribute data sets. Their characteristics are listed in Table 1, where the Arabic numerals in parentheses represent the numbers of avaliable categorical values corresponding to the discrete attributes. The data processing strategy proposed by Helal and Otero [27] was used to generate the mixed attributes for the data sets marked without ’*’ in Table 1. Then, the equal width discretization method was used to transform the continuous attributes into discrete attributes. All data sets as shown in Table 1 can be downloaded from our BaiDuPan or GitHub online storage space. The experiments were run on a PC equipped with an Intel(R) Quadcore 3.00 GHz i5-9400 CPU and 16 GB of RAM.

4.1. Feasibility Validation of MAF-NBC

A first experiment was performed to validate the feasibility of the MAF-NBC method by checking the convergence of ANN weights by the iterative update process. This experiment was conducted on the representative Vowel data set, which has three discrete attributes and seven continuous attributes. The experimental results are the average values corresponding to 10 independent ANN training. Ten ANNs with 50 hidden-layer nodes were constructed with random weights in the [−1, 1] range, a balance factor λ = 0 . 50 , and learning rates ξ = ζ = 0.01 . Figure 2 depicts the variation trends of the 1-norms for the input layer and output layer weights as the iteration number increased. Figure 3 presents the variation trends of encoding error and attribute dependence as the iteration number increased.
Figure 2 shows that the sums of absolute values corresponding to the input layer and output-layer weights become more and more stable with the number of iterations. In addition, Figure 3 shows that the encoding error and attribute dependence decrease gradually and then remain unchanged. These experimental results reveal that the updating rules of Equations (15) and (16) are effective for the determination of optimal ANN weights. In fact, Equation (17) is the approximation of the expected loss E X ¯ L X ¯ , α , β α r l , where X ¯ is the domain of the objective function L , α , β . Although the gradient of input layer weights cannot be analytically calculated due to the existence of the attribute dependence term, Equation (17) uses the mean of the objective function values corresponding to all samples in the one-hot encoded attribute data set to approximate the gradient. The convergence of output layer weights guarantees the convergence of input layer weights, because the update of α r l depends on the update of β l r . This experiment demonstrates the feasibility of transforming the one-hot encoded attributes into independent encoded attributes.

4.2. Rationality Validation of MAF-NBC

A second experiment was carried out to evaluate the rationality of MAF-NBC in terms of whether the designed ANN can transform the one-hot encoded attributes corresponding to the original mixed attributes into independent encoded attributes. On the representative Page_small data set, an ANN with 50 hidden-layer nodes was constructed with random weights in the [−1,1] interval, a balance factor λ = 0 . 50 , and learning rates ξ = ζ = 0.01 . Ten representative encoded attributes were selected from the hidden-layer outputs of ANN. The dependence between the two encoded attributes was measured with the mutual information, which is calculated with sklearn.metrics.mutual_info_score (https://scikit-learn.org/stable/modules/generated/sklearn.metrics.mutual_info_score.html, accessed on 14 October 2022) package of the scikit-learn machine learning library.
Figure 4 displays a series of heatmaps corresponding to iterations #1, #50, #100, and #150. These heatmaps show the change in attribute dependence as the iteration number increases during the ANN’s training. It can be clearly seen that the dependence between encoded attribute gradually decreases as ANN weights are updated. It indicates that independent attributes can be obtained by transforming the original mixed attributes into the encoded attributes. The experimental results shown in Figure 4 are consistent with the experimental results shown in Figure 3b, where the attribute dependence gradually decreases with the increase in iteration number. This experiment demonstrated that the ANN is able to transform the original mixed attributes into independent encoded attributes for NBC construction.

4.3. Effectiveness Validation of MAF-NBC

A third experiment was performed to evaluate the effectiveness of MAF-NBC. This was performed by comparing the classification performances of MAF-NBC with Dis-NBC, FNBC [3], TAN [4], AODE [5], HNB [6], DFW-NBC [7], CFW-NBC [8], and ICA-NBC [10]. All Bayesian algorithms were implemented using the Python programming language. Twenty KEEL mixed-attribute data sets (described in Table 1) were selected to test the training and testing accuracies of the compared algorithms. For each data set, independent training and testing were conducted 10 times with random data set partitions, i.e., 70% samples to train the different Bayes algorithms and 30% to test their generalization capabilities. All ANNs were constructed with 2 L hidden-layer nodes, initialized with random weights in the [−1, 1] range, and setup with the learning parameters λ = 0 . 50 and ξ = ζ = 0.01 . The mean and standard derivation of the 10 training and testing accuracies are listed for each algorithm in Table 2 and Table 3, respectively.
The training and testing performances of MAF-NBC were statistically validated by comparing with eight other Bayes algorithms on the 20 data sets. For the given significance level of 0.05, the critical difference (CD) value [28] is calculated as follows:
CD = 3.102 × 9 × 9 + 1 6 × 20 2.686 ,
where the number of compared algorithms is nine, and the number of data sets is 20. In Figure 5, an interval of one CD value can be observed to the left and right of the average rank of MAF-NBC. Any algorithm with a rank outside this area is significantly different from MAF-NBC. It is found that the ranks of MAF-NBC corresponding to the training and testing accuracies are obviously smaller than the other algorithms. In addition, the number of wins for MAF-NBC on 20 data sets is at least 20 2 + 1.96 × 20 2 14 for the significance level of 0.05. Furthermore, it can be said that MAF-NBC has significantly improved testing accuracies than Dis-NBC (20 wins), FNBC (20 wins), TAN (20 wins), AODE (20 wins), HNB (20 wins), DFW-NBC (19 wins), CFW-NBC (19 wins), and ICA-NBC (18 wins) under a significance level of 0.10. This indicates that MAF-NBC is significantly better for classification than the other algorithms on the selected data sets. The experimental results and statistical analyses demonstrate the effectiveness of MAF-NBC and indicate that MAF-NBC is a viable method to handle mixed-attribute classification tasks. MAF-NBC does not modify the simple model structure of the NBC and preserves the amount of information of the original mixed-attribute data set as much as possible, because the ANN transforms the original mixed attributes into encoded attributes rather than selecting or extracting independent attributes from the original mixed attributes.

5. Conclusions and Future Works

This paper presented a novel NBC training method for the mixed-attribute data classification problem without continuous attribute discretization and complex Bayesian network structure learning. The original mixed attributes were transformed into a series of continuous attributes with minimum dependence using an autoencoder neural network. To obtain optimal network weights, an effective objective function was designed, and corresponding weight updating rules were derived. The experimental results finally demonstrated improved classification performance for the novel Bayes model in comparison with eight state-of-the-art Bayesian algorithms. The technical advantages of MAF-NBC are four-fold:
  • Preserving the information. The autoencoder neural network transforms mixed attributes into independent encoded attributes, which can also be decoded to restore the original mixed attributes. Thus, useful information from the original data set is conserved as much as possible.
  • Simple model structure. MAF-NBC effectively handles the attribute dependence problem between discrete attributes and continuous attributes by transforming original attributes into encoded attributes. The simple structure of the NBC was not modified by providing a data transformation-oriented preprocessing strategy.
  • Stable training process. The efficient weight updating strategy guarantees the convergence of the autoencoder neural network and thus provided stable data transformation results. In addition, the quasi-normal distribution of independent encoded attributes provided accurate calculations for the joint conditional probability in the NBC.
  • Low computation complexity. A single hidden-layer encoder neural network was used to fuse the original mixed attributes. Training to find optimal network weights has a low computational complexity. As mentioned in Section 3, a shallow learning mechanism was able to satisfy the demand of NBC construction. Hence, it is unnecessary to explore the availability of deep learning for data transformation-oriented NBC training.
In future studies, we plan (1) to implement MAF-NBC in a distributed environment so that it can be used to deal with large-scale mixed-attribute data-classification problems and (2) to utilize the autoencoder neural network to transform mixed attributes into dependent continuous attributes to construct a non-naive Bayesian classifier [29] based on the joint probability density function estimation technique.

Author Contributions

Data curation, formal analysis, and writing—original draft preparation, G.O.; methodology, writing—original draft preparation and review and editing, Y.H.; investigation and writing—review and editing, P.F.-V.; supervision, J.Z.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Scientific Research Foundation of National Natural Science Foundation of China (61972261) and Basic Research Foundations of Shenzhen (JCYJ20210324093609026 and JCYJ20200813091134001).

Institutional Review Board Statement

None.

Informed Consent Statement

None.

Data Availability Statement

The data are available in BaiduPan https://pan.baidu.com/s/17417zF-IlP6lW_ut7lRTqA (accessed on 14 October 2022) with extraction code fcju or GitHub platform https://github.com/ouguiliang110/DataSet_of_MAF-NBC (accessed on 14 October 2022).

Acknowledgments

The authors would like to thank the editors and anonymous reviewers who carefully read the paper and provided valuable suggestions that considerably improved the paper. This paper was recommended by the 6th Asian Conference on Artificial Intelligence Technology (ACAIT 2022) which is the 2022 Top Academic Conference recognized by China Association for Science and Technology (CAST).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wu, X.; Kumar, V.; Quinlan, J.R.; Ghosh, J.; Yang, Q.; Motoda, H.; McLachlan, G.J.; Ng, A.; Liu, B.; Yu, P.S.; et al. Top 10 algorithms in data mining. Knowl. Inf. Syst. 2008, 14, 1–37. [Google Scholar]
  2. Gil-Begue, S.; Bielza, C.; Larrañaga, P. Multi-dimensional Bayesian network classifiers: A survey. Artif. Intell. Rev. 2021, 54, 519–559. [Google Scholar]
  3. John, G.; Langley, P. Estimating continuous distributions in Bayesian classifiers. In Proceedings of the 11th Conference on Uncertainty in Artificial Intelligence, Montreal, QC, Canada, 18–20 August 1995; pp. 1–8. [Google Scholar]
  4. Friedman, N.; Geiger, D.; Goldszmidt, M. Bayesian network classifiers. Mach. Learn. 1997, 29, 131–163. [Google Scholar]
  5. Webb, G.I.; Boughton, J.R.; Wang, Z. Not so naive Bayes: Aggregating one-dependence estimators. Mach. Learn. 2005, 58, 5–24. [Google Scholar]
  6. Jiang, L.; Zhang, H.; Cai, Z. A novel Bayes model: Hidden naive Bayes. IEEE Trans. Knowl. Data Eng. 2009, 21, 1361–1371. [Google Scholar]
  7. Jiang, L.; Li, C.; Wang, S.; Zhang, L. Deep feature weighting for naive Bayes and its application to text classification. Eng. Appl. Artif. Intell. 2016, 52, 26–39. [Google Scholar]
  8. Jiang, L.; Zhang, L.; Li, C.; Wu, J. A correlation-based feature weighting filter for Naive Bayes. IEEE Trans. Knowl. Data Eng. 2018, 31, 201–213. [Google Scholar]
  9. Bressan, M.; Vitria, J. Improving naive Bayes using class-conditional ICA. In Proceedings of the 2022 Ibero-American Conference on Artificial Intelligence, Seville, Spain, 12–15 November 2002; pp. 1–10. [Google Scholar]
  10. Qin, F.; Ren, S.L.; Cheng, Z.K.; Luo, H. Bayes classification model based on ICA. Comput. Eng. Des. 2007, 28, 4873–4877. [Google Scholar]
  11. Fan, L.; Poh, K.L. A comparative study of PCA, ICA and class-conditional ICA for naive Bayes classifier. In Proceedings of the 2007 International Work-Conference on Artificial Neural Networks, Porto, Portugal, 9–13 September 2007; pp. 16–22. [Google Scholar]
  12. Jayanthi, S.K.; Sasikala, S. Naive Bayesian classifier and PCA for web link spam detection. Comput. Sci. Telecommun. 2014, 41, 3–15. [Google Scholar]
  13. Zhang, B.; Liu, Z.; Jia, Y.; Ren, J.; Zhao, X. Network intrusion detection method based on PCA and Bayes algorithm. Secur. Commun. Netw. 2018, 2018, 1914980. [Google Scholar]
  14. Kurgan, L.A.; Cios, K.J. CAIM discretization algorithm. IEEE Trans. Knowl. Data Eng. 2004, 16, 145–153. [Google Scholar]
  15. Li, M.; Zhao, S.; Chen, S.; Huang, D. A new method of online fault diagnosis based on incremental continuous attribute naive Bayesian. In Proceedings of the 2018 IEEE International Conference on Prognostics and Health Management, Seattle, WA, USA, 11–13 June 2018; pp. 1–7. [Google Scholar]
  16. Scanagatta, M.; Salmerón, A.; Stella, F. A survey on Bayesian network structure learning from data. Prog. Artif. Intell. 2019, 8, 425–439. [Google Scholar]
  17. Parzen, E. On estimation of a probability density function and mode. Ann. Math. Stat. 1962, 33, 1065–1076. [Google Scholar]
  18. Yang, Y.; Webb, G.I. A comparative study of discretization methods for naive-Bayes classifiers. In Proceedings of the 2002 Pacific Rim Knowledge Acquisition Workshop, Tokyo, Japan, 18–22 August 2002; pp. 159–173. [Google Scholar]
  19. Yang, Y.; Webb, G.I. On why discretization works for naive-Bayes classifiers. In Proceedings of the 2003 Australasian Joint Conference on Artificial Intelligence, Perth, Australia, 3–5 December 2003; pp. 440–452. [Google Scholar]
  20. Yang, Y.; Webb, G.I. Discretization for naive-Bayes learning: Managing discretization bias and variance. Mach. Learn. 2009, 74, 39–74. [Google Scholar]
  21. He, Y.L.; Ye, X.; Haung, D.F.; Fournier-Viger, P.; Huang, J.Z. A hybrid method to measure distribution consistency of mixed attribute data sets. IEEE Trans. Artif. Intell. 2022. [Google Scholar] [CrossRef]
  22. Qi, Z.; Wang, B.; Tian, Y.; Zhang, P. When ensemble learning meets deep learning: A new deep support vector machine for classification. Knowl.-Based Syst. 2016, 107, 54–60. [Google Scholar]
  23. Feng, J.; Yu, Y.; Zhou, Z.H. Multi-layered gradient boosting decision trees. In Proceedings of the 32nd Conference on Neural Information Processing Systems, Montreal, QC, Canada, 3–8 December 2018; pp. 1–11. [Google Scholar]
  24. Guo, M.; Fei, J.; Meng, Y. Deep nearest neighbor website fingerprinting attack technology. Secur. Commun. Netw. 2021, 2021, 5399816. [Google Scholar]
  25. De Freitas, J.F.; Niranjan, M.; Gee, A.H.; Doucet, A. Sequential Monte Carlo methods to train neural network models. Neural Comput. 2000, 12, 955–993. [Google Scholar]
  26. Alcalá-Fdez, J.; Fernández, A.; Luengo, J.; Derrac, J.; García, S.; Sánchez, L.; Herrera, F. KEEL data-mining software tool: Data set repository, integration of algorithms and experimental analysis framework. J.-Mult.-Valued Log. Soft Comput. 2011, 17, 255–287. [Google Scholar]
  27. Helal, A.; Otero, F.E. A mixed attribute approach in ant-miner classification rule discovery algorithm. In Proceedings of the Genetic and Evolutionary Computation Conference, Denver, CO, USA, 20–24 July 2016; pp. 13–20. [Google Scholar]
  28. Demšar, J. Statistical comparisons of classifiers over multiple data sets. J. Mach. Learn. Res. 2006, 7, 1–30. [Google Scholar]
  29. Wang, X.Z.; He, Y.L.; Wang, D.D. Non-naive Bayesian classifiers for classification problems with continuous attributes. IEEE Trans. Cybern. 2013, 44, 21–39. [Google Scholar]
Figure 1. Graphical comparison between sigmoid activation and normal distribution. (a) Sigmoid activation function. (b) Normal probability distribution functions.
Figure 1. Graphical comparison between sigmoid activation and normal distribution. (a) Sigmoid activation function. (b) Normal probability distribution functions.
Applsci 12 10443 g001
Figure 2. Convergence of the ANN’s weights. (a) Input-layer weights. (b) Output-layer weights.
Figure 2. Convergence of the ANN’s weights. (a) Input-layer weights. (b) Output-layer weights.
Applsci 12 10443 g002
Figure 3. Convergence of the ANN’s objective function. (a) Encoding error. (b) Attribute dependence.
Figure 3. Convergence of the ANN’s objective function. (a) Encoding error. (b) Attribute dependence.
Applsci 12 10443 g003
Figure 4. Dependence change of the encoded continuous attributes.
Figure 4. Dependence change of the encoded continuous attributes.
Applsci 12 10443 g004
Figure 5. CD diagrams corresponding to comparisons in Table 2 and Table 3. (a) CD diagram of training accuracies. (b) CD diagram of testing accuracies.
Figure 5. CD diagrams corresponding to comparisons in Table 2 and Table 3. (a) CD diagram of training accuracies. (b) CD diagram of testing accuracies.
Applsci 12 10443 g005
Table 1. Details of 20 mixed-attribute data sets.
Table 1. Details of 20 mixed-attribute data sets.
Data setsAbalone_3_4Abalone_9_10Adult *Band *
Number of samples42310735000540
Number of continuous attributes66918
Number of discrete attributes2 (2, 4)2 (2, 4)5 (5, 7, 5, 4, 2)5 (3, 5, 2, 3, 3)
Number of classes2222
Data setsBdBpHeartIonosphere
Number of samples569198270351
Number of continuous attributes2325925
Number of discrete attributes7 (4, 6, 5, 3, 2, 4, 3)8 (6, 3, 5, 4, 5, 3, 4, 2)4 (4, 3, 3, 4)7 (4, 6, 5, 3, 4, 3, 5)
Number of classes2222
Data setsLibrasPage blocksParkinsonsRing
Number of samples360547195500
Number of continuous attributes8571715
Number of discrete attributes15 (4, 5, 7, 5, 8, 4, 5, 6, 4, 5, 6, 7, 3, 4, 5)3 (3, 5, 4)5 (4, 5, 6, 5, 3)5 (3, 4, 5, 3, 4)
Number of classes15522
Data setsSegmentSonarSPECTFVehicle
Number of samples2310208267846
Number of continuous attributes1450366
Number of discrete attributes5 (3, 6, 5, 4, 3)10 (6, 4, 3, 5, 6, 7, 5, 3, 4, 6)8 (4, 6, 5, 5, 3, 6, 7, 4)2 (4, 3)
Number of classes7224
Data setsVowelWineWineQRWineQW
Number of samples5281781599489
Number of continuous attributes71088
Number of discrete attributes3 (3, 3, 4)3 (4, 5, 4)3 (4, 4, 3)3 (3, 3, 3)
Number of classes11366
Table 2. Comparison of the training accuracies of MAF-NBC, Dis-NBC, FNB, TAN, AODE, HNB, DFW-NBC, CFW-NBC, and ICA-NBC.
Table 2. Comparison of the training accuracies of MAF-NBC, Dis-NBC, FNB, TAN, AODE, HNB, DFW-NBC, CFW-NBC, and ICA-NBC.
Data SetMAF-NBCDis-NBCFNBCTANAODEHNBDFW-NBCCFW-NBCICA-NBC
Abalone_3_40.803 ± 0.0360.721 ± 0.0320.736 ± 0.0260.767 ± 0.0380.755 ± 0.0320.744 ± 0.0290.747 ± 0.0180.755 ± 0.0170.781 ± 0.045
Abalone_9_100.607 ± 0.0290.597 ± 0.0120.592 ± 0.0210.595 ± 0.0250.596 ± 0.0240.587 ± 0.0410.617 ± 0.0080.605 ± 0.0140.619 ± 0.017
Adult0.870 ± 0.0130.782 ± 0.0140.792 ± 0.0150.807 ± 0.0130.829 ± 0.0080.811 ± 0.0170.805 ± 0.0110.813 ± 0.0060.851 ± 0.015
Band0.718 ± 0.0270.669 ± 0.0280.637 ± 0.0340.671 ± 0.0350.663 ± 0.0240.681 ± 0.0280.703 ± 0.0150.718 ± 0.0170.714 ± 0.016
Bd0.991 ± 0.0020.919 ± 0.0040.938 ± 0.0100.961 ± 0.0090.959 ± 0.0080.953 ± 0.0110.959 ± 0.0070.970 ± 0.0040.983 ± 0.007
Bp0.852 ± 0.0220.788 ± 0.0230.793 ± 0.0270.803 ± 0.0290.801 ± 0.0340.791 ± 0.0410.841 ± 0.0360.824 ± 0.0330.837 ± 0.027
Heart0.924 ± 0.0140.699 ± 0.0130.824 ± 0.0170.858 ± 0.0150.849 ± 0.0210.867 ± 0.0290.886 ± 0.0130.905 ± 0.0150.901 ± 0.017
Ionosphere0.981 ± 0.0050.901 ± 0.0060.917 ± 0.0110.934 ± 0.0090.932 ± 0.0130.926 ± 0.0210.928 ± 0.0120.941 ± 0.0150.974 ± 0.004
Libras0.905 ± 0.0070.938 ± 0.0150.857 ± 0.0190.899 ± 0.0150.847 ± 0.0160.864 ± 0.0150.911 ± 0.0160.910 ± 0.0150.908 ± 0.011
Page_small0.952 ± 0.0100.915 ± 0.0050.911 ± 0.0070.909 ± 0.0180.919 ± 0.0140.914 ± 0.0170.931 ± 0.0060.933 ± 0.0080.937 ± 0.007
Parkinsons0.872 ± 0.0290.782 ± 0.0210.801 ± 0.0220.808 ± 0.0250.814 ± 0.0270.821 ± 0.0210.827 ± 0.0220.841 ± 0.0250.852 ± 0.024
Ring0.937 ± 0.0150.801 ± 0.0170.836 ± 0.0150.855 ± 0.0280.869 ± 0.0240.858 ± 0.0270.916 ± 0.0100.948 ± 0.0060.956 ± 0.008
Segment0.940 ± 0.0070.878 ± 0.0100.889 ± 0.0190.897 ± 0.0160.895 ± 0.0140.901 ± 0.0160.928 ± 0.0060.945 ± 0.0090.959 ± 0.005
Sonar0.865 ± 0.0190.823 ± 0.0290.834 ± 0.0270.839 ± 0.0250.844 ± 0.0180.840 ± 0.0250.841 ± 0.0270.847 ± 0.0290.851 ± 0.024
Spectf0.908 ± 0.0110.768 ± 0.0230.771 ± 0.0170.783 ± 0.0190.779 ± 0.0150.756 ± 0.0290.779 ± 0.0160.789 ± 0.0140.883 ± 0.019
Vehicle0.771 ± 0.0160.537 ± 0.0050.544 ± 0.0110.576 ± 0.0160.589 ± 0.0190.567 ± 0.0260.604 ± 0.0220.611 ± 0.0210.705 ± 0.017
Vowel0.869 ± 0.0170.829 ± 0.0210.839 ± 0.0170.836 ± 0.0190.840 ± 0.0160.847 ± 0.0190.843 ± 0.0210.848 ± 0.0160.833 ± 0.023
Wine0.999 ± 0.0010.963 ± 0.0210.967 ± 0.0170.975 ± 0.0060.981 ± 0.0040.968 ± 0.0070.999 ± 0.0010.999 ± 0.0010.968 ± 0.019
WineQR0.738 ± 0.0230.563 ± 0.0140.578 ± 0.0340.580 ± 0.0330.591 ± 0.0310.602 ± 0.0370.638 ± 0.0200.717 ± 0.0190.688 ± 0.036
WineQW0.584 ± 0.0280.487 ± 0.0140.524 ± 0.0160.537 ± 0.0240.533 ± 0.0160.541 ± 0.0210.551 ± 0.0280.561 ± 0.0250.577 ± 0.021
Average0.854 ± 0.0170.768 ± 0.0160.779 ± 0.0190.795 ± 0.0210.794 ± 0.0160.792 ± 0.0240.813 ± 0.0160.824 ± 0.0150.839 ± 0.018
Table 3. Comparison of the testing accuracies of MAF-NBC, Dis-NBC, FNB, TAN, AODE, HNB, DFW-NBC, CFW-NBC, and ICA-NBC.
Table 3. Comparison of the testing accuracies of MAF-NBC, Dis-NBC, FNB, TAN, AODE, HNB, DFW-NBC, CFW-NBC, and ICA-NBC.
Data SetMAF-NBCDis-NBCFNBCTANAODEHNBDFW-NBCCFW-NBCICA-NBC
Abalone_3_40.759 ± 0.0300.685 ± 0.0420.701 ± 0.0490.742 ± 0.0320.733 ± 0.0370.728 ± 0.0290.723 ± 0.0310.751 ± 0.0270.755 ± 0.043
Abalone_9_100.603 ± 0.0280.587 ± 0.0240.588 ± 0.0260.587 ± 0.0270.589 ± 0.0280.581 ± 0.0300.589 ± 0.0310.582 ± 0.0310.602 ± 0.032
Adult0.842 ± 0.0140.795 ± 0.0110.788 ± 0.0190.801 ± 0.0250.817 ± 0.0120.807 ± 0.0270.804 ± 0.0120.811 ± 0.0140.824 ± 0.009
Band0.647 ± 0.0450.606 ± 0.0280.601 ± 0.0280.626 ± 0.0360.618 ± 0.0380.609 ± 0.0310.619 ± 0.0340.630 ± 0.0360.619 ± 0.041
Bd0.973 ± 0.0110.915 ± 0.0170.921 ± 0.0130.956 ± 0.0120.950 ± 0.0100.946 ± 0.0140.933 ± 0.0150.944 ± 0.0160.948 ± 0.014
Bp0.764 ± 0.0400.739 ± 0.0470.746 ± 0.0390.740 ± 0.0570.739 ± 0.0520.737 ± 0.0650.755 ± 0.0360.747 ± 0.0410.746 ± 0.049
Heart0.876 ± 0.0280.703 ± 0.0220.711 ± 0.0250.739 ± 0.0370.728 ± 0.0340.789 ± 0.0390.859 ± 0.0320.864 ± 0.0350.851 ± 0.038
Ionosphere0.947 ± 0.0270.894 ± 0.0240.906 ± 0.0210.917 ± 0.0170.907 ± 0.0250.901 ± 0.0270.923 ± 0.0290.927 ± 0.0200.931 ± 0.026
Libras0.778 ± 0.0390.743 ± 0.0470.756 ± 0.0580.751 ± 0.0210.749 ± 0.0480.748 ± 0.0390.759 ± 0.0360.764 ± 0.0390.759 ± 0.038
Page_small0.901 ± 0.0190.896 ± 0.0150.889 ± 0.0130.891 ± 0.0160.881 ± 0.0190.898 ± 0.0170.912 ± 0.0150.913 ± 0.0110.899 ± 0.012
Parkinsons0.826 ± 0.0420.756 ± 0.0430.769 ± 0.0490.779 ± 0.0590.771 ± 0.0470.782 ± 0.0390.786 ± 0.0340.807 ± 0.0440.779 ± 0.030
Ring0.928 ± 0.0170.781 ± 0.0410.831 ± 0.0380.849 ± 0.0490.834 ± 0.0410.879 ± 0.0490.870 ± 0.0280.910 ± 0.0110.925 ± 0.016
Segment0.913 ± 0.0160.869 ± 0.0190.874 ± 0.0140.886 ± 0.0180.895 ± 0.0160.889 ± 0.0200.899 ± 0.0150.906 ± 0.0130.922 ± 0.010
Sonar0.816 ± 0.0360.758 ± 0.0460.769 ± 0.0530.779 ± 0.0380.806 ± 0.0340.787 ± 0.0230.796 ± 0.0360.801 ± 0.0420.837 ± 0.029
Spectf0.823 ± 0.0230.716 ± 0.0370.727 ± 0.0270.778 ± 0.0410.794 ± 0.0300.757 ± 0.0410.748 ± 0.0350.757 ± 0.0300.801 ± 0.029
Vehicle0.611 ± 0.0290.483 ± 0.0370.506 ± 0.0390.529 ± 0.0320.520 ± 0.0490.518 ± 0.0540.537 ± 0.0290.549 ± 0.0320.589 ± 0.028
Vowel0.859 ± 0.0230.781 ± 0.0570.781 ± 0.0520.799 ± 0.0580.795 ± 0.0570.801 ± 0.0330.811 ± 0.0290.821 ± 0.0240.816 ± 0.027
Wine0.961 ± 0.0190.928 ± 0.0240.926 ± 0.0290.937 ± 0.0270.946 ± 0.0210.939 ± 0.0260.958 ± 0.0170.945 ± 0.0220.931 ± 0.016
WineQR0.637 ± 0.0180.489 ± 0.0250.509 ± 0.0380.519 ± 0.0370.528 ± 0.0340.524 ± 0.0310.563 ± 0.0280.587 ± 0.0300.559 ± 0.041
WineQW0.534 ± 0.0360.477 ± 0.0430.503 ± 0.0330.508 ± 0.0380.501 ± 0.0410.509 ± 0.0290.508 ± 0.0410.515 ± 0.0280.529 ± 0.039
Average0.800 ± 0.0270.730 ± 0.0320.740 ± 0.0330.756 ± 0.0340.755 ± 0.0340.756 ± 0.0330.768 ± 0.0270.777 ± 0.0270.781 ± 0.028
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Ou, G.; He, Y.; Fournier-Viger, P.; Huang, J.Z. A Novel Mixed-Attribute Fusion-Based Naive Bayesian Classifier. Appl. Sci. 2022, 12, 10443. https://doi.org/10.3390/app122010443

AMA Style

Ou G, He Y, Fournier-Viger P, Huang JZ. A Novel Mixed-Attribute Fusion-Based Naive Bayesian Classifier. Applied Sciences. 2022; 12(20):10443. https://doi.org/10.3390/app122010443

Chicago/Turabian Style

Ou, Guiliang, Yulin He, Philippe Fournier-Viger, and Joshua Zhexue Huang. 2022. "A Novel Mixed-Attribute Fusion-Based Naive Bayesian Classifier" Applied Sciences 12, no. 20: 10443. https://doi.org/10.3390/app122010443

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop