Next Article in Journal
Unstable Points, Ergodicity and Born’s Rule in 2d Bohmian Systems
Next Article in Special Issue
Network Analytics Enabled by Generating a Pool of Network Variants from Noisy Data
Previous Article in Journal
Low Noise Opto-Electro-Mechanical Modulator for RF-to-Optical Transduction in Quantum Communications
Previous Article in Special Issue
On the Analysis of Regularized Fuzzy Systems of Uncertain Differential Equations
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Upgrading the Fusion of Imprecise Classifiers

by
Serafín Moral-García
,
María D. Benítez
and
Joaquín Abellán
*
Department of Computer Science and Artificial Intelligence, University of Granada, 18012 Granada, Spain
*
Author to whom correspondence should be addressed.
Entropy 2023, 25(7), 1088; https://doi.org/10.3390/e25071088
Submission received: 21 June 2023 / Revised: 12 July 2023 / Accepted: 18 July 2023 / Published: 19 July 2023
(This article belongs to the Special Issue Selected Featured Papers from Entropy Editorial Board Members)

Abstract

:
Imprecise classification is a relatively new task within Machine Learning. The difference with standard classification is that not only is one state of the variable under study determined, a set of states that do not have enough information against them and cannot be ruled out is determined as well. For imprecise classification, a mode called an Imprecise Credal Decision Tree (ICDT) that uses imprecise probabilities and maximum of entropy as the information measure has been presented. A difficult and interesting task is to show how to combine this type of imprecise classifiers. A procedure based on the minimum level of dominance has been presented; though it represents a very strong method of combining, it has the drawback of an important risk of possible erroneous prediction. In this research, we use the second-best theory to argue that the aforementioned type of combination can be improved through a new procedure built by relaxing the constraints. The new procedure is compared with the original one in an experimental study on a large set of datasets, and shows improvement.

1. Introduction

The standard classification task tries to predict one state of a class variable when a new item or instance appears. A classifier normally learns from a set of data. In situations where there is not enough information to point out a unique state of the class variable, Imprecise classification arises. In this case, a set of states of that variable are predicted, which are the states that have no clear information against them. It can be said that Imprecise Classification discards some states and considers the ones that cannot be discarded via the available information.
Credal Decision Tree (CDT) [1,2] is a single decision tree that uses imprecise probability models and uncertainty-based information measures in the tree-building process, specifically, the maximum of entropy measure. This classifier has been used, similar to other comparable models, in many ensemble procedures [3,4], and has been adapted for Imprecise Classification (ICDT) [5]. However, ICDT was not the first imprecise classifier presented in the literature; the Naive Credal Classifier (NCC) [6] was the first imprecise classifier presented. It combines the naive assumption (i.e., all the attributes are independent given the class variable) and the Imprecise Dirichlet Model (IDM) to produce an imprecise classification. This NCC procedure provides worse results than the ICDT model [5], and we do not use it here.
An imprecise classifier might obtain a set of states of the class variable, often called the set of non-dominated states. It is constituted by those class states for which there is no other “better” choice according to an established criterion, usually known as the dominance criterion. Intuitively, an Imprecise Classification evaluation measure must consider whether the real class value belongs to the non-dominated states set as well as the precision of the predicted set of class values, which is measured based on its cardinality.
Imprecise predictions tend to consist of sets of states. Therefore, it is not trivial to combine multiple imprecise predictions. Until recently there was no technique for this. If the predictions are not properly combined, then it is highly likely that the ensemble will perform no better than a unique classifier, as excessive information reduction may result.
The first ensemble of imprecise classifiers was recently presented by Moral-García et al. [7]. This ensemble is based on a bagging scheme [8], which has obtained satisfactory performance in precise classification, especially when used with CDTs. For these reasons a bagging scheme with ICDTs was introduced in [7]. The proposed technique for combining imprecise predictions aims to maximize the precision of the bagging scheme. If the combination procedure is too conservative, then little information will be obtained, and the performance of a single classifier might be worsened. The procedure of [7] only considers the class values with the minimum level of dominance. Hence, it could be considered that such a procedure assumes too much risk, although it achieves good results when compared with the simple ICDT model.
In a procedure for combining imprecise classifiers, we consider the set of states belonging to the set of the minimum level of dominance as the winner states. In the optimal case of a clear set of winners states, i.e., if the difference with respect to the others is relatively large, the minimum level of dominance could be considered a very good procedure. The problem is that in the majority of real cases this does not occur. Consequently, the risk that we must assume can produce bad informative results. This information is obtained via the measure that we need to apply to quantify the performance.
Here, we consider the trade-off between risk and success in the procedure proposed in [7] to be less than ideal. Thus, our starting point is to reduce the risk that we must assume in a procedure to combine imprecise classifiers. This can be considered as a reduction of the constraints taken into account to improve the level of information that such a type of procedure should offer. When the optimal conditions cannot be satisfied, we can apply the theory of the second-best [9], which was originally presented in the area of economics.
Considering the reasons above exposed, we present a new procedure for combining imprecise classifiers that reduces the constraint of the minimum level of dominance, i.e., we consider a scenario in which the optimal situation is not normally attained. In the new procedure, we add a parameter to control the degree of risk we are willing to take. If we decrease that degree, the risk is higher, while the information can be higher as well. Hence, we find it interesting to consider a trade-off between risk and information. This is obtained by relaxing the minimum level of dominance considered by the original method through a parameter. Greater values of that parameter provide less information; however, they provide less risk as well, because a larger set of class values can be obtained.
The new procedure presented here can provide a greater amount of information via the measure for this aim presented in [5]. In this paper, exhaustive experimentation on 34 different datasets is carried out. All these datasets have at least three states of the class variables to ensure that our research on Imprecise Classification makes more sense. We compare the procedure presented in [7] with our new proposal. To compare the results, we apply the measure from [5]. This experimentation shows that the new procedure, which is more cautious in its predictions, obtains clearly better results based on this measure. To reinforce this assertion, our statistical tests show a performance improvement in favor of the new procedure with a strong level of significance.
The rest of this paper is organized as follows: in Section 2, the Imprecise Credal Decision Tree algorithm and the existing bagging scheme for Imprecise Classification are described; Section 2.4 introduces our proposed procedure of combining imprecise classifiers; and the experimental study carried out in this work is detailed in Section 4. Finally, Section 5 provides conclusions and ideas for future work.

2. Background

Let C be the class variable and let c 1 , c 2 , , c K be its set of possible values.

2.1. Imprecise Credal Decision Tree

The Imprecise Credal Decision Tree (ICDT) algorithm introduced in [5] adapts the Credal Decision Tree method for Imprecise Classification. The tree-building process of both methods is identical.
Let D be the dataset corresponding to a certain node. Suppose that such a dataset contain N D instances. Let n D ( c j ) be the number of instances in D that satisfy C = c j , j = 1 , 2 , , K . In order to represent the uncertainty-based information about C in D , ICDT uses the Imprecise Dirichlet Model (IDM) [10], a formal imprecise probability model based on probability intervals. Specifically, the IDM predicts that the probability that C takes each possible value c j belongs to the following interval:
I D ( c j ) = n D ( c j ) N D + s , n D ( c j ) + s N D + s , j = 1 , 2 , , K ,
where s > 0 is a given parameter of the model that indicates the estimated degree of imprecision in the data. This set of intervals yields the following credal set (a credal set being a convex and closed set of probability distributions) on C [11]:
P D ( C ) = p P ( C ) p ( c j ) I D ( c j ) , j = 1 , 2 , , K ,
where P ( C ) is the set of all probability distributions on C. ICDT quantifies the uncertainty about C through the maximum entropy on P D ( C ) :
S * P D ( C ) = max p P D ( C ) S ( p ) ,
where S ( p ) is the Shannon entropy [12] of the probability distribution p, determined by
S ( p ) = j = 1 K p ( c j ) log 2 p ( c j ) .
It must be remarked that the maximum entropy is a well-established uncertainty measure on credal sets, and satisfies the essential mathematical properties [13].
Let X be an attribute taking the possible values x 1 , , x t . The split criterion of ICDT, called the Imprecise Information Gain (IIG), is defined as follows:
I I G D ( C ) = S * P D ( C ) i = 1 t p D ( x j ) S * P D ( C X = x i ) ,
where p D ( x j ) is the probability of X = x j according to the probability distribution that leads to the maximum entropy on the IDM credal set on X on D , while P D ( C X = x i ) is the IDM credal set on C on the subset of D constituted by those instances such that X = x i , i = 1 , 2 , , t .
ICDT differs from CDT in the criterion used to make a prediction at a leaf node. While CDT uses majority vote, ICDT applies a dominance criterion to the IDM probability intervals to obtain the set of non-dominated states at a leaf node.
Let L be a leaf node; let N L be the total number of instances in L and let n L ( c j ) be the number of instances in L for which C = c j , j = 1 , 2 , , K . Then, we have the following set of IDM probability intervals at L :
n L ( c j ) N L + s , n L ( c j ) + s N L + s , j = 1 , 2 , , K .
ICDT uses the stochastic dominance criterion on this set of intervals, which asserts that a class value c j dominates another one c k if and only if
n L ( c j ) N L + s n L ( c k ) + s N L + s n L ( c j ) n L ( c k ) + s .
According to the results proved in [14], stochastic dominance is the well-established dominance criterion for IDM probability intervals.

2.2. Bagging of Credal Decision Trees for Imprecise Classification

Thus far, the only combined method for imprecise classification is called the Bagging of Imprecise Credal Decision Trees (Bagging-ICDT) method [7]. The idea of this algorithm is similar to the bagging scheme for standard classification. For each base classifier, a bootstrapped sample of the training set is first chosen; then, using the selected sample and ICDT, an Imprecise Classification model is learned.
The key issue with Bagging-ICDT is combining the predictions of the base imprecise classifiers. We remark here that this is not a trivial question, as imprecise classifiers predict a set of non-dominated states; in fact, there are multiple ways of combining multiple imprecise predictions, as certain classifiers may predict that a class value will be dominated while others predict it will be non-dominated. The fundamental point is to find a trade-off between information and risk, where the term information indicates the precision of the prediction, i.e, the number of non-dominated states set, and the term risk refers to the probability that the real class value does not belong to the set of non-dominated states. Obviously, a high level of information gives rise to a high level of risk.
The combination technique proposed in [7] tries to ensure that the ensemble scheme is as informative as possible, though it implies a higher risk of misclassification. In this approach, to classify an instance, the number of classifiers that predict such a state as dominated is counted for each class state. The set of non-dominated states predicted by Bagging-ICDT is composed of those class values predicted as dominated by the minimum number of classifiers.
Figure 1 summarizes the Bagging-ICDT method proposed in [7].
The problem with this procedure is that it is possible that the set of states with the minimum level of dominance has support very close to one of the other states. In this case, the level of risk is very high. This situation appears in the majority of real cases. Hence, the risk that we must assume can produce bad informative results. We find it interesting to consider a trade-off between risk and information by relaxing the constraint represented by the minimum level of dominance.

2.3. Evaluation Metrics for Imprecise Classification

To evaluate the performance of imprecise classifiers, the MIC evaluation measure [5] can be employed. This metric takes into account the possibility that in certain situations the errors may have different degrees of importance. Here, consider the same level of importance for all errors, as the different weights of each error must generally be quantified by experts in a particular area.
MIC is defined as follows:
M I C = 1 N t s t i : S u c c e s s log U i K + 1 K 1 i : E r r o r log K ,
where N t s t is the number of test instances, K is the number of class states, and U i is the predicted set of non-dominated states for the i-th test instance, i = 1 , 2 , , N _ t s t .
It is obvious that the higher the value of MIC, the better the performance. We can observe that the optimal value of MIC, which is reached when all the predictions are precise and correct, is equal to log 1 k = log k . Moreover, when it is verified that U i = k , i = 1 , , N T e s t , i.e, when the imprecise classifier always predicts all possible class values as non-dominated states for a given instance, the value of M I C is equal to 0. This is intuitively more correct, as in these cases the classifier is not informative.
We find that this measure is the better one to use, as it does not have the problems presented by others, as can be seen in [7].

2.4. The Second-Best Theory

The theory of the second-best [9] was first presented in the area of economics, and essentially stands as follows:
“It concerns the situation when one or more optimality conditions cannot be satisfied: if one optimality condition in an economic model cannot be satisfied, then the next-best solution might involve changing other variables away from the values that would otherwise be optimal. Politically, the theory implies that if it is infeasible to remove a particular market distortion, introducing one or more additional market distortions in an interdependent market may partially counteract the first, and lead to a more efficient outcome.”
When a market distortion appears or is introduced (i.e., optimal conditions do not exist), then the application of the best approach for those conditions, known as the second-best (i.e., not optimal) can produce clear advantages. This is known as the second-best equilibrium.

3. Application of the Second-Best Theory to the Combination of Imprecise Classifiers

In our case, the trade-off between risk and accuracy when combining the results of different imprecise classifiers can produce a similar situation. As said above, the difference between the set of winners and the rest of the states may not be large enough to consider that set of winners to be a good result. We have found that this excessive risk can produce a decrease in accuracy and a clear decrease in information via the MIC measure, which takes into account the errors in the results in an important way. Hence, we think that it is necessary to limit the risk in favor of accuracy. This is where we consider the second-best set of non-dominated states.
Based on the above assessments, we consider the need to add a constraint for rejecting states with similar information in the set of non-dominated states during the combination process. We think that at least those states very close to the one selected in the original method must be taken into account. We illustrate this issue via the following example.
Example 1. 
Consider a case where the class variable has five states { c 1 , c 2 , , c 5 } and a bagging procedure is applied using the standard size of 100 trees. To simplify, we can suppose that c 1 has the minimum number of votes against; in this case that value is 50 (dominated for 50 trees). However, c 2 has 52 votes against, while c 3 , c 4 , and c 5 each have 80 votes against. The original method of combining the minimum dominated states outputs the set { c 1 } , which represents excessive risk when taking the real situation into account.
In the resulting situation, the set we consider the optimal result is clearly { c 1 , c 2 } because of its small difference in votes against with respect to the set { c 1 } .
We can analyze possible situations associated with the above numbers via the MIC measure. We consider extreme situations around the number of votes against in concordance with those numbers (the example numbers cannot be the most extreme ones, but they are very close to the extreme possible values) in a bagging procedure of 100 trees to calculate the maximum and minimum possible degrees of information obtained via the MIC measure.
1.
The result is { c 1 } : the best situation (B1) is 50 success, i.e., 50 trees where c 1 is the only state non-dominated and 50 errors with that result, i.e., the set of non-dominated values is in the set { c 2 , , c 5 } . By this, we mean that with these values the possible real state is always accurate when the resulting set is { c 1 } , i.e., no other option can be reached. For the cases where the result is, for example, the set { c 1 , c 3 } , the real state associated with the instance can be c 3 . The worst situation (W1) is two trees, where c 1 is the only non-dominated state (because those where c 2 is non-dominated could coincide with the 48 where c 1 is non-dominated as well), and 98 errors, i.e., the set of non-dominated values is in the set { c 2 , , c 5 } .
2.
The result is { c 1 , c 2 } : the best situation (B2) is 98 success, i.e., 50 trees where c 1 and c 2 are in the set of non-dominated states but never coincide, and two errors, i.e., the set of non-dominated values of those two trees is in the set { c 3 , c 4 , c 5 } . The worst situation (W2) is 50 trees where c 1 is one of non-dominated states and c 2 is non-dominated by the same trees, and 50 errors, i.e., the set of non-dominated values of those 50 trees is in the set { c 3 , c 4 , c 5 } .
The idea here is to think about the possible real value of the class variable associated with an instance while considering the numbers discussed above. Using the MIC measure, we can obtain the following values:
Using MIC measure (7):
M I C = 1 N t s t i : S u c c e s s log U i K + 1 K 1 i : E r r o r log K
The values compatible with the above situations are
  • M I C ( A 1 ) = 1 100 ( 50 log ( 1 5 ) + 50 4 log ( 5 ) ) = 1 , 006
  • M I C ( W 1 ) = 1 100 ( 2 log ( 1 5 ) + 98 4 log ( 5 ) ) = 0 , 413
  • M I C ( A 2 ) = 1 100 ( 98 log ( 2 5 ) + 2 4 log ( 5 ) ) = 0 , 906
  • M I C ( W 2 ) = 1 100 ( 50 log ( 2 5 ) + 50 4 log ( 5 ) ) = 0 , 659 .
We can observe that from one result to the other the upper value number is reduced by around 10 % (from 1 , 006 to 0 , 906 ), whereas the lower value is increased by around 60 % (from 0 , 413 to 0 , 659 ). In terms of bets, the upper value could be considered as the maximum win (greater values imply the possibility of a greater win) and the lower one as complement of the maximum loss or maximum risk (greater values imply lower levels of risk).
In this case, for a decrease in the possibility of the maximum win of 10 % , the risk can be reduced by around 60 % . Hence, in our opinion, the more conservative strategy of considering the set { c 1 , c 2 } in stead of { c 1 } makes more sense for predicting the real value of the class variable.
In Example 1, it can be seen that the use of the second-best theory can help us to obtain an increase of information expressed in form of extreme values of gains. We can consider that we are introducing the constraint of not taking on excessive risk, even this results in a reduced maximum probability of success. Clearly, in this example, considering the second-best issue { c 1 , c 2 } instead of { c 1 } notably reduces the likelihood of loss.
For our combination procedure, the problem is now how to quantify the optimal difference among the best states to determine the maximum difference of votes against in terms of percentage. If we call this percentage γ % , then the method explained in Section 2.2 is modified only in terms of the procedure provided in Figure 2.
Here, the parameter γ represents the degree of risk we are willing to take. If we decrease its value, the risk is higher, though the information can be higher as well. We consider it important to use a trade-off between risk and information. Greater values of γ provide less information with less risk because a larger set of class values can be obtained.

4. Experimentation

4.1. Experimental Settings

For experimentation, we used the implementation in the Weka software [15] (https://en.wikipedia.org/wiki/Weka_(software)) for ICDT, as used in [5]. The structures provided in Weka for the bagging scheme were utilized to add all the necessary methods for implementing the bagging of our ICDT procedures. We call the method presented in [7] Bagging-ICDT and the new method presented here with a parameter γ = 5 %  Bagging-ICDT2B, which can be considered a good trade-off between information and risk. The bagging schemes were all applied with 100 trees, a number of classifiers established as suitable for bagging [16].
Both methods were tested on 34 known datasets found in the UCI Machine Learning repository [17]. These datasets are varied with regard to the number of instances, number of discrete and continuous features, ranges of values of discrete attributes, number of states of the class variable, etc. Consistent with the experimental analysis carried out in [5], in all selected datasets the class variable had at least three possible values, as with only two class values all of them are either predicted or a single one. Table 1 shows the most relevant characteristics of each dataset.
In accordance with the experiments carried out in [7], missing values were replaced by modal (mean) values for discrete (continuous) features. Then, continuous attributes were discretized by means of Fayyad and Irani’s discretization procedure [18]. A ten-fold cross-validation procedure was repeated ten times for each preprocessed dataset.
For statistical comparisons, following the recommendations in [19] when there are two algorithms, the Wilcoxon test [20] was used with a level of significance of α = 0.01 . For each dataset, this test ranks the differences between the performance of two methods regardless of the signs. Then, the ranks for the positive differences are compared with those for negative differences.

4.2. Results

Table 2 shows the MIC results obtained by each algorithm on each dataset. A summary of the MIC results is presented in Table 3, which shows the average value, the result of the Wilcoxon test, and the number of datasets where one algorithm beats the other algorithm as exspressed by the MIC measure.
Considering the MIC results in Table 2, Bagging-ICDT obtains lower values than Bagging-ICDT2B. While the difference is not extremely large, the values are almost always in favor of the new procedure. The differences are greater on datasets with a large number of states of the class variable, which makes sense due to the range of the MIC measure. The number of wins for each method compared to the other one can be seen in the third row of Table 3. There is a large difference in the wins of the new procedure compared to the original one (28 vs. 3).
The most important comparison can be seen in the second row of Table 3, where the results of the Wilcoxon test are presented. This test was carried out with a very strong level of significance of 0.01 . Here, we can observe that Bagging-ICDT2B significantly outperforms the original Bagging-ICDT method, demonstrating that the trade-off between risk and information in the original method is not optimal.

5. Conclusions and Future Work

In this work, we have shown that the first method to combine imprecise classifiers based on the minimum level of non-dominance, while a very strong method of combination, has the important drawback of possible information loss. This is because the difference between the set of winning states and the rest can be very small in real situations. Here, we have used the second-best theory to argue that this method can be improved to obtain better results in the combination of imprecise classifiers by relaxing the constraints taken into account.
The experimental analysis carried out in our research has revealed that the new method presented here is clearly superior to the original method. The new method outperforms the existing one on a statistical test with a high level of significance. It has been demonstrated that better results can be produced by relaxing the constraints taken into account in the original method to minimize the loss of information at the cost of assuming greater risk, as the trade-off between risk and information in the original method is not optimal.
The new method presented here can be tuned for specific datasets to obtain better results in future work. Moreover, the procedure presented here for imprecise classification can be inserted in other known ensemble schemes, such as Boosting [21] or Random Forest [8].

Author Contributions

Formal analysis, S.M.-G. and J.A.; Funding acquisition, J.A.; Investigation, S.M.-G., M.D.B. and J.A.; Methodology, S.M.-G. and J.A.; Project administration, J.A.; Supervision, J.A.; Writing—original draft, S.M.-G., M.D.B. and J.A.; Writing—review and editing, S.M.-G. and J.A. All authors have read and agreed to the published version of the manuscript.

Funding

This work has been supported by UGR-FEDER funds under Project A-TIC-344-UGR20 and by the “FEDER/Junta de Andalucía-Consejería de Transformación Económica, Industria, Conocimiento y Universidades” under Project P20_00159.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Abellán, J.; Moral, S. Building classification trees using the total uncertainty criterion. Int. J. Intell. Syst. 2003, 18, 1215–1225. [Google Scholar] [CrossRef] [Green Version]
  2. Abellán, J.; Moral, S. Upper entropy of credal sets. Applications to credal classification. Int. J. Approx. Reason. 2005, 39, 235–255. [Google Scholar] [CrossRef] [Green Version]
  3. Abellán, J.; Mantas, C.J. Improving experimental studies about ensembles of classifiers for bankruptcy prediction and credit scoring. Expert Syst. Appl. 2014, 41, 3825–3830. [Google Scholar] [CrossRef]
  4. Abellán, J. Ensembles of decision trees based on imprecise probabilities and uncertainty measures. Inf. Fusion 2013, 14, 423–430. [Google Scholar] [CrossRef]
  5. Abellán, J.; Masegosa, A.R. Imprecise classification with credal decision trees. Int. J. Uncertain. Fuzziness Knowl.-Based Syst. 2012, 20, 763–787. [Google Scholar] [CrossRef]
  6. Zaffalon, M. The naive credal classifier. J. Stat. Plan. Inference 2002, 105, 5–21. [Google Scholar] [CrossRef]
  7. Moral-García, S.; Mantas, C.J.; Castellano, J.G.; Benítez, M.D.; Abellán, J. Bagging of credal decision trees for imprecise classification. Expert Syst. Appl. 2020, 141, 112944. [Google Scholar] [CrossRef]
  8. Breiman, L. Random Forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef] [Green Version]
  9. Lipsey, R.; Lancaster, K. The General Theory of Second Best. Rev. Econ. Stud. 1956, 24, 11–32. [Google Scholar] [CrossRef]
  10. Walley, P. Inferences from multinomial data; learning about a bag of marbles (with discussion). J. R. Stat. Soc. Ser. B Methodol. 1996, 58, 3–57. [Google Scholar] [CrossRef]
  11. Abellán, J. Uncertainty measures on probability intervals from the imprecise Dirichlet model. Int. J. Gen. Syst. 2006, 35, 509–528. [Google Scholar] [CrossRef] [Green Version]
  12. Shannon, C.E. A Mathematical Theory of Communication. Bell Syst. Tech. J. 1948, 27, 379–423. [Google Scholar] [CrossRef] [Green Version]
  13. Klir, G.J. Uncertainty and Information: Foundations of Generalized Information Theory; John Wiley And Sons, Inc.: Hoboken, NJ, USA, 2005. [Google Scholar] [CrossRef]
  14. Abellán, J. Equivalence relations among dominance concepts on probability intervals and general credal sets. Int. J. Gen. Syst. 2012, 41, 109–122. [Google Scholar] [CrossRef]
  15. Witten, I.H.; Frank, E. Data Mining: Practical Machine Learning Tools and Techniques, 2nd ed.; Morgan Kaufmann Series in Data Management Systems; Morgan Kaufmann Publishers Inc.: San Francisco, CA, USA, 2005. [Google Scholar]
  16. Breiman, L. Bagging Predictors. Mach. Learn. 1996, 24, 123–140. [Google Scholar] [CrossRef] [Green Version]
  17. Lichman, M. UCI Machine Learning Repository. 2013. Available online: https://archive.ics.uci.edu/ (accessed on 20 June 2023).
  18. Fayyad, U.; Irani, K. Multi-valued Interval Discretization of Continuous-valued Attributes for classification Learning. In Proceedings of the 13th International joint Conference on Artificial Inteligence, Chambery, France, 28 August 1993; Morgan Kaufmann: Burlington, MA, USA, 1993; pp. 1022–1027. [Google Scholar]
  19. Demšar, J. Statistical Comparisons of Classifiers over Multiple Data Sets. J. Mach. Learn. Res. 2006, 7, 1–30. [Google Scholar]
  20. Wilcoxon, F. Individual Comparisons by Ranking Methods. Biom. Bull. 1945, 1, 80–83. [Google Scholar] [CrossRef]
  21. Freund, Y.; Schapire, R.E. Experiments with a New Boosting Algorithm. In Proceedings of the Thirteenth International Conference on Machine Learning (ICML 1996), Bari, Italy, 3–6 July 1996; Saitta, L., Ed.; Morgan Kaufmann: Burlington, MA, USA, 1996; pp. 148–156. [Google Scholar]
Figure 1. Summary of the Bagging-ICDT method.
Figure 1. Summary of the Bagging-ICDT method.
Entropy 25 01088 g001
Figure 2. New bagging scheme with ICDT: Bagging-ICDT2B.
Figure 2. New bagging scheme with ICDT: Bagging-ICDT2B.
Entropy 25 01088 g002
Table 1. Dataset description. Column “N” is the number of instances, column “Feat” is the number of features or attribute variables, column “Cont” is the number of continuous variables, column “Disc” is the number of discrete variables, column “K” is the number of states of the class variable, and column “Range” is the range of value of the discrete attributes.
Table 1. Dataset description. Column “N” is the number of instances, column “Feat” is the number of features or attribute variables, column “Cont” is the number of continuous variables, column “Disc” is the number of discrete variables, column “K” is the number of states of the class variable, and column “Range” is the range of value of the discrete attributes.
DatasetNFeatContDiscKRange
anneal8983863262–10
arrhythmia45227920673162
audiology22669069242–6
autos20525151072–22
balance-scale6254403-
bridges-version1107113862–54
bridges-version21071101162–54
car172860643–4
cmc147392732–4
dermatology3663413362–4
ecoli3667707-
flags1943022882–13
hypothyroid37723072342–4
iris1504403-
letter20,0001616026-
lymphography1461831542–8
mfeat-pixel20002400240104–6
nursery12,96080842–4
optdigits56206464010-
page-blocks5473101005-
pendigits10,9921616010-
postop-patient-data9090932–4
primary-tumor33917017212–3
segment2310191607-
soybean68335035192–7
spectrometer5311011001484
splice31906006034–6
sponge764404432–9
tae15153232
vehicle946181804-
vowel99011101112
waveform5000404003-
wine178131303-
zoo1011611672
Table 2. Complete results obtained for the MIC measure, with the best results marked in bold.
Table 2. Complete results obtained for the MIC measure, with the best results marked in bold.
DatasetBagging-ICDTBagging-ICDT2B
anneal1.78471.7847
arrhythmia1.93161.9440
audiology2.59362.6132
autos1.55531.5618
balance-scale0.60060.6049
bridges-version11.04461.0340
bridges-version20.97670.9910
car1.25681.2576
cmc0.26360.2682
dermatology1.68441.6834
ecoli1.61821.6208
flags1.13981.1615
hypothyroid1.37441.3746
iris0.99820.9985
letter2.67712.6967
lymphography0.94170.9558
mfeat-pixel2.00662.0208
nursery1.53981.5405
optdigits1.95791.9715
page-blocks1.54181.5424
pendigits2.09252.0966
postoperative-patient-dat0.62070.6207
primary-tumor1.22781.2504
segment1.83311.8344
soybean2.72032.7256
spectrometer1.95272.0219
splice1.00771.0080
sponge1.01211.0133
tae0.22160.2216
vehicle0.83720.8417
vowel1.85941.8680
waveform0.73250.7354
wine0.98170.9837
zoo1.85781.8571
Average1.42481.4325
Table 3. Summary of the results obtained for the MIC measure. In the “Wilcoxon test” rows, if one classifier is significantly better than the other one this is expressed by “*”. The level of significance used is 0.01 . The row “Number of wins” indicates the number of datasets on which one algorithm beat the other in terms of MIC.
Table 3. Summary of the results obtained for the MIC measure. In the “Wilcoxon test” rows, if one classifier is significantly better than the other one this is expressed by “*”. The level of significance used is 0.01 . The row “Number of wins” indicates the number of datasets on which one algorithm beat the other in terms of MIC.
Bagging-ICDTBagging-ICDT2B
MIC:Average1.36521.4248
Wilcoxon t-test *
Number of wins328
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Moral-García, S.; Benítez, M.D.; Abellán, J. Upgrading the Fusion of Imprecise Classifiers. Entropy 2023, 25, 1088. https://doi.org/10.3390/e25071088

AMA Style

Moral-García S, Benítez MD, Abellán J. Upgrading the Fusion of Imprecise Classifiers. Entropy. 2023; 25(7):1088. https://doi.org/10.3390/e25071088

Chicago/Turabian Style

Moral-García, Serafín, María D. Benítez, and Joaquín Abellán. 2023. "Upgrading the Fusion of Imprecise Classifiers" Entropy 25, no. 7: 1088. https://doi.org/10.3390/e25071088

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop