Enhancing Hierarchical Classification in Tree-Based Models Using Level-Wise Entropy Adjustment
Abstract
:1. Introduction
- It introduces Penalized Information Gain (PIG), a novel metric that integrates Information Gain (IG) with Taxonomic Informativeness (TI), enabling a hierarchy-aware evaluation of splits in decision tree models. This approach ensures that splits respect both statistical and taxonomic criteria.
- It provides a detailed empirical validation of the proposed metric using two real-world datasets structured according to the GS1 Global Product Classification (GPC) system. These experiments demonstrate the practical benefits of PIG, including improved classification accuracy and semantic alignment with hierarchical taxonomies.
- It critically analyzes hierarchical classification metrics, comparing IG, TI, and PIG to highlight their respective strengths and limitations. This analysis provides valuable insights for developing more effective split evaluation criteria in hierarchical classification tasks.
2. Relevant Research
3. Materials and Methods
3.1. Decision Tree Splitting Criteria
3.1.1. Information Gain (IG) [40]
- is the entropy of S;
- and are the proportion of samples , respectively.
3.1.2. Average Taxonomic Informativeness (ATI) [43]
- N—total number of samples.
- L—total number of levels in the taxonomy.
- —weight assigned to level l in the hierarchy, representing the informativeness or specificity of that level.
- —indicator function that evaluates whether the prediction at level l matches the true label for sample i
3.1.3. Penalized Information Gain (PIG)
- α > 0 is a scaling coefficient that controls the sensitivity of PF to changes in ATI.
- The logarithm provides a non-linear relationship, ensuring a gradual increase in PF as ATI improves.
- Hierarchy Awareness: Aligns splits with label hierarchies, preserving meaningful taxonomic distinctions.
- Improved Interpretability: Produces decision boundaries that respect semantic relationships, enhancing model transparency.
- Enhanced Generalization: Leverages hierarchical structures for better multi-level classification performance.
3.2. Addressing Challenges in Imbalanced Hierarchical Datasets
- Penalization of size-dominant splits that group taxonomically distinct classes.
- Promotion of balanced splits, ensuring that underrepresented branches are adequately considered.
3.3. Encouraging Hierarchical Balance
3.4. Quantitative Foundation for Balance Preservation
- Bias in Standard Information Gain: Conventional Information Gain (IG(S)) inherently favors larger branches due to the proportional weighting of subsets in entropy calculations. This bias skews splits toward majority branches, often overlooking smaller but semantically significant ones.
- Adjustment Through PIG: PIG addresses this bias by scaling IG(S) with the penalty factor Tableware introducing a hierarchical adjustment. This adjustment ensures that splits consider the taxonomic relationships between classes, reducing the undue influence of size imbalances and better representing smaller branches.
- Preservation of Rare Branches: PIG ensures that smaller, underrepresented branches remain distinct, enabling the model to generalize effectively to these classes.
- Alignment with Taxonomy: By aligning splits with the hierarchical structure, PIG reduces semantic distortion and ensures that meaningful distinctions are preserved.
- Improved Model Fairness: PIG prevents overfitting to the majority of branches, fostering models that are more robust and equitable across imbalanced datasets.
4. Experimental Settings
4.1. Dataset
4.2. Experimental Validation of PIG Performance
5. Experimental Setup
- Byte Pair Encoding (BPE): Applied as a subword tokenization method to address the challenges posed by abbreviated and short texts. BPE ensures compact and meaningful text representation while preserving semantic integrity [49].
- BERT Embeddings: Contextualized dense vector representations were extracted using BERT to capture semantic relationships within the input text [50]. These embeddings served as high-dimensional feature inputs to the machine learning models.
6. Results
- To compare the performance of random forest and XGBoost using different split criteria (IG, ATI, and PIG) in preserving both statistical entropy reduction and hierarchical taxonomic alignment.
- To analyze the impact of hierarchy-aware metrics (ATI and PIG) on mitigating the challenges of imbalanced data, particularly within deeply nested and sparsely populated branches of the GPC taxonomy.
7. Discussion
8. Conclusions
- Comparative analysis to evaluate PIG against other state-of-the-art approaches for hierarchical classification, including neural network-based and hybrid methods.
- Scalability by applying the proposed methodology to larger and more diverse datasets across domains such as bioinformatics, product taxonomy, and medical hierarchies.
- Model extensions by exploring the integration of PIG into ensemble methods like gradient boosting, as well as hierarchical multi-label and deep learning frameworks.
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Junardi, W.; Khodra, M.L. Automatic Multi-Label Classification for GDP Economic-Phenomenon News. In Proceedings of the 2020 International Conference on ICT for Smart Society (ICISS), Bandung, Indonesia, 19–20 November 2020; IEEE: Bandung, Indonesia, 2020; pp. 1–6. [Google Scholar] [CrossRef]
- Cotacallapa, H.; Saboya, N.; Canas Rodrigues, P.; Salas, R.; Linkolk López-Gonzales, J. A Flat-Hierarchical Approach Based on Machine Learning Model for e-Commerce Product Classification. IEEE Access 2024, 12, 72730–72745. [Google Scholar] [CrossRef]
- Uddin, M.A.; Aryal, S.; Bouadjenek, M.R.; Al-Hawawreh, M.; Talukder, M.A. Hierarchical Classification for Intrusion Detection System: Effective Design and Empirical Analysis. arXiv 2024. [Google Scholar] [CrossRef]
- Cao, Y.-K.; Wei, Z.-Y.; Tang, Y.-J.; Jin, C.-K. Hierarchical Label Text Classification Method with Deep-Level Label-Assisted Classification. In Proceedings of the 2023 IEEE 12th Data Driven Control and Learning Systems Conference (DDCLS), Xiangtan, China, 12–14 May 2023; IEEE: Xiangtan, China, 2023; pp. 1467–1474. [Google Scholar] [CrossRef]
- Chang, C.-M.; Mishra, S.D.; Igarashi, T. A Hierarchical Task Assignment for Manual Image Labeling. In Proceedings of the 2019 IEEE Symposium on Visual Languages and Human-Centric Computing (VL/HCC), Memphis, TN, USA, 14–18 October 2019; IEEE: Memphis, TN, USA, 2019; pp. 139–143. [Google Scholar] [CrossRef]
- Fan, Q.; Qiu, C. Hierarchical Multi-Label Text Classification Method Based on Multi-Level Decoupling. In Proceedings of the 2023 3rd International Conference on Neural Networks, Information and Communication Engineering (NNICE), Guangzhou, China, 24–26 February 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 453–457. [Google Scholar]
- Rejeb, A.; Keogh, J.G.; Martindale, W.; Dooley, D.; Smart, E.; Simske, S.; Wamba, S.F.; Breslin, J.G.; Bandara, K.Y.; Thakur, S.; et al. Charting Past, Present, and Future Research in the Semantic Web and Interoperability. Future Internet 2022, 14, 161. [Google Scholar] [CrossRef]
- Global Product Classification (GPC). Available online: https://www.gs1.org/standards/gpc (accessed on 17 December 2024).
- Sepúlveda-Rojas, J.P.; Aravena, S.; Carrasco, R. Increasing Efficiency in Furniture Remanufacturing with AHP and the SECI Model. Sustainability 2024, 16, 10339. [Google Scholar] [CrossRef]
- Kong, X.; Zhu, X.; Wang, M.; Wang, X.; Zou, M. Text Classification for Social Governance: A Novel Strategy with Adaptive Reward Mechanisms. In Proceedings of the 2024 6th International Conference on Communications, Information System and Computer Engineering (CISCE), Guangzhou, China, 10–12 May 2024; IEEE: Piscataway, NJ, USA, 2024; pp. 339–343. [Google Scholar] [CrossRef]
- Zhang, Q.; Chai, B.; Song, B.; Zhao, J. A Hierarchical Fine-Tuning Based Approach for Multi-Label Text Classification. In Proceedings of the 2020 IEEE 5th International Conference on Cloud Computing and Big Data Analytics (ICCCBDA), Chengdu, China, 10–13 April 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 51–54. [Google Scholar] [CrossRef]
- Li, J.; Wang, H.; Song, C.; Han, R.; Hu, T. Research on Hierarchical Clustering Undersampling and Random Forest Fusion Classification Method. In Proceedings of the 2021 IEEE International Conference on Progress in Informatics and Computing (PIC), Shanghai, China, 17–19 December 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 53–57. [Google Scholar] [CrossRef]
- Tangirala, S. Evaluating the Impact of GINI Index and Information Gain on Classification Using Decision Tree Classifier Algorithm*. Int. J. Adv. Comput. Sci. Appl. 2020, 11, 277. [Google Scholar] [CrossRef]
- Peng, H.; Li, J.; Wang, S.; Wang, L.; Gong, Q.; Yang, R.; Li, B.; Yu, P.S.; He, L. Hierarchical Taxonomy-Aware and Attentional Graph Capsule RCNNs for Large-Scale Multi-Label Text Classification. IEEE Trans. Knowl. Data Eng. 2021, 33, 2505–2519. [Google Scholar] [CrossRef]
- Rao, S.X.; Egger, P.H.; Zhang, C. Hierarchical Classification of Research Fields in the “Web of Science” Using Deep Learning 2024. arXiv 2023. [Google Scholar] [CrossRef]
- Silla, C.N.; Freitas, A.A. A Survey of Hierarchical Classification across Different Application Domains. Data Min. Knowl. Discov. 2011, 22, 31–72. [Google Scholar] [CrossRef]
- Nowozin, S. Improved Information Gain Estimates for Decision Tree Induction. In Proceedings of the 29th International Conference on Machine Learning (ICML’12), Edinburgh, Scotland, 26 June–1 July 2012; Omnipress: Madison, WI, USA, 2012; pp. 571–578. [Google Scholar] [CrossRef]
- Naik, A.; Rangwala, H. Filter Based Taxonomy Modification for Improving Hierarchical Classification. arXiv 2016. [Google Scholar] [CrossRef]
- Cai, X.; Xiao, M.; Ning, Z.; Zhou, Y. Resolving the Imbalance Issue in Hierarchical Disciplinary Topic Inference via LLM-Based Data Augmentation. In Proceedings of the 2023 IEEE International Conference on Data Mining (ICDM), Shanghai, China, 1–4 December 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 956–961. [Google Scholar]
- Heinsen, F.A. Tree Methods for Hierarchical Classification in Parallel. arXiv 2022. [Google Scholar] [CrossRef]
- Asadi, A.R. An Entropy-Based Model for Hierarchical Learning 2023. arXiv 2022. [Google Scholar] [CrossRef]
- Wadud, M.d.A.H.; Rashadul, M.d. Text Coherence Analysis Based on Misspelling Oblivious Word Embeddings and Deep Neural Network. Int. J. Adv. Comput. Sci. Appl. 2021, 12, 124. [Google Scholar] [CrossRef]
- Doroshenko, A. Application of Global Optimization Methods to Increase the Accuracy of Classification in the Data Mining Tasks. Comput. Model. Intell. Syst. 2019, 2353, 98–109. [Google Scholar] [CrossRef]
- Dindar, S.; Kaewunruen, S.; An, M. A Hierarchical Bayesian-Based Model for Hazard Analysis of Climate Effect on Failures of Railway Turnout Components. Reliab. Eng. Syst. Saf. 2022, 218, 108130. [Google Scholar] [CrossRef]
- Busson, A.J.G.; Rocha, R.; Gaio, R.; Miceli, R.; Pereira, I.; Moraes, D.d.S.; Colcher, S.; Veiga, A.; Rizzi, B.; Evangelista, F.; et al. Hierarchical Classification of Financial Transactions Through Context-Fusion of Transformer-Based Embeddings and Taxonomy-Aware Attention Layer. In Proceedings of the Anais do II Brazilian Workshop on Artificial Intelligence in Finance (BWAIF 2023), João Pessoa, Paraíba, 6–11 August 2023; pp. 13–24. [Google Scholar] [CrossRef]
- Dini, L.; Brunato, D.; Dell’Orletta, F.; Caselli, T. TEXT-CAKE: Challenging Language Models on Local Text Coherence. In Proceedings of the 31st International Conference on Computational Linguistics, Abu Dhabi, United Arab Emirates, 19–24 January 2025; Rambow, O., Wanner, L., Apidianaki, M., Al-Khalifa, H., Eugenio, B.D., Schockaert, S., Eds.; Association for Computational Linguistics: Stroudsburg, PA, USA, 2025; pp. 4384–4398. [Google Scholar]
- Shu, H.; Cao, L.; Xu, Z.; Liu, K. The Research of Multidimensional Information Decision Mining Based on Information Entropy. In Proceedings of the 2009 International Forum on Information Technology and Applications, Chengdu, China, 15–17 May 2009; IEEE: Piscataway, NJ, USA, 2009; pp. 732–735. [Google Scholar]
- Song, J.; Zhang, P.; Qin, S.; Gong, J. A Method of the Feature Selection in Hierarchical Text Classification Based on the Category Discrimination and Position Information. In Proceedings of the 2015 International Conference on Industrial Informatics-Computing Technology, Intelligent Technology, Industrial Information Integration, Wuhan, China, 3–4 December 2015; IEEE: Piscataway, NJ, USA, 2015; pp. 132–135. [Google Scholar]
- Mutsaddi, A.; Jamkhande, A.; Thakre, A.; Haribhakta, Y. BERTopic for Topic Modeling of Hindi Short Texts: A Comparative Study. arXiv 2025. [Google Scholar] [CrossRef]
- Williams, L.; Anthi, E.; Burnap, P. Comparing Hierarchical Approaches to Enhance Supervised Emotive Text Classification. Big Data Cogn. Comput. 2024, 8, 38. [Google Scholar] [CrossRef]
- Zangari, A.; Marcuzzo, M.; Rizzo, M.; Giudice, L.; Albarelli, A.; Gasparetto, A. Hierarchical Text Classification and Its Foundations: A Review of Current Research. Electronics 2024, 13, 1199. [Google Scholar] [CrossRef]
- Joachims, T. Text Categorization with Support Vector Machines: Learning with Many Relevant Features. In Machine Learning: ECML-98; Nédellec, C., Rouveirol, C., Eds.; Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 1998; Volume 1398, pp. 137–142. ISBN 978-3-540-64417-0. [Google Scholar]
- Yang, Z.; Yang, D.; Dyer, C.; He, X.; Smola, A.; Hovy, E. Hierarchical Attention Networks for Document Classification. In Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, San Diego, CA, USA, 12–17 June 2016; Association for Computational Linguistics: Stroudsburg, PA, USA, 2016; pp. 1480–1489. [Google Scholar] [CrossRef]
- Basu, S.; Bilenko, M.; Mooney, R.J. A Probabilistic Framework for Semi-Supervised Clustering. In Proceedings of the Tenth ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Seattle, WA, USA, 22–25 August 2004; ACM: New York, NY, USA, 2004; pp. 59–68. [Google Scholar] [CrossRef]
- Sebastiani, F. Classification of Text, Automatic. In Encyclopedia of Language & Linguistics; Elsevier: Amsterdam, The Netherlands, 2006; pp. 457–462. ISBN 978-0-08-044854-1. [Google Scholar]
- Read, J.; Pfahringer, B.; Holmes, G.; Frank, E. Classifier Chains for Multi-Label Classification. Mach. Learn. 2011, 85, 333–359. [Google Scholar] [CrossRef]
- Sun, A.; Lim, E.-P. Hierarchical Text Classification and Evaluation. In Proceedings of the 2001 IEEE International Conference on Data Mining, San Jose, CA, USA, 29 November–2 December 2001; IEEE: Piscataway, NJ, USA, 2001; pp. 521–528. [Google Scholar]
- Narushynska, O.; Teslyuk, V.; Doroshenko, A.; Arzubov, M. Data Sorting Influence on Short Text Manual Labeling Quality for Hierarchical Classification. Big Data Cogn. Comput. 2024, 8, 41. [Google Scholar] [CrossRef]
- Kosmopoulos, A.; Paliouras, G.; Androutsopoulos, I. Probabilistic Cascading for Large Scale Hierarchical Classification. arXiv 2015. [Google Scholar] [CrossRef]
- Kononenko, I.; Kukar, M. Measures for Evaluating the Quality of Attributes. In Machine Learning and Data Mining; Elsevier: Amsterdam, The Netherlands, 2007; pp. 153–180. ISBN 978-1-904275-21-3. [Google Scholar]
- Kosmopoulos, A.; Partalas, I.; Gaussier, E.; Paliouras, G.; Androutsopoulos, I. Evaluation Measures for Hierarchical Classification: A Unified View and Novel Approaches. Data Min. Knowl. Discov. 2015, 29, 820–865. [Google Scholar] [CrossRef]
- Plaud, R.; Labeau, M.; Saillenfest, A.; Bonald, T. Revisiting Hierarchical Text Classification: Inference and Metrics 2024; Association for Computational Linguistics: Stroudsburg, PA, USA, 2024. [Google Scholar] [CrossRef]
- Warwick, R.M. Average Taxonomic Diversity and Distinctness. In Encyclopedia of Ecology; Elsevier: Amsterdam, The Netherlands, 2008; pp. 300–305. ISBN 978-0-08-045405-4. [Google Scholar]
- Hwang, S.; Yeo, H.G.; Hong, J.-S. A New Splitting Criterion for Better Interpretable Trees. IEEE Access 2020, 8, 62762–62774. [Google Scholar] [CrossRef]
- Naik, A.; Rangwala, H. Inconsistent Node Flattening for Improving Top-down Hierarchical Classification. arXiv 2017. [Google Scholar] [CrossRef]
- Yang, J.; Gao, M.; Kuang, K.; Ni, B.; She, Y.; Xie, D.; Chen, C. Hierarchical Classification of Pulmonary Lesions: A Large-Scale Radio-Pathomics Study. In Medical Image Computing and Computer Assisted Intervention; Springer: Berlin/Heidelberg, Germany, 2020; Volume 12266, pp. 497–507. [Google Scholar] [CrossRef]
- Doroshenko, A.; Tkachenko, R. Classification of Imbalanced Classes Using the Committee of Neural Networks. In Proceedings of the 2018 IEEE 13th International Scientific and Technical Conference on Computer Sciences and Information Technologies (CSIT), Lviv, Ukraine, 11–14 September 2018; IEEE: Lviv, Ukraine, 2018; pp. 400–403. [Google Scholar] [CrossRef]
- Product Categories. Available online: https://www.directionsforme.org/categories (accessed on 17 December 2024).
- Vilar, D.; Federico, M. A Statistical Extension of Byte-Pair Encoding. In Proceedings of the 18th International Conference on Spoken Language Translation (IWSLT 2021), Bangkok, Thailand, 5–6 August 2021; Association for Computational Linguistics: Stroudsburg, PA, USA, 2021; pp. 263–275. [Google Scholar] [CrossRef]
- Devlin, J.; Chang, M.-W.; Lee, K.; Toutanova, K. BERT: Pre-Training of Deep Bidirectional Transformers for Language Understanding; Association for Computational Linguistics: Stroudsburg, PA, USA, 2019; Volume 1, pp. 4171–4186. [Google Scholar] [CrossRef]
- Teslyuk, V.; Doroshenko, A.; Narushynska, O. Preprocessing Product Descriptions with Byte Pair Encoding: A Solution for Abbreviation-Heavy Texts//CEUR Workshop Proceedings. In Proceedings of the Computational Intelligence Application Workshop (CIAW 2024), Lviv, Ukraine, 10–12 October 2024; CEUR: Lviv, Ukraine, 2024; Volume 3861, pp. 57–76. [Google Scholar]
- Romeo, L.; Frontoni, E. A Unified Hierarchical XGBoost Model for Classifying Priorities for COVID-19 Vaccination Campaign. Pattern Recognit. 2022, 121, 108197. [Google Scholar] [CrossRef] [PubMed]
Product Name | Real Level 1 | Real Level 2 | Real Level 3 | Real Level 4 | Split |
---|---|---|---|---|---|
Example 1 | |||||
“Pics 10” “Flour Burrito Tortillas” | Food/Beverage | Bread/Bakery Products | Bread | Bread (Perishable) | Left |
“Schmidt’s Delicatessen” “No Seeds” “Rye Bread” | Food/Beverage | Bread/Bakery Products | Bread | Bread (Perishable) | Left |
Ahold 100% Whole Wheat Bread | Food/Beverage | Bread/Bakery Products | Bread | Bread (Perishable) | Right |
Tia Rosa Flour Tortillas Burrito | Food/Beverage | Bread/Bakery Products | Bread | Bread (Shelf Stable) | Right |
Turano Gourmet Sandwich Rolls—8 CT | Food/Beverage | Bread/Bakery Products | Bread | Bread (Shelf Stable) | Right |
Udi’s Gluten Free Classic Hamburger Buns—4 CT | Food/Beverage | Bread/Bakery Products | Bread | Bread (Shelf Stable) | Left |
Example 2 | |||||
Tia Rosa Flour Tortillas Burrito | Food/Beverage | Bread/Bakery Products | Bread | Bread (Shelf Stable) | Left |
Turano Gourmet Sandwich Rolls—8 CT | Food/Beverage | Bread/Bakery Products | Bread | Bread (Shelf Stable) | Left |
Udi’s Gluten Free Classic Hamburger Buns—4 CT | Food/Beverage | Bread/Bakery Products | Bread | Bread (Shelf Stable) | Right |
Ahold Fiber Select Brownies Chocolate Fudge—6 CT | Food/Beverage | Cereal/Grain/ Pulse Products | Processed Cereal Products | Cereals Products— Not Ready to Eat (Shelf Stable) | Right |
American Classic Gourmet Muffins Raisin Bran—4 CT | Food/Beverage | Cereal/Grain/ Pulse Products | Processed Cereal Products | Cereals Products— Not Ready to Eat (Shelf Stable) | Right |
Balconi Choco & Latte Sponge Cakes—10 CT | Food/Beverage | Cereal/Grain/ Pulse Products | Processed Cereal Products | Cereals Products— Not Ready to Eat (Shelf Stable) | Left |
Example 3 | |||||
Tia Rosa Flour Tortillas Burrito | Food/Beverage | Bread/Bakery Products | Bread | Bread (Shelf Stable) | Left |
Turano Gourmet Sandwich Rolls—8 CT | Food/Beverage | Bread/Bakery Products | Bread | Bread (Shelf Stable) | Left |
Udi’s Gluten Free Classic Hamburger Buns—4 CT | Food/Beverage | Bread/Bakery Products | Bread | Bread (Shelf Stable) | Right |
Wheatena Toasted Wheat Cereal | Food/Beverage | Cereal/Grain/ Pulse Products | Processed Cereal Products | Cereals Products— Not Ready to Eat (Shelf Stable) | Right |
Bobs Red Mill Old Fashioned Whole Grain Rolled Oats 32 oz STAND PACK | Food/Beverage | Cereal/Grain/ Pulse Products | Processed Cereal Products | Cereals Products— Not Ready to Eat (Shelf Stable) | Right |
Hodgson Mill Oat Bran Hot Cereal | Food/Beverage | Cereal/Grain/ Pulse Products | Processed Cereal Products | Cereals Products— Not Ready to Eat (Shelf Stable) | Left |
Example | IG | ATI | PIG |
---|---|---|---|
1 | 0.082 | 0.92 | 0.023 |
2 | 0.082 | 0.83 | 0.021 |
3 | 0.082 | 0.75 | 0.020 |
Name | Brick_Test_Accuracy | Brick_Test_F1 Score | Brick_Test_Precision | Brick_Test_Recall | Class_Test_Accuracy | Class_Test_F1 Score | Class_Test_Precision | Class_Test_Recall | Family_Test_Accuracy | Family_Test_F1 Score | Family_Test_Precision | Family_Test_Recall | Segment_Test_Accuracy | Segment_Test_F1 Score | Segment_Test_Precision | Segment_Test_Recall |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
1_RandomForest_IG (standart) | 0.54 | 0.52 | 0.56 | 0.54 | 0.62 | 0.60 | 0.62 | 0.62 | 0.69 | 0.68 | 0.68 | 0.69 | 0.96 | 0.96 | 0.96 | 0.96 |
1_XGBoost_IG (standart) | 0.56 | 0.56 | 0.57 | 0.56 | 0.63 | 0.63 | 0.63 | 0.63 | 0.70 | 0.69 | 0.70 | 0.70 | 0.96 | 0.96 | 0.96 | 0.96 |
2_RandomForest_ATI | 0.55 | 0.52 | 0.57 | 0.55 | 0.62 | 0.61 | 0.63 | 0.62 | 0.69 | 0.68 | 0.69 | 0.69 | 0.96 | 0.96 | 0.96 | 0.96 |
2_XGBoost_ATI | 0.65 | 0.63 | 0.68 | 0.65 | 0.74 | 0.72 | 0.74 | 0.74 | 0.80 | 0.80 | 0.80 | 0.80 | 0.98 | 0.98 | 0.98 | 0.98 |
3_RandomForest_PIG | 0.66 | 0.63 | 0.68 | 0.66 | 0.74 | 0.72 | 0.74 | 0.74 | 0.80 | 0.80 | 0.80 | 0.80 | 0.98 | 0.98 | 0.98 | 0.98 |
3_XGBoost_PIG | 0.69 | 0.69 | 0.70 | 0.69 | 0.77 | 0.77 | 0.77 | 0.77 | 0.84 | 0.84 | 0.84 | 0.84 | 0.99 | 0.99 | 0.99 | 0.99 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Narushynska, O.; Doroshenko, A.; Teslyuk, V.; Antoniv, V.; Arzubov, M. Enhancing Hierarchical Classification in Tree-Based Models Using Level-Wise Entropy Adjustment. Big Data Cogn. Comput. 2025, 9, 65. https://doi.org/10.3390/bdcc9030065
Narushynska O, Doroshenko A, Teslyuk V, Antoniv V, Arzubov M. Enhancing Hierarchical Classification in Tree-Based Models Using Level-Wise Entropy Adjustment. Big Data and Cognitive Computing. 2025; 9(3):65. https://doi.org/10.3390/bdcc9030065
Chicago/Turabian StyleNarushynska, Olga, Anastasiya Doroshenko, Vasyl Teslyuk, Volodymyr Antoniv, and Maksym Arzubov. 2025. "Enhancing Hierarchical Classification in Tree-Based Models Using Level-Wise Entropy Adjustment" Big Data and Cognitive Computing 9, no. 3: 65. https://doi.org/10.3390/bdcc9030065
APA StyleNarushynska, O., Doroshenko, A., Teslyuk, V., Antoniv, V., & Arzubov, M. (2025). Enhancing Hierarchical Classification in Tree-Based Models Using Level-Wise Entropy Adjustment. Big Data and Cognitive Computing, 9(3), 65. https://doi.org/10.3390/bdcc9030065