Next Article in Journal
Low-Power Radiation-Hardened Static Random Access Memory with Enhanced Read Stability for Space Applications
Previous Article in Journal
GAN-Based Map Generation Technique of Aerial Image Using Residual Blocks and Canny Edge Detector
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Enhanced Feature Selection via Hierarchical Concept Modeling

by
Jarunee Saelee
1,
Patsita Wetchapram
2,
Apirat Wanichsombat
2,
Arthit Intarasit
1,
Jirapond Muangprathub
2,*,
Laor Boongasame
3 and
Boonyarit Choopradit
4
1
Faculty of Science and Technology, Prince of Songkla University, Pattani Campus, Pattani 94000, Thailand
2
Faculty of Science and Industrial Technology, Prince of Songkla University, Surat Thani Campus, Surat Thani 84000, Thailand
3
Department of Mathematics, Faculty of Science, King Mongkut’s Institute of Technology Ladkrabang, Bangkok 10520, Thailand
4
Department of Mathematics and Statistics, Faculty of Science and Technology, Thammasat University (Rangsit Campus), Pathumthani 12120, Thailand
*
Author to whom correspondence should be addressed.
Appl. Sci. 2024, 14(23), 10965; https://doi.org/10.3390/app142310965
Submission received: 24 October 2024 / Revised: 21 November 2024 / Accepted: 22 November 2024 / Published: 26 November 2024

Abstract

:
The objectives of feature selection include simplifying modeling and making the results more understandable, improving data mining efficiency, and providing clean and understandable data preparation. With big data, it also allows us to reduce computational time, improve prediction performance, and better understand the data in machine learning or pattern recognition applications. In this study, we present a new feature selection approach based on hierarchical concept models using formal concept analysis (FCA) and a decision tree (DT) for selecting a subset of attributes. The presented methods are evaluated based on all learned attributes with 10 datasets from the UCI Machine Learning Repository by using three classification algorithms, namely decision trees, support vector machines (SVM), and artificial neural networks (ANN). The hierarchical concept model is built from a dataset, and it is selected by top-down considering features (attributes) node for each level of structure. Moreover, this study is considered to provide a mathematical feature selection approach with optimization based on a paired-samples t-test. To compare the identified models in order to evaluate feature selection effects, the indicators used were information gain (IG) and chi-squared (CS), while both forward selection (FS) and backward elimination (BS) were tested with the datasets to assess whether the presented model was effective in reducing the number of features used. The results show clearly that the proposed models when using DT or using FCA, needed fewer features than the other methods for similar classification performance.

1. Introduction

At present, big data are often confronted, and dataset sizes have increased dramatically in machine learning classification applications. Feature selection is typically applied in these classification tasks. The aim of feature selection is to improve the predictive performance of classifiers by removing redundant or irrelevant features. A single feature subset to generate a binary classifier tends to have features that are useful for distinguishing these specific classes but useless for distinguishing others [1,2,3]. Feature selection is a common task in applications of pattern recognition, data mining, and machine learning since it can help improve prediction quality, reduce computation time, and allow building more human-understandable models. Thus, it is used in many applications before training a classifier [3]. However, while there are a lot of state-of-the-art approaches for feature selection in standard feature space [4], only a few approaches for feature selection in hierarchical feature space have been proposed in the literature [1,5].
Hierarchical modeling is one approach to feature selection for enhanced predictive performance of classifiers. It is not only advantageous for identifying the hierarchical model itself but is also helpful for selecting a feature subset for each node [1,5,6,7,8]. Classes in a hierarchical structure have both parent–children relationships and sibling relationships [2,9,10]. Moreover, classes with a parent–children relationship are similar to each other and may share common features for classification, while distinguishing between classes with a sibling relationship may require different features [1,8,11]. The authors developed a method for joint feature selection and hierarchical classifier design using genetic algorithms. Thus, this work focuses on using hierarchical structure in feature selection to improve the predictive performance in classification. In this study, we designed a feature selection algorithm based on the hierarchical information structure.
The hierarchical models were built by using two strategies: bottom-up and top-down. The bottom-up strategy starts by placing each object in its own cluster and then merges these atomic clusters into larger and larger clusters until all the objects are in a single cluster or until some other termination conditions are satisfied. Later, the top-down strategy performs the reverse by starting with all objects in one cluster. It subdivides the cluster into smaller and smaller pieces until each object forms a cluster on its own or until some other termination conditions are met, such as a desired number of clusters or that the distance between the two closest clusters is above a certain threshold [12,13,14]. Thus, this work applied both bottom-up and top-down approaches, which were performed with formal concept analysis (FCA) and a decision tree, respectively. Both FCA and the decision tree approach were considered for feature selection in order to reduce the attribute count while still retaining the efficiency of data classification.
In the context of feature selection, the hierarchical approach provides distinct advantages over non-hierarchical or flat methods, particularly when applied to complex datasets where relationships among features are not uniform. Traditional feature selection techniques, while effective for certain tasks, often operate in a flat structure and may overlook nuanced, multi-level relationships between features. Hierarchical models, such as those using FCA and DT, address these limitations by organizing features in a structured, multi-level manner that reflects inherent data patterns more accurately. Thus, FCA is advantageous for building hierarchical feature selection models because it organizes features based on generalization and specialization relationships. This lattice-based hierarchy helps identify clusters of related features and isolate distinctive attributes within classes. FCA’s structured approach reduces noise and redundancy more effectively than flat methods, as it respects the hierarchical dependencies among features. Meanwhile, the DT model is well suited for hierarchical feature selection due to its top-down approach, which prioritizes attributes based on their classification power at each level. By constructing a hierarchy where critical features emerge at the root levels, DT minimizes the feature space efficiently. This targeted selection enables better classification performance with fewer attributes, particularly in cases where some features are only relevant at certain levels within the dataset hierarchy. For these reasons, hierarchical models can better manage complex data with interrelated features. FCA’s logical structure and DT’s decision nodes allow these models to adapt to varying feature relevance across levels, providing more accurate classification outcomes than flat selection methods. This adaptability is particularly beneficial in tasks where features exhibit varying importance depending on context or subclass.
FCA [15,16,17,18], invented by Rudolf Wille in 1982, is a method for data analysis based on concept lattice. It is widely used in information science to describe attributes and objects of information that can be represented in a hierarchical structure. FCA provides the relationships for generalization and specialization among concepts in the concept lattice. This method is rarely applied to feature selection in classification or other tasks. This study applied the advantage of general knowledge through calculation and selection of a root node in the hierarchical structure. The alternative approach of decision trees uses these powerful and popular tools for classification and prediction. Such classifier has a tree structure, where each node is either a leaf node or a branching decision node [11,19]. This approach is widely used to perform feature selection. The advantages of decision trees are intuitive appeal for knowledge expression, simple implementation, and high classification accuracy. Thus, this work applied both FCA and decision trees to reduce the number of features that need to be collected and also to provide better classification accuracy. Due to the different patterns of selecting general knowledge in both FCA and decision trees in hierarchical concept structure, FCA will be constructed a concept lattice structure based on logic and set theory while decision trees will be built a tree structure based on information gain. Moreover, the decision tree in this study is considered to provide a mathematical selection of the features with the optimization level based on paired-samples t-test. Next, the proposed models were tested and evaluated by using three popular algorithms for classification—namely, decision tree, support vector machine (SVM), and artificial neural network (ANN)—to assess the predictive performances with the presented approach using data from the UCI repository.
This article is organized as follows. Section 2 briefly describes the machine learning techniques applied in this work. Section 3 presents the research methodology, followed by the experimental results and discussion in Section 4. Finally, Section 5 concludes the article.

2. Related Works

Feature selection methods have been used for a wide variety of applications [1,2]. Such methods proceed to eliminate unfavorable features that might be noisy, redundant, or irrelevant and could penalize the performance of a classifier [2]. Feature selection thus contributes to a reduction in the dimensionality of data and a restriction of the inputs, which can contain missing values in one or several features [1,4]. The various feature selection methods fall into three main categories: filters, wrappers, and embedded methods [2,3,4,20]. The first category of filter methods selects the features by using weights that indicate correlations between each feature and a class. The largest weights are selected in rank order. The calculated weights may be based on information gain [1,3,20], chi-square [21,22], or other saliency measures. The second category is wrapper methods that evaluate feature subsets by directly using a learning algorithm and selecting features based on their impact on the model’s performance [3,20]. While this appears reasonable as a direct approach, it tends to be too computationally expensive. Finally, embedded methods incorporate feature selection within the learning algorithm during the training process [1,2,3]. Currently, several studies have explored innovative methods to improve learning frameworks, particularly in meta-learning, robustness, and advanced classification strategies. For instance, Liu et al. investigated the adaptability of model-agnostic meta-learning (MAML) in NLP applications, providing insights into optimizing task generalization [23]. Similarly, the robustness of learning frameworks was examined in adversarial scenarios by Yang et al., emphasizing the importance of reliability in sensitive domains like healthcare [24]. Additionally, Ju et al. introduced hypergraph-enhanced semi-supervised classification, highlighting the utility of complex relational modeling [25]. These works provide valuable insights that contextualize and contrast our hierarchical feature selection approach. In prior studies, many feature selection methods were developed to identify the most relevant and informative features from a dataset. Selecting the right set of features can improve model performance, reduce overfitting, enhance interpretability, and speed up the learning process. This study focuses on applying a hierarchical concept model to select the features from available data.
In recent years, feature selection based on a hierarchical information structure has been proposed because of its rational learning inside the structure. Zhao et al. [26] proposed a hierarchical feature selection framework by considering the parent–child relationship, sibling relationship, and family relationship. These relationships were modeled and implemented using a data matrix concept [5]. Tuo et al. [13] presented the hierarchical feature selection with subtree-based graph regularization by exploring two-way dependence among different classes. Zhao et al. [5] designed a feature selection strategy for hierarchical classification based on fuzzy rough sets to compute the lower and upper approximations of classes organized in a class hierarchy. The authors developed an efficient hierarchical feature selection algorithm based on sibling nodes [5]. Huang et al. [27] proposed a feature selection framework based on semantic and structural information of labels for hierarchical feature selection by transforming to semantic regularization and adapted the proposed model to a directed acyclic graph case. Liu et al. [28] proposed a robust hierarchical feature to reduce the adverse effects of data outliers and learn relatively robust and discriminative feature subsets for hierarchical classification.
Hierarchical feature selection has gained attention due to its effectiveness in managing structured data relationships and optimizing feature sets within multi-level classification tasks. Conventional feature selection methods, like filter, wrapper, and embedded approaches, often struggle to capture the complex, layered interactions between features in datasets where hierarchical relationships exist. This has led to the development of hierarchical feature selection techniques, among which FCA and DT have been widely applied. While FCA and DT form a foundational basis for hierarchical selection, other methods have emerged, such as fuzzy rough sets and multi-granularity clustering structures, which provide additional insights. For instance, the fuzzy rough sets method combines the robustness of rough set theory with the flexibility of fuzzy logic, allowing for the handling of uncertain or imprecise data in a hierarchical format. Studies show that fuzzy rough sets are particularly effective for data with varying levels of uncertainty across different hierarchical layers, making them well suited for classification tasks where feature relevance may be ambiguous or context-dependent [5]. The multi-granularity clustering structures approach organizes features at different granularity levels, allowing for a flexible clustering of features based on varying levels of detail. Multi-granularity structures can adapt to both broad and specific feature relationships, providing a more nuanced hierarchy than flat clustering methods. They have shown promise in applications requiring both high-level summaries and detailed feature differentiation. These alternative methodologies highlight the flexibility of hierarchical models in capturing feature dependencies across multiple layers, further emphasizing the limitations of traditional, non-hierarchical methods in structured classification tasks [12]. Recent advancements have explored hybrid methodologies that combine hierarchical approaches with other selection techniques to enhance the adaptability and performance of feature selection [16]. The hybrid hierarchical-filter methods such as information gain or chi-square allow for the quick preliminary reduction of features before building a hierarchical structure. This hybrid approach leverages the strengths of each method, achieving both computational efficiency and structured feature prioritization. Hybrid methods incorporating fuzzy logic into hierarchical models, such as fuzzy FCA, enhance flexibility in handling complex and imprecise feature relationships. This advancement allows for more nuanced feature selection in datasets where feature relevance is context-dependent or fluctuates across hierarchical levels.
All the above-mentioned studies assumed specific relationships between categories for hierarchical regularization. However, most of the existing hierarchical feature selection methods are not robust when dealing with the inevitable data outliers, resulting in a serious inter-level error propagation problem in the classification process that follows. Thus, FCA is applied in this current study because the previous experimental studies have shown that FCA is not sensitive to outliers [7]. FCA provides a well-defined mathematical framework to discover implicit and explicit knowledge in an easily understood format by using formal context and a Hasse diagram that clearly represents the concepts’ generalization/specialization relationships. In addition, FCA can construct an informative concept hierarchy providing valuable information on various specific domains. To support this work, Trabelsi et al. [15] also applied the FCA method to present a new filter for feature selection, called H-Ratio, which can identify pertinent features from data based on the Shannon entropy, also known as the diversity Shannon index, which reflects how many different types there are in a dataset to select features without considering classification accuracy. However, the classification accuracy is an important aspect of selection. Thus, this work applied the classification accuracy to formulate for feature selection from concept lattice.

3. The Proposed Models for Feature Selection

FCA uses a concept lattice structure to systematically organize features based on their generalization and specialization relationships. By identifying clusters of features, FCA can pinpoint attributes that contribute significantly to classification accuracy. In this hierarchical structure, each level of the concept lattice corresponds to a specific set of feature combinations that add distinct value to the classification task. Features that do not contribute to these essential relationships are pruned, allowing FCA to retain a streamlined set of features that still preserves the necessary classification information. This lattice-based approach reduces the dimensionality of the feature set without compromising accuracy, as FCA naturally filters out features that may add noise or redundancy to the model. The retained features are those that contribute most to accurate classification outcomes, ensuring that the reduced feature set is both efficient and effective.
The proposed approaches based on hierarchical models focused on two methods, namely FCA and DT. FCA and DT were applied to build the hierarchical structures for feature selection from data so that each node in the structure represented a feature (attribute). The choice of FCA and DT for hierarchical feature selection was motivated by their ability to structure features in a multi-level hierarchy. Unlike traditional flat feature selection methods, which may overlook complex relationships between features, FCA and DT enable the preservation of feature dependencies across levels. This hierarchical organization allows for more informed feature selection, enhancing the model’s classification accuracy and interpretability, especially in datasets with intricate class structures. An overview of the proposed models is depicted in Figure 1.
In developing the hierarchical models for feature selection using FCA and DT, we make the following assumptions: (1) We assume that the dataset contains hierarchical relationships among features that can be effectively captured through a concept lattice or decision tree structure. This structure is assumed to reflect generalization and specialization among features, allowing us to reduce feature dimensionality meaningfully. (2) Each level in the hierarchical model is assumed to contain features that independently contribute to classification accuracy. Redundant features within a level are minimized under the assumption that the FCA and DT structures accurately filter irrelevant or less useful attributes. (3) It is assumed that the feature contributions to classification accuracy remain consistent across the hierarchy. This means that a feature’s impact at one level should either be retained or improved at higher levels. This allows us to optimize feature selection sequentially through the hierarchy.
Initially, the hierarchical structure form was built, where the top node represents general knowledge, whereas the bottom represents specific knowledge. Next, the top node of structure was chosen by considering each level of structure. In this study, we selected a feature for each level of the structure, and then the selected feature was used to represent data for classification in the next step of evaluating the performance.
Afterward, the relationship between the number count of features (attributes) and classification accuracy was considered. We found that the number of attributes for each level affects classification accuracy at some level. We attempted to optimize the feature selection from the hierarchical structure. Thus, the paired-samples t-test was applied to choose appropriate features for each level. This approach has advantages, including the estimation of mean differences between paired observations and the identification of statistically significant changes in two related measurements. The paired design helps manage individual differences, minimizing the impact of confounding variables and optimizing data economy by leveraging the paired nature of observations. Furthermore, it serves as an efficient method for testing differences between two related conditions or measurements, offering a straightforward and practical means for comparing means within the same group. We performed feature selection based on a hierarchical structure as follows.
On level l of hierarchical structure, for any l { 1 , 2 , , k } , we defined the total accuracy as follows.
T o t a l   A c c l = i = 1 n A c c l , i n
where A c c l , i { 0 , 1 } is the ith accuracy on level l of the decision tree, and the values 1 and 0 mean that prediction of A c c l , i is true or false, respectively. To find the set of chosen levels, we denoted by S o C L where level 0 is definitely chosen. Next, we consider sequentially the next level of decision tree by using the two-samples z test between total accuracy on level l and l 1 defined as
Z l , l 1 = T o t a l   A c c l T o t a l   A c c l 1 p ^ q ^ 1 n + 1 n
= n 2 p ^ q ^ T o t a l   A c c l T o t a l   A c c l 1
where p ^ = i = 1 n A c c l , i + i = 1 n A c c l 1 , i n + n and q ^ = 1 p ^ for any l { 1 , 2 , , k } . The level l will be added to set S o C L depending on the level of significance α . This means we use α to consider the significant difference between total accuracy of level l and l – 1. If the total accuracy of level l is significantly different from the total accuracy of level l – 1, we will add l into the set S o C L . Otherwise, we do not add l into the set S o C L and stop the level finding process. Table 1 shows some values of the significance level α and their corresponding confidence intervals for z.
For example, when we consider the level of significance α =0.01, if | Z ( l , l 1 ) | 2.576 , then we add l to the set S o C L . Otherwise, l is not added to set S o C L , and we stop the process. We denote the set of chosen attributes by S o C A . If m is the last level added to the set S o C L , we have set S o C L = { 0 , 1 , 2 , , m } and set S o C A = i = 0 m A t t r i , where A t t r i is the set of all attributes on level i. We present the proposed feature selection algorithm using FCA and decision tree, presented as Algorithm 1.
Algorithm 1: Feature selection for FCA and decision tree.
Input     : The training dataset comprises all attributes and n instances.
               : Tree in decision tree or concept lattice including k level.
               : α and z // for example: α = 0.01 , z = [ 2.576 , 2.576 ]
Output  : Set of selected attributes ( S o C A ).
Method :
                1. S o C L = { 0 } , l = 0
                2. S o C A = { A t t r l }
                3. T o t a l A c c [ l ] = i = 1 n A c c l , i n // Accuracy of using selected attributes in level l
                4. For + + l to k // l=1 to k
                5.     p ^ = ( i = 1 n A c c l , i + i = 1 n A c c l 1 , i ) / ( n + n )
                6.     q ^ = 1 p ^
                7.     Z l , l 1 = n 2 p ^ q ^ T o t a l A c c [ l ] T o t a l A c c [ l 1 ]
                8.     If ( | Z ( l , l 1 ) | 2.576 )
                9.        S o C L = { 0 l }
                10.        S o C A = i = 0 l A t t r i
                11.     Else Break();
                12. End For
                13. Return ( S o C A )

4. Research Methodology

4.1. Dataset Description and Tools

This study obtained all required datasets from the UCI machine learning repository [29]. These datasets belong to many different fields, such as engineering, social science, and business. Among them, we selected 11 datasets covering many areas for our evaluation. Table 2 shows the different samples of data used. We also provide the number of attributes, instances, and classes. Moreover, in the final column, we describe the attribute data type in each dataset. These data are used to demonstrate the applicability and performance of feature selection. Afterwards, classification is used to evaluate the accuracy when each dataset is randomly split into training and test sets.
This study used R for data management and analysis, specifically R-3.5.1 for Windows and RStudio Version 1.1.456. They were used to classify using the DT, SVM, and ANN classifiers. We also used Weka to provide the previous feature selection algorithms compared to our models. The feature selection methods tested were filter methods (information gain and chi-square), wrapper methods, and forward selection. In addition, in the proposed model using FCA for feature selection, we applied Concept Explorer (ConExp) [30] to generate the hierarchical structure for FCA. ConExp has the functionality to create and edit contexts, generating lattices from contexts and finding the relative attribute that actually occurs in the context and the basic relationship that is true in the context.

4.2. The Experimental Design

We designed experiments for evaluating the performance in classification tasks. The overall scheme is illustrated in Figure 2.
In these numerical experiments, datasets from the UCI Machine Learning Repository (Asuncion, 2007) were employed, each divided into training and testing groups. The training set was processed with feature selection using three methods: the proposed models, prior models, and the original features. Initially, the proposed model involves the utilization of FCA and decision tree within a hierarchical concept model. These structures were created, and significant features were selected, as explained in Section 3. Subsequently, traditional feature selection methods such as IG, CS, FS, and BS were employed to compare the proposed feature selection model and the original features. These conventional methods are widely utilized for feature selection and operate based on filter (e.g., IG and CS) and wrapper (e.g., FS and BS) methods. Finally, the original features were used as a baseline for comparison. Next, the selected feature(s) for each dataset were used to train a classifier, namely DT, SVM, and ANN. These models are popular classifier models. Afterward, the test data were used to evaluate the classification accuracies of these classifiers. The accuracies served as the performance measures for each model.
Subsequently, the chosen feature(s) for each dataset were employed to train classifiers, specifically, DT, SVM, and ANN. These models are well known in the field of classification. Following the training phase, the test data were utilized to assess the classification accuracies of these classifiers. The accuracies then served as performance metrics for each model.

5. Results and Discussion

5.1. The Original Performance Results

Practically, the datasets in Table 2 were run using 10-fold cross-validation, which is the most used method. Cross-validation metrics further validated the stability and reliability of our selected features, providing key metrics such as mean accuracy, F1-score, and standard deviation. In this work, we selected accuracy as the primary metric to gauge the stability of model performance across different data splits. Accuracy was chosen for its straightforward interpretation and its widespread use as an effective measure of a model’s ability to correctly classify instances across classes. For this, each dataset was divided into complementary subsets, performing the analysis on one subset, called the training set, and validating the analysis on the other subset, called the testing set. The results from our experiments for classification accuracy are shown in Table 3. To improve the interpretability of the classification accuracy results, we provide a color-coded heatmap visualization for Table 3 shown in Figure 3. This heatmap illustrates classification performance across different classifiers—DT, SVM, and ANN—for each dataset. Darker shades indicate higher classification accuracy, making it easier to identify which classifiers performed best for each dataset at a glance.
In this experiment, the input data had not been reduced; instead, all original attributes were used for training. In Table 3, the results show the standard classification accuracy for each dataset when using DT, SVM, and ANN. Mostly, DT was the superior one. The accuracy results are used as benchmarks for evaluation the feature selection in the next section.

5.2. The Hierarchical Concept Model Performance

This section describes the proposed model for feature selection based on hierarchical concept model using DT or FCA. Thus, the results will be divided into three parts as follows.

5.2.1. Feature Selection Using Decision Tree

In order to evaluate our models, we used the well-known software RapidMiner Studio to generate a tree structure using the DT algorithm. The datasets in Table 2 were used to create tree structures. Next, we selected the attributes from each level of the tree structure. In this experiment, we selected attributes by filtering from the first to the fifth level of the tree. Table 4 shows the classification accuracy (%) in each level of the decision tree. We compared classification accuracy with our feature selection using DT to the benchmark with original features (without using feature selection at all) in Section 5, as shown in Figure 4.
From Figure 4, we see that the classification accuracy did not differ between the original and the proposed models for each level. Indeed, if we add the number of attributes from each level, it leads to a trend of improving accuracy.

5.2.2. Feature Selection Using FCA

FCA’s ability to create a structured concept lattice enables it to systematically reduce the number of features while retaining classification robustness. Within the concept lattice, FCA organizes features according to their generalization and specialization relationships, which helps distinguish between essential and redundant attributes. Features that consistently contribute to accurate classification outcomes are preserved within the hierarchical structure, while those that do not support the classification process are pruned. This lattice-based hierarchy allows FCA to efficiently identify a minimal yet comprehensive feature subset that captures the necessary information for classification. As a result, the feature set derived using FCA maintains a level of classification accuracy comparable to the original dataset while achieving significant dimensionality reduction. This is particularly beneficial in large datasets, where FCA’s structure can simplify models, reducing computational load without compromising accuracy.
The other hierarchical method, FCA, was then applied to generate the hierarchical construct. We applied the ConExp program to build the concept lattice structure. Similarly, we selected the attributes from each level of concept lattice from the first to the fifth level. The classification accuracy results are shown in Table 5. From this table, we generate the color-coded heatmap showed as Figure 5 showing classification accuracy across various levels (LV.1 to LV.5) for different classifiers (DT, SVM, ANN) using FCA-based feature selection. This visualization highlights the classification performance across datasets, with darker shades representing higher accuracy, facilitating a quick comparison across levels and classifiers.
The results show that as the levels increase from LV.1 to LV.5, certain datasets, such as Pageblocks, Movement, and Tunadromd, show improvements in classification accuracy. This pattern suggests that higher levels in the FCA hierarchy incorporate additional relevant features, enhancing the classifiers’ ability to make accurate predictions. This improvement across levels aligns with FCA’s role in selecting more specific feature combinations as the hierarchy expands, allowing for more refined data representation. For datasets like Dermatology and Zoo, accuracy remains relatively stable across all levels, with minor fluctuations. This stability indicates that FCA-based selection effectively identifies a core set of important features early on, and additional levels do not significantly alter the classification performance. This result suggests that, for these datasets, FCA’s early levels capture the most relevant information, making deeper feature selection levels less impactful. The heatmap highlights variability in performance across different classifiers. For instance, ANN achieves high accuracy on datasets such as Segmentation and Soybean at all levels, while SVM and DT sometimes lag in accuracy. This variability underscores the adaptability of FCA-based selection: certain classifiers, especially those more sensitive to specific features or complex interactions (like ANN), benefit more from the hierarchical structure than others. In some cases, specific levels offer optimal feature sets. For example, Glass and Iris datasets achieve high accuracy with minimal levels (LV.1 or LV.2), indicating that these datasets require only a few select features for effective classification. This observation supports the advantage of FCA-based selection in balancing dimensionality reduction with performance, especially for simpler datasets where adding more features would not yield substantial accuracy gains. A few datasets, such as Lung-cancer and Zoo, demonstrate lower accuracy across most levels, particularly for certain classifiers. This pattern may reflect inherent complexities in these datasets that are not fully captured by FCA-selected features or limitations in the classifiers used. It suggests that alternative selection strategies or more complex models may be required to handle such cases effectively.
Likewise, we compared classification accuracy of our feature selection using FCA to the benchmark original features (without using feature selection at all) in Section 5.1, as shown in Figure 6.
From Figure 6, we see that the classification accuracy is not different between the original and the proposed model at each level. Indeed, if we add the number of attributes from each level, the trend in accuracy is increasing.
From the testing above, we find that the classification accuracy is not different between the original and the proposed model for each level. Thus, we can compare using paired-samples t-tests as mentioned in Section 3.
On considering the relationship between the number of selected attributes and the average classification accuracy for datasets, as shown in Figure 7, we found that the proposed model using a decision tree gave a better classification result than the model using FCA.

5.2.3. An Example of Feature Selection Optimization

In our feature selection, we select the top node of structure to be representative of dataset. However, the proper selection of relevant attribute leads to an efficient classification. For example, in Table 4 and Table 5, we chose datasets (Dermatology, Glass, and Iris) and applied three classifiers (DT, SVM, and ANN) to illustrate a line plot of classification accuracy across hierarchical levels (Levels 1 to 5), as shown in Figure 8. This visualization demonstrates how classification performance improves with higher hierarchical levels, providing insight into the incremental impact of feature selection at each level for each classifier and dataset. Both graphs show that as the hierarchical level increases, accuracy gradually improves. However, after reaching a certain level, the accuracy plateaus and remains constant. By systematically adjusting the number of hierarchical levels (from LV.1 to LV.5) and observing corresponding accuracy changes, we gained insights into the optimal level configurations for each dataset. Specifically, we used paired-samples statistical tests to compare accuracy across levels, ensuring that each selected level added meaningful improvement without redundant features. Thus, we experimented with using the paired-samples t-test to optimize the selected level. We use the experimental results from Table 4 to demonstrate the proper selection of the node attribute in each level. The Dermatology dataset was investigated to show the calculation for selecting attribute in each level of the tree. We assume the hierarchical structure of the Dermatology dataset, and the classification accuracies are shown in Figure 9.
From Figure 9, we can apply the proposed feature selection in an example calculation. The initial attribute node (or root) is selected with an accuracy of 64.94%. Next, we consider the next step for selection or not by using the Equation (2) where n= 366, p ^ = l = i n A c c l , i + l = i n A c c l 1 , i n + n = 245 + 270 366 + 366 = 0.704 , q ^ = 1 p ^ = 0.296 , thus
Z l , l 1 = 366 22 0.704 0.296 73.77 64.94 = 42.18 .
From Table 1, if we consider the level of significance 0.01, if | Z l , l 1 | 2.576 , then we select level 1 (LV.1 in Figure 9). Next, we still consider the next level (LV.2). We calculate z 2 , 1 = 1.61 and find that this value is less than z in Table 1. Thus, LV.2 is not selected, and we stop the process. Thus, in this structure, we choose attributes from the structure in LV.0 to LV.2.
In addition, we estimate for all datasets which levels should be chosen. We applied the statistical paired-samples t-test to estimate this across all the datasets shown in Table 6, where x is the mean of the accuracy values of the data set for each level and S D is the standard deviation. We find that the feature selection from each level depends on the classifier model. Namely, we can use two levels in a decision tree classifier, whereas we should select four levels for using SVM or ANN classifiers.

5.3. The Performances of Prior Feature Selection Approaches

This section evaluates the classification performance of feature selection using IG, CS, FS, and BS. IG and CS algorithms function as filter methods, while FS and BS algorithms operate as wrappers. These algorithms were implemented in the RapidMiner Studio program, representing traditional feature selection methods for comparison with our proposed model. The 10 datasets listed in Table 2 were employed for training and testing through a 10-fold cross-validation approach. The experimental results, presented in Table 7, present the use of DT, SVM, and ANN as classifiers, with the “No.” column indicating the number of selected attributes. The classification accuracies reported in this table will be further compared in the subsequent section.
To improve clarity and facilitate comparisons, we provide a color-coded heatmap visualization for Table 7, shown in Figure 10. This heatmap represents the classification accuracy results across various feature selection methods (IG, CS, FS, BS) and classifiers (DT, SVM, ANN) for each dataset. The color intensity in each cell corresponds to the accuracy level, with darker shades indicating higher accuracy. Figure 10 illustrates the classification performance for each method and classifier on different datasets. This visual format allows for quick identification of the most effective feature selection method and classifier combination for each dataset. The heatmap makes it easier to compare performance across different configurations, addressing a reviewer’s suggestion to include enhanced data interpretation.
As shown in the results, the IG and CS methods prioritize features with strong discriminatory power, which directly boosts classifier performance, particularly in datasets with well-defined feature relevance, such as Dermatology and Glass. IG, in particular, enhances the performance of DT and SVM by focusing on features that most impact classification decisions. Similarly, FS and BS methods refine the feature space by iteratively selecting or removing features based on their importance. This targeted approach leads to high accuracy in datasets like Pageblocks and Segmentation, where complex feature relationships are effectively reduced to only the most critical attributes for classification. Feature importance also varies across classifiers. For instance, DT benefits from feature selection methods that retain features with hierarchical significance, aligning well with its split-based structure. In contrast, SVM performs best with methods that maintain a broad set of influential features, such as those identified by CS. ANN, on the other hand, benefits from a more nuanced feature selection approach, as seen in the results for Soybean and Wine. The layered structure of ANN models can leverage diverse feature sets, making it especially effective with IG and FS methods that capture non-linear relationships within the data. The highest classification accuracy occurs when feature importance aligns with each classifier’s inherent strengths. For example, datasets with high inter-feature relevance perform well with BS and FS, which selectively retain essential features while discarding irrelevant ones. This alignment between feature importance and classifier capabilities ensures that the chosen features directly support classification goals, resulting in improved model performance.

5.4. Comparison of Performances

We compared the classification accuracies of our feature selection using decision trees in Section 5.2 with others in Section 5.3, as illustrated in Figure 11. The figure depicts the classification performance after utilizing different numbers of attributes at each level (from Level 1 to 3, from top to bottom respectively) for each classification method denoted as Figure 11a–c, respectively. Detailed data can be found in Table 4 and Table 7. From the graphs in each dataset, considering the accuracy of attribute selection at each level, it appears that the performance is relatively consistent with other feature selection methods. For some datasets, the presented method may exhibit superior accuracy from the initial level, such as the Soybean dataset and Segmentation dataset. Conversely, the presented method may have lower accuracy for certain datasets, as observed in the Dermatology dataset.
Similarly, we conducted a comparison of the classification accuracy of our models using FCA in Section 5.2 with others in Section 5.3, illustrated in Figure 12. The figure demonstrates the classification performance after employing varying numbers of attributes at each level, with detailed data available in Table 5 and Table 7. Upon examining the accuracy trends from this graph, we observe a similar pattern to use the decision tree approach, where some datasets may exhibit comparable performance to other methods, such as the Movement dataset, while others may show comparatively lower classification accuracy.
From the above results, the outcomes obtained by employing a hierarchical structure for feature selection before the classification task indicate that it could lead to either favorable or less favorable results. This corresponds to the observed patterns seen in other feature selection methods used for comparison. Importantly, the classification performance of the proposed method does not consistently appear inferior. However, the presented key point in this study is its ability to significantly reduce the number of attributes compared to previous methods. To highlight the comparative advantages of the hierarchical concept model (using FCA and DT), we compared its performance with traditional feature selection methods, such as flat filters and wrappers, across multiple datasets. Our results reveal specific cases where the hierarchical model demonstrates notable benefits over traditional approaches. The hierarchical concept model’s structured approach is particularly advantageous in datasets with layered feature relationships and complex class structures. Hierarchical methods like FCA and DT can capture multi-level dependencies between features. For instance, in datasets where features have a natural hierarchy or are interdependent, these methods outperform traditional flat methods by preserving relational structures within the data. This leads to a more meaningful reduction in feature space and improved classification accuracy. In datasets with many classes or intricate class relationships, the hierarchical approach organizes features according to relevance at each level, leading to better feature prioritization and reducing the computational load. For example, in the Soybean dataset, FCA’s concept lattice helped isolate critical features while minimizing noise, thereby enhancing performance compared to non-hierarchical methods.
Considering the count of selected features, the results show clearly that the DT retained fewer features than the other methods, as shown in Figure 13. This figure will illustrate the number of selected attributes at each level from 1 to 3 in comparison with other methods that exhibit classification performance as shown in Figure 6 and Figure 7 mentioned earlier. Figure 13a displays the count of selected attributes from the decision tree structure, while Figure 13b shows the count of selected attributes from the FCA structure. From as the results, the hierarchical structure using decision tree method can be applied to select the feature for data dimension reduction yields favorable outcomes. This is due to the decision tree’s method, which involves a top-down data entry process using specific filtering and selecting root node. However, FCA produces less satisfactory results due to the creation of a layer-wise structure designed for bidirectional ordering—both from top to bottom and from bottom to top. This may lead to effect of feature selection. The experimental results confirm that FCA’s concept lattice can achieve dimensionality reduction without compromising classification accuracy. By systematically identifying and retaining only the most informative features, FCA produces a reduced feature set that achieves comparable accuracy to the original, full feature set. This performance validates FCA’s effectiveness in filtering out features that are irrelevant or redundant while preserving those that contribute to robust classification outcomes. Compared to traditional feature selection methods, FCA’s lattice structure organizes attributes in a way that maintains high classification reliability while minimizing the feature set. This approach optimally balances computational efficiency and model interpretability with performance, making FCA an ideal choice for applications that require both robust classification and efficient feature reduction.
While the hierarchical feature selection model demonstrated robust performance with DT, SVM, and ANN classifiers, certain conditions could further enhance its results. For instance, tuning the parameters for each classifier, such as adjusting hyperparameters in SVM or layer configurations in ANN, could improve classification accuracy. Moreover, increasing the hierarchical levels of feature selection allows finer distinctions among features, which can be beneficial when working with high-dimensional datasets. This model can be adapted to other classification algorithms, such as k-nearest neighbors (k-NN) and random forest, by modifying the selection criteria within the hierarchical structure. The flexibility in adapting this model to other classifiers makes it a versatile approach across different machine learning tasks.
This hierarchical feature selection model is particularly beneficial in applications requiring efficient and accurate feature reduction. For example, in medical diagnosis systems, where high-dimensional data (e.g., from genetic or imaging data) can impact computational efficiency, this model enables the reduction of features without sacrificing diagnostic accuracy. Another practical use case is in financial analysis for credit scoring, where a subset of relevant features can streamline decision-making processes. Additionally, environmental monitoring systems, where IoT data often contains redundant attributes, can benefit from this model’s ability to filter and retain only the most informative features.
However, one interesting limitation is class imbalance. Class imbalance can impact the performance of feature selection methods, as features correlated with the minority class may be underrepresented or overlooked during selection. To mitigate this, we employed re-sampling based on using cross-validation metrics to ensure balanced representation across classes. However, this is still a limitation of this study. Our analysis indicates that while the hierarchical feature selection method performs robustly in balanced datasets, its effectiveness in imbalanced datasets improves with re-sampling techniques. Future work could explore integrating class-sensitive feature scoring directly into the hierarchical selection framework, ensuring equitable consideration of features across all classes. Moreover, we plan to address key challenges, such as multicollinearity and class imbalance, through established techniques like variance inflation factor (VIF) analysis, SMOTE for balancing classes, and comparison with advanced feature selection methods. Future studies will involve comprehensive testing on both benchmark and synthetic datasets to validate the efficacy of the hierarchical framework.

6. Conclusions

Feature selection methods are used in a wide variety of applications. Such methods proceed to eliminate unfavorable features that may be noisy, redundant or irrelevant, as these can penalize the performance of a classifier. Feature selection reduces the dimensionality of data and restricts the inputs to a subset of the original, but the features can have missing values. In this paper, we propose a new feature selection method based on a hierarchical concept model to improve the performance of supervised learning. We applied a decision tree and FCA to build the hierarchical structures from data to generate knowledge-based systems. In these hierarchical structures that embed knowledge, the top and bottom nodes represent general and specific knowledge, respectively. For this reason, we selected the top node of each structure to be representative of the dataset. However, the proper selection of relevant attributes leads to an efficient classification. Thus, we also used paired-samples t-tests to optimize the selected levels in the hierarchical structures. In this study, numerical experiments used available software, RapidMiner Studio v. 9.2.0 and ConExp v. 1.3 programs, to generate the hierarchical structures for decision tree and FCA, respectively. Afterward, the attributes in top node of the obtained structure were selected to use in learning classifier models. Three popular classifier algorithms were used in this study: decision trees, SVM, and ANN. To compare the performances of our feature selection methods with available alternatives: all original features (without using feature selection) and the prior feature selection methods. We used 10 datasets from the UCI Machine Learning Repository for classification. The prior feature selection methods tested were IG, CS, FS, and BS. The results show clearly that the proposed models with DT and FCA gave smaller set of retained attributes than the other methods, while still classification performance remained comparable. In other words, the new approach provided superior data reduction to the least dimensionality without performance sacrifice. In summary, this study demonstrates the strengths of hierarchical feature selection methods, specifically FCA and DT, over traditional flat methods in scenarios involving complex feature interdependencies or layered class structures. The hierarchical model’s ability to retain structured relationships and prioritize relevant features enables it to outperform conventional approaches in both feature reduction and classification accuracy for multi-level datasets. This makes hierarchical methods a valuable approach for feature selection in data-intensive applications requiring nuanced feature analysis.
Future work could focus on refining the hierarchical feature selection model by exploring more dynamic methods of defining hierarchical relationships among features. Integrating adaptive methods, such as reinforcement learning, to determine optimal hierarchical levels could further improve model efficiency and adaptability to diverse datasets. Additionally, investigating other classifiers, including ensemble models like gradient boosting machines (GBMs) and random forest, could provide insights into the model’s adaptability across various machine learning paradigms.

Author Contributions

Conceptualization, J.M.; Methodology, J.M. and B.C.; Software, P.W. and J.S.; Validation, A.W., A.I., J.S., L.B. and B.C.; Formal analysis, A.W., L.B. and B.C.; Writing – original draft, J.M.; Writing—eview & editing, P.W., A.W., A.I., J.S., L.B. and B.C.; Visualization, A.I.; Supervision, J.M.; Funding acquisition, J.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by National Science, Research and Innovation Fund (NSRF) and Prince of Songkla University (Ref. No. (SIT6601016S)).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

This data can be found here: https://archive.ics.uci.edu/, accessed on 6 January 2023.

Acknowledgments

The authors are deeply grateful to the Faculty of Science and Industrial Technology, Prince of Songkla University, Surat Thani Campus, Thailand. This research was supported by the National Science, Research, and Innovation Fund (NSRF) and Prince of Songkla University (Ref. No. (SIT6601016S)). The authors gratefully acknowledge the helpful comments and suggestions of the reviewers, which have improved the presentation.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Dhal, P.; Azad, C. A comprehensive survey on feature selection in the various fields of machine learning. Appl. Intell. 2022, 52, 4543–4581. [Google Scholar] [CrossRef]
  2. Khaire, U.M.; Dhanalakshmi, R. Stability of feature selection algorithm: A review. J. King Saud Univ. Comput. Inf. Sci. 2022, 34, 1060–1073. [Google Scholar] [CrossRef]
  3. Solorio-Fernández, S.; Carrasco-Ochoa, J.A.; Martínez-Trinidad, J.F. A new hybrid filter–wrapper feature selection method for clustering based on ranking. Neurocomputing 2016, 214, 866–880. [Google Scholar] [CrossRef]
  4. Cai, J.; Luo, J.; Wang, S.; Yang, S. Feature selection in machine learning: A new perspective. Neurocomputing 2018, 300, 70–79. [Google Scholar] [CrossRef]
  5. Zhao, H.; Wang, P.; Hu, Q.; Zhu, P. Fuzzy Rough Set Based Feature Selection for Large-Scale Hierarchical Classification. IEEE Trans. Fuzzy Syst. 2019, 27, 1891–1903. [Google Scholar] [CrossRef]
  6. Bolón-Canedo, V.; Alonso-Betanzos, A. Ensembles for feature selection: A review and future trends. Inf. Fusion 2019, 52, 1–12. [Google Scholar] [CrossRef]
  7. Zebari, R.; Abdulazeez, A.; Zeebaree, D.; Zebari, D.; Saeed, J. A Comprehensive Review of Dimensionality Reduction Techniques for Feature Selection and Feature Extraction. J. Appl. Sci. Technol. Trends 2020, 1, 56–70. [Google Scholar] [CrossRef]
  8. Wan, C.; Freitas, A.A. An empirical evaluation of hierarchical feature selection methods for classification in bioinformatics datasets with gene ontology-based features. Artif. Intell. Rev. 2018, 50, 201–240. [Google Scholar] [CrossRef]
  9. Wetchapram, P.; Muangprathub, J.; Choopradit, B.; Wanichsombat, A. Feature Selection Based on Hierarchical Concept Model Using Formal Concept Analysis. In Proceedings of the 2021 18th International Conference on Electrical Engineering/Electronics, Computer, Telecommunications and Information Technology (ECTI-CON), Chiang Mai, Thailand, 19–22 May 2021; pp. 299–302. [Google Scholar] [CrossRef]
  10. Hancer, E.; Xue, B.; Zhang, M. A survey on feature selection approaches for clustering. Artif. Intell. Rev. 2020, 53, 4519–4545. [Google Scholar] [CrossRef]
  11. Cerrada, M.; Sánchez, R.V.; Pacheco, F.; Cabrera, D.; Zurita, G.; Li, C. Hierarchical feature selection based on relative dependency for gear fault diagnosis. Appl. Intell. 2016, 44, 687–703. [Google Scholar] [CrossRef]
  12. Guo, S.; Zhao, H.; Yang, W. Hierarchical feature selection with multi-granularity clustering structure. Inf. Sci. 2021, 568, 448–462. [Google Scholar] [CrossRef]
  13. Tuo, Q.; Zhao, H.; Hu, Q. Hierarchical feature selection with subtree based graph regularization. Knowl. Based Syst. 2019, 163, 996–1008. [Google Scholar] [CrossRef]
  14. Zheng, J.; Luo, C.; Li, T.; Chen, H. A novel hierarchical feature selection method based on large margin nearest neighbor learning. Neurocomputing 2022, 497, 1–12. [Google Scholar] [CrossRef]
  15. Trabelsi, M.; Meddouri, N.; Maddouri, M. A New Feature Selection Method for Nominal Classifier based on Formal Concept Analysis. Procedia Comput. Sci. 2017, 112, 186–194. [Google Scholar] [CrossRef]
  16. Azibi, H.; Meddouri, N.; Maddouri, M. Survey on Formal Concept Analysis Based Supervised Classification Techniques. In Machine Learning and Artificial Intelligence; IOS Press: Amsterdam, The Netherlands, 2020; pp. 21–29. [Google Scholar]
  17. Wang, C.; Huang, Y.; Shao, M.; Hu, Q.; Chen, D. Feature Selection Based on Neighborhood Self-Information. IEEE Trans. Cybern. 2020, 50, 4031–4042. [Google Scholar] [CrossRef]
  18. Wille, R. Formal concept analysis as mathematical theory of concepts and concept hierarchies. Lect. Notes Artificial Intell. (LNAI) 2005, 3626, 1–33. [Google Scholar] [CrossRef]
  19. Zhou, H.; Zhang, J.; Zhou, Y.; Guo, X.; Ma, Y. A feature selection algorithm of decision tree based on feature weight. Expert Syst. Appl. 2021, 164, 113842. [Google Scholar] [CrossRef]
  20. Venkatesh, B.; Anuradha, J. A Review of Feature Selection and Its Methods. Cybern. Inf. Technol. 2019, 19, 3–26. [Google Scholar] [CrossRef]
  21. Bahassine, S.; Madani, A.; Al-Sarem, M.; Kissi, M. Feature selection using an improved Chi-square for Arabic text classification. J. King Saud Univ. Comput. Inf. Sci. 2020, 32, 225–231. [Google Scholar] [CrossRef]
  22. Trivedi, S.K. A study on credit scoring modeling with different feature selection and machine learning approaches. Technol. Soc. 2020, 63, 101413. [Google Scholar] [CrossRef]
  23. Liu, Z.; Zhang, R.; Song, Y.; Ju, W.; Zhang, M. When does maml work the best? an empirical study on model-agnostic meta-learning in nlp applications. arXiv 2020, arXiv:2005.11700. [Google Scholar] [CrossRef]
  24. Yang, J.; Xu, H.; Mirzoyan, S.; Chen, T.; Liu, Z.; Liu, Z.; Ju, W.; Liu, L.; Xiao, Z.; Zhang, M.; et al. Poisoning medical knowledge using large language models. Nat. Mach. Intell. 2024, 6, 1156–1168. [Google Scholar] [CrossRef]
  25. Ju, W.; Mao, Z.; Yi, S.; Qin, Y.; Gu, Y.; Xiao, Z.; Wang, Y.; Luo, X.; Zhang, M. Hypergraph-enhanced Dual Semi-supervised Graph Classification. arXiv 2024, arXiv:2405.04773. [Google Scholar] [CrossRef]
  26. Zhao, H.; Hu, Q.; Zhu, P.; Wang, Y.; Wang, P. A Recursive Regularization Based Feature Selection Framework for Hierarchical Classification. IEEE Trans. Knowl. Data Eng. 2021, 33, 2833–2846. [Google Scholar] [CrossRef]
  27. Huang, H.; Liu, H. Feature selection for hierarchical classification via joint semantic and structural information of labels. Knowl. Based Syst. 2020, 195, 105655. [Google Scholar] [CrossRef]
  28. Liu, X.; Zhao, H. Robust hierarchical feature selection with a capped ℓ2-norm. Neurocomputing 2021, 443, 131–146. [Google Scholar] [CrossRef]
  29. UCI Machine Learning Repository. Available online: https://archive.ics.uci.edu (accessed on 6 January 2023).
  30. Yevtushenko, S. Concept Explorer, Open Source JAVA Software. 2009. Available online: http://sourceforge.net/projects/conexp (accessed on 6 January 2023).
Figure 1. A conceptual overview of the proposed models.
Figure 1. A conceptual overview of the proposed models.
Applsci 14 10965 g001
Figure 2. The experimental design.
Figure 2. The experimental design.
Applsci 14 10965 g002
Figure 3. The classification accuracy comparison across DT, SVM, and ANN.
Figure 3. The classification accuracy comparison across DT, SVM, and ANN.
Applsci 14 10965 g003
Figure 4. The comparison DT-SVM-ANN classifiers with features selected using DT from the first to the fifth level and with all original features.
Figure 4. The comparison DT-SVM-ANN classifiers with features selected using DT from the first to the fifth level and with all original features.
Applsci 14 10965 g004
Figure 5. The classification accuracy using FCA-based feature selection across levels and classifiers (Table 5).
Figure 5. The classification accuracy using FCA-based feature selection across levels and classifiers (Table 5).
Applsci 14 10965 g005
Figure 6. A comparison DT-SVM-ANN classifiers with feature selection using FCA from the first to the fifth level and with all original features.
Figure 6. A comparison DT-SVM-ANN classifiers with feature selection using FCA from the first to the fifth level and with all original features.
Applsci 14 10965 g006
Figure 7. A comparison of average classification accuracies with each classifier model between features selected using DT or FCA.
Figure 7. A comparison of average classification accuracies with each classifier model between features selected using DT or FCA.
Applsci 14 10965 g007
Figure 8. The classification accuracy across hierarchical levels for selected datasets. (a) Feature selection using decision tree. (b) Feature selection using FCA.
Figure 8. The classification accuracy across hierarchical levels for selected datasets. (a) Feature selection using decision tree. (b) Feature selection using FCA.
Applsci 14 10965 g008
Figure 9. An example hierarchical structure using decision tree and the accuracies with selected features.
Figure 9. An example hierarchical structure using decision tree and the accuracies with selected features.
Applsci 14 10965 g009
Figure 10. The classification performance comparison in Table 7.
Figure 10. The classification performance comparison in Table 7.
Applsci 14 10965 g010
Figure 11. Comparison of proposed feature selection using DT on each level and others with DT-SVM-ANN classifiers.
Figure 11. Comparison of proposed feature selection using DT on each level and others with DT-SVM-ANN classifiers.
Applsci 14 10965 g011
Figure 12. Comparison of proposed feature selection using FCA in each level to others on using DT-SVM-ANN classifiers.
Figure 12. Comparison of proposed feature selection using FCA in each level to others on using DT-SVM-ANN classifiers.
Applsci 14 10965 g012
Figure 13. The number of selected features by (a) DT and (b) FCA methods in each level, compared to the other selection methods.
Figure 13. The number of selected features by (a) DT and (b) FCA methods in each level, compared to the other selection methods.
Applsci 14 10965 g013
Table 1. The mapping from α to confidence interval for z.
Table 1. The mapping from α to confidence interval for z.
α 0.010.050.1
z 2.576 z 2.576 1.96 Z 1.96 1.645 z 1.645
Table 2. Datasets used in this study.
Table 2. Datasets used in this study.
DatasetNo. AttributesNo. InstancesNo. ClassesData Type
Dermatology333666Integer
Glass102147real
Iris41503real
Lung-cancer56323Integer
Movement9136015Integer
Pageblocks1054735Integer, real
Segmentation1923107Integer, real
Soybean3530719Integer
Tunadromd24244652Integer
Wine131783Integer, real
Zoo171017Nominal, Integer
Table 3. Summary of datasets and classifier performances.
Table 3. Summary of datasets and classifier performances.
DatasetNo. AttributesThe Classification Accuracy (%)
Org-DTOrg-SVMOrg-ANN
Dermatology3384.9984.5085.09
Glass10100.095.2699.05
Iris478.6776.9877.00
Lung-cancer5680.8380.5681.00
Movement9173.0672.6972.88
Pageblocks1097.9496.0597.88
Segmention1989.0988.6388.70
Soybean3596.0095.9896.00
Tunadromd24296.6499.1099.22
Wine1383.1780.8282.98
Zoo1796.0995.5395.72
Table 4. The classification accuracy on using each level of decision tree for feature selection.
Table 4. The classification accuracy on using each level of decision tree for feature selection.
DatasetThe Classification Accuracy (%) Using Decision Tree-Based Feature Selection
LV.1 LV.2 LV.3 LV.4 LV.5
No. Attribute DT SVM ANN No. Attribute DT SVM ANN No. Attribute DT SVM ANN No. Attribute DT SVM ANN Attribute DT SVM ANN
Dermatology166.9466.9566.94373.7775.1573.50573.5087.4573.50689.6589.6589.65896.1696.1696.16
Glass1100.0082.2499.533100.0085.19100.004100.0085.1999.535100.0099.0599.057100.00100.0099.52
Iris195.3382.6795.33295.3383.8395.33396.0083.8396.00497.3397.3397.33497.3397.3397.33
Lung-cancer187.5074.1787.50387.5062.5078.12587.5074.1781.25980.8380.8378.331180.8380.8378.33
Movement167.2255.0063.61268.3351.6768.33575.5689.4471.11875.2875.2875.281780.8380.8380.83
Page-blocks196.8696.2996.53297.9499.6096.55598.0699.6097.20798.0697.4497.44998.0598.0597.44
Segmentation186.1585.6785.89287.0692.6486.67589.0091.2189.00998.0098.0098.001096.4196.4196.41
Soybean187.2370.5078.72397.8762.0097.87595.7470.0091.871198.0098.0098.001598.0098.0098.00
Tunadromd193.2893.2893.28296.6496.6496.64496.8696.7596.75696.9897.7697.42696.9897.7697.42
Wine183.7183.8283.71387.6488.2487.64587.6488.2487.64787.1687.1687.161086.5086.5086.50
Zoo196.0496.0996.04396.0476.2796.04596.0476.2796.047100.00100.00100.00898.0098.0098.00
Table 5. The classification accuracy when using each level of FCA concept lattice for feature selection.
Table 5. The classification accuracy when using each level of FCA concept lattice for feature selection.
DatasetThe Classification Accuracy (%) Using FCA-Based Feature Selection
LV.1 LV.2 LV.3 LV.4 LV.5
No. Attribute DT SVM ANN No. Attribute DT SVM ANN No. Attribute DT SVM ANN No. Attribute DT SVM ANN No. Attribute DT SVM ANN
Dermatology1377.5877.5877.582177.5877.5877.582477.5877.5877.582877.5877.5877.583377.5877.5877.58
Glass1100.0092.6099.55394.8793.9693.96593.0191.1788.381093.0191.1788.381093.0191.1788.38
Iris294.6794.0094.00373.3349.3358.67480.0084.0084.00480.0084.0084.00480.0084.0084.00
Lung-cancer597.5097.5097.50849.1755.8355.251049.1755.8355.831649.1755.8355.833249.1755.8355.83
Movement763.3360.2860.281263.3360.2860.281999.4499.1799.172899.4499.1799.173299.4499.1799.17
Page-blocks490.3985.3185.31490.3985.3185.31799.8099.8599.851099.8099.8599.851099.8099.8599.85
Segmentation899.7092.8199.701099.7091.0299.701957.3691.0297.491957.3691.0297.491957.3691.0297.49
Soybean2100.00100.001001283.5087.5087.501661.5068.0064.003561.5068.0064.003561.5068.0064.00
Tunadromd593.2893.8493.281394.9696.6497.421996.1997.8797.983496.7598.8898.884496.7599.1099.33
Wine381.5083.7683.76669.7168.5668.56978.1481.5081.501378.1481.5081.501378.1481.5081.50
Zoo788.1888.1888.181088.1888.1888.181261.5561.5559.551761.5561.5559.551761.5561.5559.55
Table 6. The paired-samples t-test results.
Table 6. The paired-samples t-test results.
MethodsLevelx ± SDPaired t-TestSig.
DTLV.186.69 ± 8.392.970.016
LV.289.14 ± 6.742.550.031
LV.391.00 ± 7.131.590.146
LV.492.43 ± 3.072.220.053
LV.592.51 ± 2.173.030.014
SVMLV.179.34 ± 3.124.880.001
LV.277.70 ± 4.074.150.002
LV.384.54 ± 3.522.850.019
LV.492.27 ± 0.942.460.036
LV.593.21 ± 0.881.560.153
ANNLV.186.28 ± 12.672.550.031
LV.288.00 ± 6.503.200.011
LV.389.51 ± 6.202.590.029
LV.492.02 ± 3.452.350.043
LV.592.85 ± 3.331.650.133
Table 7. The classification performances of prior feature selection approaches.
Table 7. The classification performances of prior feature selection approaches.
DatasetIGCSFSBS
DT SVM ANN No. DT SVM ANN No. DT SVM ANN No. DT SVM ANN No.
Dermatology92.3593.0092.351286.0787.2986.07785.7987.8584.64985.7987.8585.7933
Glass99.53100.00100.00499.5399.82100.003100.00100.00100.009100.00100.00100.003
Iris95.3395.3395.00295.3394.2295.33296.0096.0496.00196.0096.0496.673
Lung-cancer68.5068.5069.00287.5086.5086.501687.5087.5083.505587.5087.5087.502
Movement66.6466.6565.451675.0074.5176.023476.9482.5077.858879.4482.5083.336
Pageblocks97.9297.9297.80498.1997.6398.02598.1498.0697.94898.3298.1298.064
Segmentation90.9189.3390.871088.6188.6188.52789.2289.4880.78689.4889.5297.3610
Soybean97.87100.0098.8714100.00100.0098.8710100.00100.0097.872100.00100.00100.0034
Wine82.5882.5082.00583.1583.1584.22685.9688.2290.45987.0888.4889.8910
Zoo99.0099.50100.00895.0996.5696.00496.0496.08100.00696.0496.0896.0416
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Saelee, J.; Wetchapram, P.; Wanichsombat, A.; Intarasit, A.; Muangprathub, J.; Boongasame, L.; Choopradit, B. Enhanced Feature Selection via Hierarchical Concept Modeling. Appl. Sci. 2024, 14, 10965. https://doi.org/10.3390/app142310965

AMA Style

Saelee J, Wetchapram P, Wanichsombat A, Intarasit A, Muangprathub J, Boongasame L, Choopradit B. Enhanced Feature Selection via Hierarchical Concept Modeling. Applied Sciences. 2024; 14(23):10965. https://doi.org/10.3390/app142310965

Chicago/Turabian Style

Saelee, Jarunee, Patsita Wetchapram, Apirat Wanichsombat, Arthit Intarasit, Jirapond Muangprathub, Laor Boongasame, and Boonyarit Choopradit. 2024. "Enhanced Feature Selection via Hierarchical Concept Modeling" Applied Sciences 14, no. 23: 10965. https://doi.org/10.3390/app142310965

APA Style

Saelee, J., Wetchapram, P., Wanichsombat, A., Intarasit, A., Muangprathub, J., Boongasame, L., & Choopradit, B. (2024). Enhanced Feature Selection via Hierarchical Concept Modeling. Applied Sciences, 14(23), 10965. https://doi.org/10.3390/app142310965

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop