Next Article in Journal
Archimedean Copulas-Based Estimation under One-Parameter Distributions in Coherent Systems
Next Article in Special Issue
Enhancing Medical Decision Making: A Semantic Technology-Based Framework for Efficient Diagnosis Inference
Previous Article in Journal
A Routing Model for the Distribution of Perishable Food in a Green Cold Chain
Previous Article in Special Issue
An Out-of-Distribution Generalization Framework Based on Variational Backdoor Adjustment
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Knowledge Granularity Attribute Reduction Algorithm for Incomplete Systems in a Clustering Context

1
Guangxi Key Lab of Multi-Source Information Mining & Security, Guangxi Normal University, Guilin 541004, China
2
School of Computer Science and Engineering, Guangxi Normal University, Guilin 541004, China
3
School of Computer and Artificial Intelligence, Chaohu University, Hefei 238000, China
4
CBICA, University of Pennsylvania, Philadelphia, PA 19104, USA
*
Author to whom correspondence should be addressed.
Mathematics 2024, 12(2), 333; https://doi.org/10.3390/math12020333
Submission received: 22 December 2023 / Revised: 16 January 2024 / Accepted: 18 January 2024 / Published: 19 January 2024

Abstract

:
The phenomenon of missing data can be seen everywhere in reality. Most typical attribute reduction models are only suitable for complete systems. But for incomplete systems, we cannot obtain the effective reduction rules. Even if there are a few reduction approaches, the classification accuracy of their reduction sets still needs to be improved. In order to overcome these shortcomings, this paper first defines the similarities of intra-cluster objects and inter-cluster objects based on the tolerance principle and the mechanism of knowledge granularity. Secondly, attributes are selected on the principle that the similarity of inter-cluster objects is small and the similarity of intra-cluster objects is large, and then the knowledge granularity attribute model is proposed under the background of clustering; then, the IKAR algorithm program is designed. Finally, a series of comparative experiments about reduction size, running time, and classification accuracy are conducted with twelve UCI datasets to evaluate the performance of IKAR algorithms; then, the stability of the Friedman test and Bonferroni–Dunn tests are conducted. The experimental results indicate that the proposed algorithms are efficient and feasible.

1. Introduction

Rough set theory (RST) [1], initiated by Pawlak, is an effectively mathematical tool to deal with imprecise, fuzzy, and incomplete data. RST has been successfully applied in machine learning [2,3,4], knowledge discovery [5,6], expert system [7], disease diagnostics [8,9,10], decision support [11,12,13], and other areas [14,15]. Attribute reduction is one of the research hotspots in RST. As an important technology in the process of data preprocessing, attribute reduction has captured researchers’ attention in big data and knowledge discovery [16,17,18]. The main objective of attribute reduction is to remove some irrelevant or non-important attributes while keeping the original distinguishing ability unchanged. In this way, the effect of data dimension reduction can be achieved, and a lot of time and space resources can be saved for the process of knowledge discovery and rule extraction.
With the rapid development of network and information technology, we have gradually entered the era of big data. The datasets have the characteristics of large volume, rapid change, and diverse data forms [19]. At the same time, due to the influence of the collection method and environment during the data collection process, there are a large number of missing data or wrong data in the datasets. The existence of these disturbed data will seriously affect the decision making and judgement of big data, and even mislead decision makers. After a long period of unremitting endeavor, scholars have achieved outstanding results in attribute reduction [20,21,22,23,24,25,26]. For example, Dai [20] proposed a semi-supervised attribute reduction based on attribute indiscernibility. Cao [21] put forward a three-way approximate reduction approach by using information-theoretic measure. Yang [22] presented a novel incremental attribute reduction method via quantitative dominance-based neighborhood self-information. Lin [16] et al. developed a feature selection way by using neighborhood multi-granulation fusion. In a variable precision rough set model, Yu [24] et al. raised a novel attribute reduction based on local attribute significance. In the literature [27], Devi proposed a new dimension reduce technology by considering the picture of fuzzy soft matrices in the decision-making process. Wen [28] et al. raised an unsupervised attribute reduction algorithm for mixed data based on fuzzy optimal approximation set. These above-mentioned classic reduction models are suitable only for complete systems.
As with most of the datasets in various real-world applications, the classical rough set model defined with equivalence relation leads to the limitation in handling data in incomplete systems. The important attributes cannot be selected correctly, which leads to a decrease in the classification accuracy of the reduction set. In order to reduce the incomplete system, Zhang and Chen proposed a lambda-reduction approach based on the similarity degree respect to a conditional attribute subset for incomplete set-valued information systems [25]. For incomplete interval-valued information systems, Li [29] proposed the concept of similarity degree and tolerance relation between two information values of a given attribute. Then, three reduction algorithms based on theta-discernibility matrix, theta-information entropy, and theta-significance were designed. Liu introduced a new attribute reduction approach by using conditional entropy based on the fuzzy alpha-similarity relation [30]. Subsequently, Dai [31] proposed interval-valued fuzzy min–max similarity relations and designed two attribute reduction algorithms based on interval-valued fuzzy discernibility pairs model. Song [32] put forward the similarity degree between information values on each attribute and an attribute reduction method was designed by using information granulation and information entropy. Zhou presented a heuristic attribute reduction algorithm with a binary similarity matrix and attribute significance as heuristic knowledge under incomplete information systems [33]. Zhang [34] presented a novel approach for knowledge reduction by using the discernibility techniques in multi-granulation rough set model. He and Qu [35] put forward the fuzzy-rough iterative computation model based on symmetry relations for an incomplete categorical decision information system. Srirekha et al. proposed an attribute reduction in SE-ISI concept lattice based on the concept of object ranking for incomplete information systems [36]. Cornelis et al. put forward a generalized model of attribute reduction using fuzzy tolerance relation within the context of fuzzy rough set theory [37]. Liu applied the concept of accurate reduction and reduced invariant matrix for reducing attribute under information systems [38]. To reduce unnecessary tolerance classes for the original cover, Nguyen [39] introduced a new concept of stripped neighborhood covers and proposed an efficient heuristic algorithm in mixed and incomplete decision tables.
Although the above reduction algorithms can effectively reduce incomplete information systems, the classification accuracy of the reduction set is not ideal. The main reason is that only the importance of attributes is considered when selecting attributes, and the impact of attributes on classification is not considered. Usually, people take the best result of clustering as a reference standard for classification work and classify similar samples into the same cluster. In order to solve the problems mentioned above, this paper proposes an attribute reduction method from the perspective of clustering.
At present, the studies of attribute reduction on using the clustering idea to construct a feature selection model are relatively infrequent. In order to avoid or reduce the loss of some original information after discretizing continuous values, Zhang [40] proposed a feature selection method based on fuzzy clustering, but this method has no reliable theoretical support. Jia [41] proposed a spectral clustering method based on neighborhood information entropy feature selection, which uses a feature selection method to remove redundant features before clustering. In order to take the classification effect of the dataset into consideration when selecting features, Zhao proposed a fuzzy C-Means clustering fuzzy rough feature selection method [42], which can improve the classification accuracy of the reduction set to a certain degree, but the effect is not obvious. Jia proposed a similarity attribute reduction in the context of clustering in the literature [43], which can greatly improve the classification accuracy of the reduction set but needs to continuously adjust the parameters to achieve the best classification effect. Such a feature set has certain random limitations and increases the time consumption of the system. Therefore, it is necessary to design a stable model with high classification accuracy for data preprocessing.
Although these existing approaches can effectively reduce incomplete systems, they only consider the importance of the attributes themselves, and do not consider the correlation between attributes. The influence of conditional attributes on decision classification is not considered. In order to improve the classification accuracy of a reduction set, we apply the idea of clustering.
Based on the principle that the similarity of samples within a cluster is as large as possible and the similarity of samples between clusters is as small as possible, an attribute reduction algorithm for an incomplete system is designed under the background of clustering. First, according to the principle of tolerance and the theory of knowledge granularity, we define the similarity of intra-cluster and inter-cluster for an incomplete system. Secondly, a formula for calculating the similarity of intra-cluster and inter-cluster objects is designed. After normalizing the two similarities, we define the similarity of objects. Then, according to the corresponding similarity mechanism, a new attribute reduction algorithm for an incomplete system is proposed. Finally, a series of experiments have verified that the proposed algorithm in this paper is significantly better than other similar algorithms in terms of running time, accuracy, and the stability of algorithm was analyzed by using Friedman test and Bonferroni–Dunn test in statistics.
The contribution of this paper is embodied in the following four aspects:
(1)
A tolerance class calculation in incomplete information systems is proposed and applied to knowledge granularity calculation.
(2)
Knowledge granules are used as a measure of sample similarity to measure the similarity of inter-cluster samples and intra-cluster samples.
(3)
A knowledge granularity reduction algorithm based on clustering context is designed in incomplete information systems.
(4)
Lots of experiments are conducted to verify the validity of the algorithm proposed in this paper, and the stability of the algorithm is verified by mathematical statistics.
The other parts of this paper are constructed as follows. The principle of tolerance and related concepts of knowledge granularity are recalled in Section 2. In Section 3, we propose a similarity measure of intra-cluster and inter-cluster objects and discuss the reduction mechanism according to the clustering background for missing dataset. We normalize the similarity of inter-cluster and intra-cluster, and then design the corresponding reduction model in Section 4. In Section 5, a series of experiments are conducted and the performance of the algorithm is evaluated from the reduction size, running time, classification accuracy, and stability. Then, the feasibility and effectiveness of the algorithm are verified. Finally, the advantages and disadvantages of the algorithm proposed in this paper are concluded and unfolded in the future work.

2. Preliminaries

In this section, we review some basic concepts in rough set theory, the definitions of tolerance class, knowledge granularity, clustering metrics, and the significance of attribute for incomplete decision systems.

2.1. Basic Concept of RST

A decision information system is a quadruple D S = ( U , A , V , f ) , where U is a non-empty finite set of objects and A is a finite nonempty attribute sets; if A = C D , where C is the conditional attribute sets and D is the decision attribute set; then, V is the union of attribute domains, V = a A V a , V a is the value set of attribute a , called the domain of a ; f : U × A V is an information function with f ( x , a ) = V a for each a A and x U . For every attribute subset B C , a indiscernibility relation is defined as follows:
I N D ( B ) = { ( x , y ) U × U | a B , f ( x , a ) = f ( y , a ) } .
By the relation I N D ( B ) , we can obtain the partition of U denoted by U / I N D ( B ) or U / B . If B A X U , the upper approximation is denoted as
B ¯ ( X ) = { x U | [ x ] P X } .
The lower approximation of X can be denoted as
B ¯ ( X ) = { x U | [ x ] P X }
where the objects in B ¯ ( X ) may belong to X, while the objects in B ¯ ( X ) must belong to X.
Definition 1. 
In a decision system  D S = U , C , D , V , f , if a C   a n d   x U that f x , a = , then we call the decision system an incomplete system (IDS). In an incomplete decision system, if P C D the tolerance relation is as follows:
T P = { x , y U × U | a P , f x , a = f y , a f x , a = f y , a = }
where * represents missing value. T P is symmetric and reflexive, but not transitive.
Definition 2. 
Given an incomplete decision system  I D S = ( U , C D , V , f ) , P C D , T ( P ) = a B T ( a ) , and T P ( o ) is the tolerance class determined by o with respect to P, which is defined as follows:
T P ( o ) = { y U | ( o , y ) T ( P ) } .
Let U / T ( P ) denote the family set of T P ( o ) , which is the classification included by P. If P B , then T B ( o ) T P ( o ) , which is the monotonicity of the tolerance class.

2.2. Basic Concept of Knowledge Granularity

Definition 3. 
Suppose  I D S = ( U , C D , V , f ) is an incomplete decision system, P C . T P ( o i ) is the tolerance class of object o i with respect to P. The knowledge granularity of P on U is defined as follows:
G K U ( P ) = 1 | U | 2 i = 1 | U | | T P ( o i ) |
where | U | represents the number of objects in dataset U. Since the reflexivity and symmetry of T P ( o i ) , there are lot of repeated calculations when calculating i = 1 | U | | T P ( o i ) | . In order to reduce the amount of calculation, we propose Definition 4 as follows:
Definition 4. 
Suppose  I D S = ( U , C D , V , f ) is an incomplete decision system, P C . C T P ( o i ) is the simplified tolerance relation of object o i with respect to P, which is defined as follows:
C T P ( o i ) = { o j | a P   , ( f ( a , o i ) = f ( a , o j )     (   f ( a , o i ) = * f ( a , o j ) = * ) ( i < j ) )   ( f ( a , o i ) = * f ( a , o j ) * ) } .
Definition 5. 
Given an incomplete decision system IDS= ( U , C D , V , f ) , P C and o U , the simplified tolerance class of object o with respect to attribute P is defined as follows:
C T P ( o ) = { y | ( o , y ) C T ( P ) } .
We can delete the symmetric element pair and reflexive element pair from Definition 3 and obtain Definition 5, so that Definition 5 does not have the characteristics of symmetry and reflexivity.
Definition 6. 
Given an incomplete decision system  I D S = ( U , C D , V , f ) , P C , o U , C T P ( o ) , is the simplified tolerance class of object o with respect to attribute P. The equal knowledge granularity of P on U is defined as follows:
E G K U ( P ) = 1 | U | 2 o U | C T P ( o ) | .
Theorem 1. 
Given an incomplete decision system  I D S = ( U , C D , V , f ) , P C . Let U / P = { X 1 , X 2 , , X l } , X i U / P , X i = n i where 1 i l . All objects in subdivision X l are missing a value on attribute P, X l = n . For the convenience of the following description, we mark X l as X . Objects with all non-missing values on attribute P are marked with X ¯ . E G K U ( P ) represents the equal knowledge granularity of P on U, we have:
E G K U ( P ) = 1 U 2 i = 1 l C n i 2 + X X ¯
where  C n 2 = n ( n 1 ) 2 .
Proof. 
Suppose that o X i and o X . According to Definition 4, we can obtain o X i C T P ( o )   = n i ( n i 1 ) 2 + n i n and o X C T P ( o ) = n ( n 1 ) 2 . Suppose o U , according to Definition 5, then we can obtain E G K U ( P ) = 1 | U | 2 o U C T P ( o ) = 1 U 2 i = 1 l 1 o X i C T P ( o ) +   o X C T P ( o ) = 1 U 2 · i = 1 l 1 n i ( n i 1 ) 2 + n i = 1 l 1 n i + n ( n 1 ) 2 = 1 U 2 i = 1 l n i ( n i 1 ) 2 + n U n . Since C n i 2 = n i ( n i 1 ) 2 , C n 2 = n ( n 1 ) 2 and X ¯ = U n , we can obtain E G K U ( P ) = = 1 U 2 i = 1 l C n i 2 + X X ¯ ) .  
Property 1. 
Given an incomplete decision system IDS=  ( U , C D , V , f ) , P C . If the knowledge granularity of P is G K U ( P ) on U and the equal knowledge granularity of P is E G K U ( P ) , then we have:
E G K U ( P ) = ( G K U ( P ) 1 U ) / 2 .
Proof. 
Let U / P = { X 1 , X 2 , , X l 1 , X } , X i = n i , X   = n where X i is the i-th subdivision of U / P and X stands for the subdivision of missing values on attribute P, X i stands for the number of object in subdivision X i . We can obtain U i = 1 l 1 X i = X . According to Definition 2, suppose that o X i and T P ( o ) = { x | x X i x X } , we can obtain T P ( o ) = n i + n . Since the T P ( o ) value of each object in X i is n i + n , we can obtain o X i T P ( o ) = n i ( n i + n ) . In the same way, we can obtain o X T P ( x ) = n U . According to Definition 3, we can obtain G K U ( P ) = 1 U 2 ( i = 1 l 1 o X i | T P ( o ) | + o X T P ( o ) ) = 1 U 2 · i = 1 l 1 n i ( n i + n ) + n U , then G K U ( P ) 1 U = 1 U 2 ( i = 1 l 1 ( n i 2 n i ) + i = 1 l 1 n i + n i = 1 l 1 n i + n U U ) = 1 U 2 ( i = 1 l 1 ( n i 2 n i ) n 2 n + 2 n U ) = 2 U 2 i = 1 l 1 ( n i 2 n i ) 2 + n 2 n 2 + n ( U n ) = 2 E G K U ( P ) . Let X l = X , we can obtain G K U ( P ) 1 U = 2 U 2 i = 1 l ( n i 2 n i ) 2 + n ( U n ) = 2 E G K U ( P ) .  □
Due to the time complexity of calculating T P ( o ) is U , we know that the time complexity of calculating G K U ( P ) is U 2 . However, the time complexity of C T P ( o ) is U / P , the time complexity of calculating E G K U ( P ) is U / P 2 , and U / P 2 U 2 . In addition, in the process of calculating E G K U ( P ) , the sub-division with a cardinality of 1 is constantly pruned, which further speeds up the calculation. Therefore, the time of calculating E G K U ( P ) is less than G K U ( P ) for the same data set.
Example 1. 
Example of computing equivalent knowledgegranularity. Let I D S = ( U , C D , V , f ) , U = { o 1 , o 2 , o 3 , o 4 , o 5 , o 6 , o 7 , o 8 , o 9 } , C = { a , b , c , e } , D = { d } . The detailed data are shown in Table 1. Let P = { a , b } , we can obtain f ( o 6 , P ) = * , f ( o 9 , P ) = * . We use the following two methods to calculate E G K U ( P ) .
(1)
According to Definition 5, we obtain that C T P ( o 1 ) = { o 2 , o 6 , o 9 } , C T P ( o 2 ) = { o 6 , o 9 } , C T P ( o 3 ) = { o 4 , o 5 , o 6 , o 9 } , C T P ( o 4 ) = { o 5 , o 6 , o 9 } , C T P ( o 5 ) = { o 6 , o 9 } , C T P ( o 6 ) = { o 9 } , C T P ( o 7 ) = { o 6 , o 8 , o 9 } , C T P ( o 8 ) = { o 6 , o 9 } , C T P ( o 9 ) = { } . According to Definition 6, we can obtain E G K U ( P ) = 1 U 2 · i = 1 9 C T P ( o i ) 1 9 2 ( 3 + 2 + 4 + 3 + 2 + 3 + 2 + 1 + 0 ) = 20 81 .
(2)
Since U / P = { { o 1 , o 2 } , { o 3 , o 4 , o 5 } , { o 7 , o 8 } , { o 6 , o 9 } } , let X 1 = { o 1 , o 2 } , X 2 = { o 3 , o 4 , o 5 } , X 3 = { o 7 , o 8 } , X = { o 6 , o 9 } , then X = 2 , X ¯ = U 2 = 7 . According to Theorem 1, we can obtain E G K U ( P ) = 1 9 2 · [ C 2 2 + C 3 2 + C 2 2 + C 2 2 + 2 · ( 9 2 ) ] = 20 81 .
Although the above two methods achieve the same results, the calculation time is different. Since method 1 needs to scan the dataset multiple times, it consumes more time. However, method 2 only needs to scan the dataset one time and obtains each subdivision of U/P. According to the number of objects in each subdivision, we can acquire the combination value quickly, and then the value of equivalent knowledge granularity is calculated.

3. The Mechanism of Knowledge Granularity Attribute Reduction in the Background of Clustering

Most traditional attribute reduction models use equivalence class relation to compute the importance of conditional attributes. Although these methods can effectively deal with complete decision-making information systems, they cannot obtain correct reduction rules in incomplete ones. In order to deal with the loss of information effectively, this paper focuses on the reduction in incomplete decision systems.
The traditional reduction model does not consider the impact on the classification of the dataset when deleting redundant attributes. If there are inconsistent objects in the dataset, the classification accuracy of the reduced set will be affected. In order to improve the data quality, this paper uses the idea of clustering. Clustering is to divide all objects in the dataset into different clusters according to a certain standard when the target category is unknown. Objects within a cluster are as similar as possible, and objects between clusters are as dissimilar as possible. Classification is to classify all objects in the dataset according to a certain nature and level when the object category is known. Good clustering results can be used as a reference standard for accurate classification. The desired results of clustering involve the objects of the same class being gathered in intra-clustering, otherwise they will be gathered in different inter-clustering. This paper studies the labeled data objects decision information system. Therefore, we use the results of the classification to guide the process of clustering the data objects. When the data objects are clustered, they follow the principle that the objects of intra-clustering are as close as possible and the objects of inter-clustering are as far away as possible. Next, we discuss how to measure the distance of intra-clustering and inter-clustering objects.

3.1. The Intra-Cluster Similarity for Incomplete Systems

Generally, there are two approaches for clustering calculations: distance and similarity. The closer the distance between two different objects is, the weaker their ability to distinguish is. On the contrary, the farther the distance is, the stronger the ability to distinguish is. In this paper, the similarity method is used to measure the distinguishing ability of objects. Since the knowledge granularity can measure the similarity between objects, the coarser the knowledge granularity is, the stronger the distinguishing ability is. The better the knowledge granularity is, the weaker the distinguishing ability is. Next, we discuss how to use knowledge granularity information to measure the similarity of objects in an incomplete system.
Definition 7 
(The similarity of intra-cluster objects). Given an incomplete decision system  I D S = ( U , C D , V , f ) , U / D = { D 1 , D 2 , , D n } . For the sake of convenience, let U / D = π D , D i π D , P C . Suppose the equivalence division relationship of D i under attribute set P is P = { X 1 , X 2 , , X m } ; then, the similarity of objects in the cluster of D i about attribute P is defined as follows (where o D i ):
S I n t r a D i ( P ) = E G K D i ( P ) = 1 | D i | 2 o D i C T P ( o ) .
Definition 8 
(The average similarity of intra-cluster objects). Given an incomplete decision system  I D S = ( U , C D , V , f ) , U / D = { D 1 , D 2 , , D n } . P C , o D i . The knowledge granularity similarity of intra-clustering objects for subdivision D i with respect to attribute P is S I n t r a D i ( P ) , then the average intra-clustering similarity is defined as follows:
A S I n t r a π D ( P ) = 1 n i = 1 n S I n t r a D i ( P ) .
The desired effect of clustering is that the similarity of intra-clustering is high, and the similarity of inter-clustering is low.
Property 2. 
Given an incomplete system  I D S = ( U , C D , V , f ) , U / D = π D , D i π D , P , Q C . If P Q , and we have
A S I n t r a π D ( P ) A S I n t r a π D ( Q ) .
Proof. 
Let D i / P = { X 1 , X 2 , , X k X k + 1 , X k + 2 , , X n , X Y } , where X , Y represents the object sets with missing values on attribute set P. Since P Q , we can obtain D i / Q D i / P . This to say, each subdivision of D i / Q is a subset of some subdivision of D i / P . Let D i / Q = { X 1 , X 2 , , X k , X k + 1 , X k + 2 , , X n , Y , X } . According to Theorem 1 and Definition 7, we can obtain: S I n t r a D i ( P ) = 2 D i 2 ( j = 1 k 1 C X j 2 + C X k + X k + 1 2 + j = k + 2 n C X j 2 + C X + Y 2 + ( X + Y ) ( D i X Y ) )   = 2 D i 2 ( j = 1 k 1 C X j 2 + C X k 2 + C X k + 1 2 + X k X k + 1 + j = k + 2 n C X j 2 + C X 2 + C Y 2 + X Y + ( X + Y ) ( D i X Y ) ) = 2 D i 2 ( j = 1 n C X j 2 + X k X k + 1 + C X 2 + C Y 2 + X Y + ( X + Y ) ( D i X Y ) ) . Since S I n t r a D i ( Q ) = 2 D i 2 ( j = 1 n C X j 2 + C X 2 + C Y 2 + X ( D i X ) ) , then 2 D i 2 ( X k X k + 1 + Y D i X Y Y 2 ) = 2 D i 2 ( X k X k + 1 + Y ( D i X Y ) ) . Since D i > = X + Y , X k X k + 1 0 ; therefore, the results of S I n t r a D i ( P ) S I n t r a D i ( Q ) and i = 1 n S I n t r a D i ( P ) i = 1 n S I n t r a D i ( Q ) are obtained. Above all, we can obtain A S I n t r a π D ( P ) A S I n t r a π D ( Q ) . □
According to Property 2, we conclude that the intra-cluster similarity is monotonic when the conditional attribute set changes.
Example 2. 
In Table 1, let  P = { a , b } ; we have D 1 = { o 1 , o 3 } , D 2 = { o 2 , o 4 , o 5 , o 6 } , D 3 = { o 7 , o 8 , o 9 } , S I n t r a D 1 ( P ) = 0 , S I n t r a D 2 ( P ) = 0 + C 2 2 + 1 × 3 4 2 = 1 4 , S I n t r a D 3 ( P ) = C 2 2 + 1 × 2 3 2 = 1 3 , A S I n t r a π D ( P ) = 1 3 i = 1 3 S I n t r a D i ( P ) = 7 36 .

3.2. The Inter-Cluster Similarity for Incomplete Systems

Definition 9 
(The inter-cluster similarity for incomplete systems). Given  I D S = ( U , C D , V , f ) is an incomplete decision system, let π D = { D 1 , D 2 , , D n } . Suppose P C , D i , D j U / D , then the inter-cluster similarity of D i and D j with respect to attribute set P for incomplete systems is defined as the following:
S I n t e r D i , D j ( P ) = 1 ( D i + D j ) 2 o D i D j C T P ( o ) o D i C T P ( o ) o D j C T P ( o ) .
Assuming that D i and D j are objects of two different clusters, the inter-cluster similarity between D i and D j is calculated in two steps. The first step is to calculate the similarity after the two clusters are merged, and the second step is to remove the similarity information of the same cluster. The rest is the similarity information between objects in different clusters.
Property 3. 
Given an incomplete decision system  I D S = ( U , C D , V , f ) , π D = { D 1 , D 2 , , D n }   P C . Let D i / P = { X 1 , X 2 , , X k , X k + 1 , , X m , X } , X ¯ = D i X , D j / P = { Y 1 , Y 2 , , Y k , Y k + 1 , , Y n , Y } , Y ¯ = D j Y , D i D j / P = { X 1 Y 1 , X 2 Y 2 , , X k Y k , X Y , X k + 1 , , X m , Y k + 1 , , Y n } , where o X Y , f ( o , P ) = * , ‘*’ is the flag of missing value. We have:
S I n t e r D i , D j ( P ) = 1 ( D i + D j ) 2 l = 1 k X l Y l + + X Y + X Y ¯ + X ¯ Y .
Proof. 
According to Definition 5 and Theorem 1, we can obtain o D i D j C T P ( o ) = l = 1 k C X l + Y l 2 + C X + Y 2 + l = k + 1 m C X i 2 + l = k + 1 n C Y i 2 + ( X + Y ) ( Y ¯ + X ¯ ) , o D i C T P ( o ) = l = 1 m C X l 2 + C X 2 + X X ¯ , o D j C T P ( o ) = l = 1 n C Y l 2 + C Y 2 + Y Y ¯ . Since l = 1 k C X l + Y l 2 = l = 1 k X l 2 X l + Y l 2 Y l 2 + X l Y l = l = 1 k C X l 2 + l = 1 k C Y l 2 + l = 1 k X l Y l , according to Definition 9, we can conclude that ( D i + D j ) 2 S I n t e r D i , D j ( P ) = o D i D j C T P ( o ) o D i C T P ( o ) o D j C T P ( o ) = l = 1 k X l Y l + X Y + X Y ¯ + X ¯ Y . Then, we have:
S I n t e r D i , D j ( P ) = 1 ( D i + D j ) 2 l = 1 k X l Y l + X Y + X Y ¯ + X ¯ Y .
Property 4. 
Given an incomplete decision system  I D S = ( U , C D , V , f ) , π D = { D 1 , D 2 , , D n } , P Q C , D i , D j U / D , then we have S I n t e r D i , D j ( P ) S I n t e r D i , D j ( Q ) .
Proof. 
Let D i / Q = { X 1 , X 2 , , X k , X k + 1 , , X m , X , X X } , X X ¯ = D i X + X , D j / Q = { Y 1 , Y 2 , , Y k , Y k + 1 , , Y n , Y , Y Y } , Y Y ¯ = D j Y + Y . Since P Q C , then D i / Q is a refinement of D i / P , D j / Q is a refinement of D j / P . Let D i / P = { X 1 , X 2 , , X k X k + 1 , , X m , X } , X ¯ = D i X , D j / P = { Y 1 , Y 2 , , Y k Y k + 1 , , Y n , Y } , Y ¯ = D j Y . Suppose D i D j / Q   = { X 1 Y 1 , X 2 Y 2 , , X k Y k , X k + 1 Y k + 1 , X Δ Y Δ , ( X X Δ ) ( Y Y Δ ) , X k + 2 , , X m , Y k + 2 , , Y n } , D i D j / P = { X 1 Y 1 , X 2 Y 2 , , X k Y k , X k + 1 Y k + 1 , X Y , X k + 2 , , X m , Y k + 2 , , Y n } , then S I n t e r D i , D j ( P ) = l = 1 k 1 X l Y l + ( X k + X k + 1 ) ( Y k + Y k + 1 ) + X Y + X Y ¯ + X ¯ Y , S I n t e r D i , D j ( Q ) = l = 1 k 1 X l Y l + X k Y k + X k + 1 Y k + 1 + X Δ Y Δ + X X Δ Y Y Δ + X X Δ Y Y ¯ Δ + X X Δ ¯ Y Y Δ . S I n t e r D i , D j ( P ) S I n t e r D i , D j ( Q ) = X k Y k + 1 + X k + 1 Y k + X Y + X Y ¯ + X ¯ Y X Δ Y Δ X X Δ Y Y Δ X X Δ Y Y ¯ Δ + X X Δ ¯ Y Y Δ . Since X ¯ = D i X , Y ¯ = D j Y , X X Δ ¯ = D i X + X Δ , Y Y Δ ¯ = D j Y + Y Δ , X k Y k + 1 + X k + 1 Y k 0 , then S I n t e r D i , D j ( P )   S I n t e r D i , D j ( Q ) X Y + X Y ¯ + X ¯ Y X Δ Y Δ X X Δ Y Y Δ X X Δ Y Y ¯ Δ + X X Δ ¯ Y Y Δ = X · D j + Y Δ D i X Y Δ X Δ Y = X Δ ( D j Y ) + Y Δ · ( D i X ) . Since D j Y , D i X , then we have S I n t e r D i , D j ( P ) S I n t e r D i , D j ( Q ) 0 . □
From Property 4, it can be concluded that the similarity of inter-clusters has mono- tonicity with the change in conditional attributes.
Definition 10 
(The average similarity of inter-cluster objects). Given an incomplete decision system  I D S = ( U , C D , V , f ) , π D = { D 1 , D 2 , , D n } , P C . In n different clusters, the inter-cluster similarity is calculated between every two clusters and the time of comparisons is 1 2 n ( n 1 ) , then the average knowledge granularity similarity of inter-cluster in dataset U respect of attribute set P is defined as the following:
A S I n t e r π D ( P ) = 2 n ( n 1 ) i = 1 n 1 j = i + 1 n S I n t e r D i , D j ( P ) .
Example 3. 
In Table 1, with the same conditions as Example 2, let P = { a , b } , we have D 1 = { o 1 , o 3 } , D 2 = { o 2 , o 4 , o 5 , o 6 } , D 3 = { o 7 , o 8 , o 9 } , S I n t e r D 1 , D 2 ( P ) = 1 × 1 + 1 × 2 + 2 × 1 ( 2 + 4 ) 2 = 5 36 , S I n t e r D 1 , D 3 ( P ) = 0 + 0 + 1 × 2 ( 2 + 3 ) 2 = 2 25 , S I n t e r D 2 , D 3 ( P ) = 0 + 1 × 2 + 3 × 1 ( 4 + 3 ) 2 = 5 49 , A S I n t e r π D ( P ) = 1 3 i = 1 2 j = i + 1 3 S I n t e r D i , D j ( P ) = 14153 132300 .

4. Attribute Reduction of Knowledge Granularity for Incomplete Systems

Traditional attribute reduction methods are mostly aimed at datasets with no missing data. Various datasets in reality are often incomplete due to various subjective or objective factors. Therefore, we researched information systems with missing data and propose corresponding algorithms to improve the data quality of incomplete system reduction sets.
This paper discusses how to design a reduction method with the idea of clustering. For an ideal clustering effect, the objects of inter-cluster should be far away, and the objects of intra-cluster should be close together. Here, similarity is used to measure the distance between two different objects. The higher the similarity is, the closer the objects are. Conversely, the lower the similarity is, the farther the objects are. Based on the above analysis, we designed a formula to measure the importance of attributes as the following:
S I M R = S A I n t r a + λ · ( 1 A S I n t e r )
where λ is the weight, and 1 A S I n t r a is the dissimilarity of intra-cluster objects. We can set the importance of the intra-cluster similarity and inter-cluster similarity by using the size of the parameter λ . This method requires the adjustment of the parameters continuously, which consumes a lot of time. To this end, we first normalize A S I n t r a and A S I n t e r . Then, the two similarity calculations can be measured within a unified range, avoiding the parameter adjustment process.

4.1. Normalization of Inter-Cluster Similarity and Intra-Cluster Similarity

Given an incomplete decision system I D S = ( U , C D , V , f ) , π D = { D 1 , D 2 , , D n } , P C , D i , D j U / D . Since the number of elements in each sub-division may be different, then the value range of its equivalent knowledge granularity may also be different. To calculate the average similarity of all subdivisions, they must be calculated in the same domain. For the sake of generality, we normalize the inter-cluster similarity and intra-cluster similarity. According to Definition 7 and Theorem 1, we can obtain S I n t r a D i ( P ) = E G K D i ( P ) . When all data objects in the subdivision D i are indistinguishable with respect to the attribute set P, E G K D i ( P ) = 1 2 ( 1 1 D i ) takes the maximum value.
When all data objects in D i can be distinguished from each other, E G K D i ( P ) = 0 takes the minimum value. So, the result of E G K D i ( P ) 0 , D i 1 2 D i is obtained. If the value of E G K D i ( P ) is mapped to the range [0,1], the correction formula of S I n t r a D i ( P ) is defined as
S I n t r a D i ( P ) = E G K D i ( P ) / D i 1 2 D i = E G K D i ( P ) 2 D i D i 1 .
The average similarity of intra-cluster objects is corrected as follows:
A S I n t r a π D ( P ) = 1 n i = 1 n S I n t r a D i ( P )
D i and D j are two object sets of different clusters. Suppose D i / P = { X 1 , X 2 , , X n , X } , D j / P = { Y 1 , Y 2 , , Y m , Y } . If f ( X l , P ) = f ( Y l , P ) ( 1 l k ) , then D i D j / P = { X 1 Y 1 , X 2 Y 2 ,   , X k Y k , , X k + 1 , , X n , Y k + 1 , , Y m , X , Y } . According to Property 3, the similarity of inter-cluster respect to D i and D j is denoted as
S I n t e r D i , D j ( P ) = 1 ( D i + D j ) 2 l = 1 k X l Y l + X Y ¯ + X ¯ Y .
When i = 1 k X i Y i = , X = , Y = } , S I n t e r D i , D j ( P ) = 0 , takes the minimum value. When all data objects in D i and D j are indistinguishable, S I n t e r D i , D j ( P ) = D i D j ( D i + D j ) 2 takes the maximum value. Then, we can obtain S I n t e r D i , D j ( P ) 0 , D i D j ( D i + D j ) 2 . The normalized formula of S I n t e r D i , D j ( P ) is as follows:
S I n t e r D i , D j ( P ) = S I n t e r D i , D j ( P ) ( D i + D j ) 2 D i D j .
The definition of the average similarity of inter-cluster objects is revised as follows:
A S I n t e r π D ( P ) = n ( n 1 ) 2 i = 1 n 1 j = i + 1 n S I n t e r D i , D j ( P ) .
After the similarities of inter-cluster and intra-cluster objects are normalized, S I M R = A S I n t e r + λ · ( 1 A S I n t r a ) is revised as follows:
S I M R = A I n t r a π D ( P ) + 1 A I n t e r π D ( P ) .
Since A S I n t e r π D ( P ) represents the similarity of intra-cluster objects, then the dissimilarity is 1− A S I n t e r π D ( P ) . If you use the formula of S I M R to measure the effect of clustering, the larger the value of S I M R is, the better the effect is.

4.2. The Knowledge Granularity Attribute Reduction Algorithm for Incomplete Systems (IKAR)

In Section 4.1, we discussed the similarities of inter-cluster and intra-cluster objects from the perspective of clustering, which provided a clear goal for the next step of attribute selection.
Definition 11 
(Equal knowledge granularity attribute reduction). Given an incomplete decision system  I D S = ( U , C D , V , f ) , an attribute subset R C is an equal know-ledge granularity attribute reduction if and only if:
( 1 ) R = m i n P C { S I M P } ( 2 ) R R , S I M R > S I M R
In Definition 11, condition (1) is the jointly sufficient condition that guarantees that the equal knowledge granularity value induced from the reduction is minimal, and condition (2) is the individual necessary condition that guarantees the reduction is minimal.
According to Definition 11, we can find a smaller reduction set with an ideal classification effect. Property 2 proves that when the attribute decreases, the similarity of intra-cluster increases monotonically. Property 4 proves that the similarity of inter-cluster also increases when decreasing the attribute, so that the dissimilarity will decrease. Obviously, the formula S I M R cannot determine its monotonicity. To find the R when the S I M R is the largest in the conditional attribute, which is a combinatorial optimization problem, and trying the methods one by one is not the best way to solve the problem. So, we use a heuristic method to find the smallest reduced set. For any attribute a C , its inner significance is defined as follows:
S i g C a = S I M C S I M C a
The bigger the value of S i g C a is, the more important the attribute is.
In order to obtain the optimal reduction set quickly, we adopt the deletion strategy. Firstly, the importance S i g C a of attribute a is defined by the formula of S I M ; then, sort the different S i g C a . Secondly, let R = C. The value of S I M C is used as the initial condition, which ensures that the clustering effect after reduction is better than the raw dataset. Then, remove the unimportant attribute a from the remaining attributes C R and calculate the value of S I M R a . If S I M R a S I M R , delete attribute a and continue; otherwise, the algorithm terminates. The details of the IKAR algorithm are shown in Figure 1.
Example 4. 
In Table 1, since D 1 = { o 1 , o 3 } , D 2 = { o 2 , o 4 , o 5 , o 6 } , D 3 = { o 7 , o 8 , o 9 } . Let R = C , according to the definition of S I M R , we can obtain S I M R = 72 75 , S I M C a = 71 72 , S I M C b = 100 72 , S I M C c = 60 72 and S I M C e = 80 72 , then S i g C a = 4 72 , S i g C b = 25 72 , S i g C c = 15 72 , S i g C e = 5 72 . Since S I M C b = 100 72 > S I M R , we delete the attribute b from R and let R = R b , S I M R = 100 72 . In the same way, we obtain that S I M R a = 94 72 , S I M R c = 92 72 , S I M R e = 108 72 , S i g R a = 6 72 , S i g R c = 8 72 and S i g R e = 8 72 . Since S I M R e = 108 72 > S I M R , we delete the attribute e from R and let R = R e . We calculate the value of S I M R a , S I M R c , obtain the results of S I M R a = 82 72 and S I M R c = 100 72 . Now, we have S i g R a = 18 72 and S i g R c = 8 72 . If attribute c is deleted S I M R c = 100 72 < S I M R , and the algorithm is terminated. We have R = { a , c } .

4.3. Time Complexity Analysis

The time complexity of Step 1 is O ( 1 ) . In Step 2, we calculate the value of S I M C which include A S I n t r a π D ( C ) and A S I n t e r π D ( C ) . The computational time complexity of intra-cluster S I n t r a D i ( C ) similarity is O ( C D i ) , then the time complexity of calculating A S I n t r a π D ( C ) is O ( U C ) . Since the time complexity of calculating the similarity S I n t e r D i , D j ( C ) about inter-cluster D i and D j is O ( ( D i + D j ) C ) , then time complexity of A S I n t e r π D ( C ) is O ( i = 1 U / D 1 j = i + 1 U / D ( D i + D j ) C ) = O ( ( U / D 1 ) U C ) . We can obtain the time complexity of Step 2 is O ( U / D U C ) . In Step 3, the consume time is O ( U / D U C ) . Since Step 4 utilizes the results of Step 3, the time complexity is O ( 1 ) . Step 5 is to sort the importance S i g R a of each attribute, and the time complexity is O ( 1 2 C 2 ) . Since the time complexity of calculating S i g R a is O ( U / D U R ) , then the time complexity of Step 5 is O ( U / D U R ) + O ( 1 2 C 2 ) . In Step 6, the time complexity of deleting a redundant attribute is O ( U / D U R ) , and Step 6 needs to be executed C R times, R C , then the time complexity of Step 6 is O ( U / D U ( C 2 R 2 ) . In summary, the time complexity of IKAR is O ( U / D U ( C 2 R 2 ) + 1 2 C 2 ) .

5. Experiments Results Analysis

In order to evaluate the feasibility and effectiveness of the proposed algorithm in this paper, the complete dataset was preprocessed to obtain incomplete information systems, and many different attribute reduction algorithms were used for reduction. The reduction set obtained in the previous stage was classified and analyzed by multiple classifiers in the Weka Tool. It was compared with three other existing algorithms in terms of reduction set size, running time, accuracy, and algorithm stability. The specific experimental framework is shown in Figure 2.
All of the datasets are displayed in Table 2. The twelve datasets selected were downloaded from UCI. In Table 2, U , C and D represent the number of objects, conditional attributes, and the categories of the decision attribute, respectively. In order to generate an incomplete information system, we deleted 12% attribute values from the raw datasets, and the missing values were randomly and uniformly distributed. The missing value of each attribute was removed with equal probability, which eliminated the impact of the later reduction set on the classification accuracy due to different attribute selection. During data preprocessing prior to the experiment, we kept the discrete data unchanged and discretized the continuous value data. The two classifiers, RF (random forest) and SMO of the Weka v3.8 software platform, were used to demonstrate the classification effect of the reduction set. In subsequent experiments, the reduced sets of each dataset were analyzed for accuracy using the cross method. All the objects in the reduced set were randomly divided into 10 equal parts, one of which was used as the test set, and the remaining 9 were used as the training set. During the classification and analysis process, the default settings of the Weka tool were used for all parameters. In this way, each reduction set was repeated 10 times for the cross experiment. Finally, the size of the reduction set, running time, and classification accuracy obtained by the 10 experiments were averaged. We executed all experiments on a PC with Windows 10, Intel(R) Core(TM) i7-10710U CPU @ 1.10 GHz, 1.61 GHz and 8 GB memory. Algorithms were coded in python and the software that was used is PyCharm Community Edition 2020.2.3 x64.

5.1. Reduction Set Size and Running Time Analysis

This section mainly verifies the feasibility of IKAR for dataset reduction from the perspective of reduction size and calculation time. Here, we have selected three other representative incomplete system attribute reduction approaches. For the convenience of the following description, these three methods are referred to as NGLE [44], IEAR [45], and PARA [46], respectively. NGLE is the neighborhood multi-granulation rough sets-based attribute reduction using Lebesque and entropy, and the method considers both algebraic view and information view. IEAR is the information entropy attribute reduction for incomplete set-valued data using the similarity degree function and proposed λ -information entropy. PARA is the positive region attribute reduction using indiscernibility and discernibility relation.
The attribute reduction size of four algorithms is shown in Table 3. Table 4 shows the time consumed by the reduction process of the four algorithms in seconds. Ave represents the average value, Best in Table 3 represents the number of minimum reduction sets obtained, and Best in Table 4 stands for the number of times which the running time was the shortest. From Table 3, it can be seen that the average reduction set size of the IKAR algorithm is 11.833, and the average reduction set size obtained by the NGLE, IEAR, and PARA algorithms are 11.917, 11.00, and 12.083, respectively. IKAR obtained the minimum reduction set on the Ches, Spli, Mush, and Shut datasets, and the reduction effect was slightly better than the NGLE and PARA, but not as ideal as the IEAR algorithm. In the 12 datasets in Table 2, the IEAR algorithm obtains the minimum reduction set in 11 datasets.
From the consumption time of Table 4, we can find that the reduction advantage obtained by the IKAR algorithm is not obvious, but that the IKAR algorithm is obviously better than the other three algorithms. When calculating the reduction set of 12 datasets, the average time required by IKAR is 15.20 s, whilst the average times consumed by the other three algorithms are (155.40, 2003.56, and 260.38) seconds, respectively. From the experimental results in Table 4, it can also be found that the IKAR algorithm only needs 25.30 s to reduce the Shut dataset, and the running time of the other three algorithms of NGLE, IEAR, and PARA is (886.08, 3974.02, and 133.25) seconds. When reducing the Mush dataset, IKAR takes 4.45 s, and the other three algorithms take (411.09, 4544.04, and 827.16) seconds. Of course, the IKAR algorithm also has shortcomings, and when the number of sample categories is higher, the algorithm is more time-consuming. For example, if there are 10 different categories of samples in the Hand dataset, the calculation time of the IKAR algorithm takes 92.93 s. At this time, the time consumption of the NGLE and PARA algorithms is (81.61 and 91.28) seconds, respectively. It takes less time than that of the IKAR algorithm.
From the above analysis, we can obtain that the IKAR algorithm can effectively reduce the dataset, it is obviously better than similar algorithms in terms of reduction speed.

5.2. Changes of the Classification Accuracy When Missing Value

It is not enough to evaluate the overall performance of a reduction algorithm only from the size and running time of the reduction set. Here, we further analyze the performance of the above four algorithms from the perspective of the classification accuracy of the reduction set. In order to find out the influence of missing values on the IKAR algorithm, we selected four datasets in Table 2, including Heas, Prom, Hepa, and Stai. During the experiment, we divided the missing values into 5 different levels, 2%, 4%, 6%, 8%, and 10%. First, 2% of the data objects in the original dataset were randomly selected, and random deletion is conducted in these data objects of attribute values for different attributes to generate an incomplete dataset. In order to reduce the bias of attribute selection that may be caused by random deletion, we used the 2% data just obtained as a basis, then selected attributes with the same probability, and then deleted the 2% data, which generated an incomplete dataset with a missing value of 4% and so on, generating a dataset with a missing value of 6%, 8%, and 10%, respectively. After the incomplete dataset was ready, we used IKAR, NGLE, IEAR, and PARA algorithms for reduction and used SMO and RF classifiers to analyze the classification accuracy. The specific results are shown in Figure 3 and Figure 4. The horizontal axis in Figure 3 and Figure 4 represent the proportion of deletion data objects in the dataset, and the vertical axis represents the classification accuracy.
Figure 3 and Figure 4 show the change trend diagrams obtained by using the SMO and RF classifiers to analyze the accuracy of the reduced dataset. From the results of Figure 3 and Figure 4, it can be seen that when increasing the missing data, the classification accuracy of the above four algorithms has a downward trend. For example, under the SMO classifier, when the proportion of missing values in the dataset Heas changes from 2% to 4%, the accuracy changes in IKAR and other NGLE, IEAR and PARA algorithms are (84.01→83.69), (80.32→79.69), (77.01→76.86), and (79.24→78.92). Under the RF classifier, when the proportion of missing values in the dataset Heas changes from 2% to 4%, in the IKAR and other NGLE, IEAR, and PARA algorithms, the accuracy changes are (81.43→80.94), (77.85→76.28), (74.79→73.33), and (76.24→74.86), respectively. The main reason for this phenomenon is that as the proportion of missing data increases, more and more data objects cannot be distinguished, resulting in a decrease in classification accuracy. From the trend diagrams in Figure 3 and Figure 4, it can be seen that the IKAR algorithm changes smoothly. The other three algorithms have a greater impact on the classification accuracy of the reduced set as the proportion of missing data increases. For example, under the RF classifier, when the missing proportion of the Hepa dataset changes from 2% to 10%, the accuracy of IKAR changes to 2.09, while the accuracy changes of the NGLE, IEAR, and PARA algorithms are 3.2, 3.19, and 4.49, respectively. Under the SMO classifier, on the classification accuracy of the Heas dataset, the IKAR algorithm changes value is 1.12, and for the other three algorithms is 3.37, 2.28, and 2.93, respectively.

5.3. Classification Accuracy Analysis

The previous experiments have compared the changes in the classification accuracy when missing value exist. With the change in the proportion of missing values, the accuracy of the IKAR algorithm can not only change smoothly but can also obtain a better classification effect on multiple datasets. We used the SMO and RF classifiers to analyze the accuracy of the previous reduction set, and the detailed content is shown in Figure 5 and Figure 6, respectively. In Figure 5 and Figure 6, missing values represent the incomplete raw dataset and Ave denotes the average accuracy for the different datasets. Under the classifier SMO, the IKAR algorithm obtained an average accuracy of 85.31%, and the average accuracy obtained by the other three algorithms NGLE, IEAR, PARA, and raw datasets were only 78.72, 78.78, 78.81, 84.49(%), respectively. Among the 12 datasets in Table 2, the IKAR algorithm was the highest 11 times. The classification accuracy of IKAR is higher than that of the raw set.
For example, the classification accuracy of IKAR is (94.76, 99.68, 93.86)% in the datasets of Hand, Shut, and Spli. The classification accuracy of IEAR is (81.59, 85.93, 81.53)% and the average accuracy is 10 percentage points lower than IKAR. Similarly, on the RF classifier, the average classification accuracy of the IKAR (86.24%) is higher than the raw dataset (85.09%) and those obtained by the other three algorithms obtained (80.79, 79.77, 81.14)%, respectively.
From the results shown in Figure 5 and Figure 6, we can obtain three conclusions as follows: (1) The classification accuracies obtained by the above four attribute algorithms are closer to those of the raw dataset, which indicates that all four algorithms are able to effectively reduce the incomplete dataset. (2) The algorithm has better overall performance. The main reason is that the IKAR algorithm considers both the correlation between attributes and the influence of attributes on the classification two factors, so it obtains a more ideal classification accuracy. (3) The classification accuracy of the IKAR algorithm is higher than the original dataset, and the main reason for this is that the IKAR algorithm effectively eliminates redundant attributes and reduces the influence of noisy data on classification.

5.4. Lgorithm Stability Analysis

To indicate the statistical significance of classification results, the non-parametric Friedman test and Bonferroni–Dunn test methods were used to the analyze classification accuracy of each classifier with different methods in Section 5.2, where the Friedman test is a statistical test that uses the ranking of each method on each dataset. The Friedman statistic is described as follows:
χ F 2 = 12 N t ( t + 1 ) i = 1 s R i 2 3 t ( t + 1 )
F F = ( N 1 ) χ F 2 N ( t 1 ) χ F 2
where N is the number of datasets, and t is the categories of algorithms. R i represents the average rank of the classification effect ranking of the i-th algorithm on all datasets, and the statistics F F obey the Fisher distribution with t − 1 and (t − 1)(N − 1) degrees of freedom. If the value of F F is bigger than F α ( t 1 , N 1 ) , then the original hypothesis does not hold.
The F F test value can judge where these algorithms are different, but it cannot indicate the superiority of the algorithm. In order to explore which algorithm is better, we used the Bonferroni–Dunn test to calculate the critical value range of the average sequence value difference, which is defined as follows:
C D α = q α t ( t + 1 ) 6 N .
If the difference between the average ranking values of the two algorithms exceeds the critical region C D α , the hypothesis that ‘the performance of the two algorithms is the same’ will be rejected with the corresponding confidence. Otherwise, the two algorithms perform differently, and the algorithm with the higher average rank is statistically better than the algorithm with the lower average rank. Generally, we set α = 0.05.
In order to compare the stability of the IKAR algorithm in this paper, we chose three other similar algorithms, NGLE, LEAR, and PARA, to reduce the datasets in Table 2. Then, the previous reduction results were analyzed by the classifiers SMO and RF using the Weka tool. The classification accuracy was detected by the Friedman test and Bonferroni–Dunn test. When t = 4 and N = 12, then χ F 2 = 23.4 and F F = 20.429. Under the classifier SMO, the classification accuracy of the four algorithm reduction sets is sorted, and the number 1 represents the most ideal. The details of the sorting results are shown in Table 5. The average ranking of IKAR, NGLE, IEAR, and PARA algorithms are 1.08, 2.75, 3.58, and 2.58 in turn.
Since the critical value of F 0.05 ( 3 , 33 ) is 2.892 and F F 2.892 , we can reject the original hypothesis at α = 0.05 under the Friedman test. So, there are statistical differences in classification accuracy among the above four algorithms. Next, q 0.05 = 2.569 and C D 0.05 = 1.354, and the results of Bonferroni–Dunn test for these four algorithms at α = 0.05 is shown in Figure 7. The accuracy of the algorithm on the left side of the coordinate axis is relatively high in Figure 7. From Figure 7, we know that the average accuracy ranking of IKAR is s1 = 1.08. Among the other three algorithms, the better-ranking algorithm, PARA, has an average ranking value of s2 = 2.58. Since s 1 s 2 = 1.5 and s 1 s 2 C D 0.05 , we can find that the IKAR algorithm is significantly superior to the PARA. In the same way, the IKAR is better on accuracy than NGLE and LEAR. However, there is no obvious difference in the ranking of the NGLE, LEAR, and PARA algorithms.
For the same reason, under the RF classifier, the average ranking values of the classification accuracy of the above four algorithms are 1.00, 2.83, 3.67, and 2.50 as shown in Table 6. Since χ F 2 = 26.8 and F F = 32.043, then F F F 0.05 . We can see that the classification accuracy of these four algorithms is significantly different in the statistical sense under the RF classifier. From Figure 8, we can ascertain that the IKAR algorithm is significantly different from the NGLE classification, while the ranking of the NGLE, LEAR, and PARA algorithms have no obvious differences among each other.
In Table 5 and Table 6, the STD represents the standard deviation of the classification accuracy of different algorithms’ reduction datasets. From Table 5 and Table 6, we know that the STD of the IKAR algorithm is smallest among the four algorithms. The STD of the classification accuracy of algorithm IKAR is less than 1 for both SMO and RF classifiers, while the STD values of the other algorithms are bigger than 1, and some are even greater than 3. For example, the STD of IKAR in the Heas dataset is ±0.015 under the SMO, but the STD values of NGLE, LEAR, and PARA are ±1.455, ±2.318, and ±1.521, respectively. On RF classifiers, the IKAR algorithm has the same stability in classification accuracy. In the dataset Spli, the STD of IKAR’s classification accuracy is ±0.248; meanwhile, the STD values of the other three algorithms are ±1.994, ±3.565 and ±3.558, respectively.
Therefore, all the test results demonstrate that there is no consistent evidence to denote statistical differences between any two of the four approaches under the SMO and RF classifier. In general, the IKAR model is better than the other models in stability.

6. Conclusions

In the face of incomplete systems, most of the traditional attribute reduction models cannot obtain effective reduction rules and affect the classification accuracy of the reduced set. Therefore, we propose a new attribute reduction method, IKAR, based on the clustering background for incomplete system. IKAR uses the tolerance principle to calculate the information of knowledge granularity and to measure the similarity of data objects. When selecting attributes, the similarity of intra-cluster objects should be as large as possible, and the similarity of inter-cluster objects should be as small as possible. The innovations of this paper are manifested in the following four aspects: (1) Use of the tolerance principle to quickly calculate knowledge granularity; (2) Use of knowledge granularity to calculate the similarity of inter-cluster and intra-cluster objects; (3) Use of the idea of clustering to calculate the importance of attributes; (4) In addition to conventional time, space, and precision analysis, it also analyzes the stability of datasets with different percent missing value. All the experiments show that the IKAR algorithm is not only superior in reducing time compared to the other three algorithms, but it also has an excellent performance in terms of accuracy and stability. Of course, the IKAR algorithm also has some shortcomings. For example, it is unsuitable for datasets with multiple decision values, and complex data types and the dynamical changing of the datasets are not considered.
In our future research endeavors, we intend to work in the following four aspects. (1) We focus on attribute reduction of incomplete systems with mixed data types, especially on how to deal with missing data. (2) In addition, we will integrate the incremental learning method into the knowledge granularity reduction model in the background of clustering. (3) To address even big datasets more efficiently, applying GPU and MapReduce technologies to design parallel attribute reduction models or acceleration strategy is a very popular research topic.

Author Contributions

Conceptualization, B.L.; Supervision, R.H.; investigation, E.J.; Writing—review and editing, L.W. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Natural Science Foundation of China (61836016), the Key Subject of Chaohu University (kj22zdxk01), the Quality Improvement Project of Chaohu University on Discipline Construction (kj21gczx03), and the Provincial Natural Science Research Program of Higher Education Institutions of Anhui province (KJ2021A1030).

Data Availability Statement

Data are contained within the article.

Acknowledgments

The authors would like to thank the editors and the anonymous reviewers for their valuable comments and suggestions, which have helped immensely in improving the quality of the paper.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

RSTRough set theory
DSDecision system
EGKEqual knowledge granularity
GKKnowledge granularity
S I n t e r D i P The similarity of objects in the clustering of Di under attribute set P
ASIntraThe average intra-cluster similarity
SInterThe inter-cluster similarity
AInterThe average inter-cluster similarity
SIMThe importance of attribute
SigThe inner significance
IKARThe knowledge granularity attribute reduction for incomplete systems
NGLENeighborhood multi-granulation attribute reduction using Lebesque and Entropy
PARAThe positive attribute reduction algorithm
IEARThe information entropy attribute reduction

References

  1. Pawlak, Z. Rough sets. Int. J. Comput. Inform. Sci. 1982, 11, 341–356. [Google Scholar] [CrossRef]
  2. Wang, R.; Wang, X.; Kwong, S.; Xu, C. Incorporating diversity and informative-ness in multiple-instance active learning. IEEE Trans. Fuzzy Syst. 2017, 25, 1460–1475. [Google Scholar] [CrossRef]
  3. Wang, X.; Tsang, E.C.C.; Zhao, S.; Chen, D.; Yeung, D.S. Learning fuzzy rules from fuzzy samples based on rough set technique. Inf. Sci. 2017, 177, 4493–4514. [Google Scholar] [CrossRef]
  4. Wang, X.; Xing, H.; Li, Y.; Hua, Q.; Dong, C.R.; Pedrycz, W. A study on relationship between generalization abilities and fuzziness of base classifiers in ensemble learning. IEEE. Trans. Fuzzy. Syst. 2015, 23, 638–1654. [Google Scholar] [CrossRef]
  5. Liu, X.; Qian, Y.; Liang, J. A rule-extraction framework under multi-granulation rough sets. Int. J. Mach. Learn. Cybern. 2014, 5, 319–326. [Google Scholar] [CrossRef]
  6. Zhang, X.; Mei, C.; Chen, D.; Li, J. Multi-confidence rule acquisition and confidence-preserved attribute reduction in interval valued decision systems. Int. J. Approx. Reason. 2014, 55, 1787–1804. [Google Scholar] [CrossRef]
  7. Hu, Q.; Yu, D.; Xie, Z. Neighborhood classifiers. Expert Syst. Appl. 2008, 34, 866–876. [Google Scholar] [CrossRef]
  8. Cheruku, R.; Edla, D.R.; Kuppili, V.; Dharavath, R. RST-Bat-Miner: A fuzzy rule miner integrating rough set feature selection and bat optimization for detection of diabetes disease. Appl. Soft. Comput. 2018, 67, 764–780. [Google Scholar] [CrossRef]
  9. Hamouda, S.K.M.; Wahed, M.E.; Alez, R.H.A.; Riad, K. Robust breast cancer prediction system based on rough set theory at National Cancer Institute of Egypt. Comput. Methods Programs Biomed. 2018, 153, 259–268. [Google Scholar] [CrossRef]
  10. Jothi, G.; Inbarani, H.H. Hybrid tolerance rough set-firefly based supervised feature selection for MRI brain tumor image classification. Appl. Soft. Comput. 2016, 46, 639–651. [Google Scholar]
  11. Hao, C.; Li, J.; Fan, M.; Liu, W.; Tsang, E.C. Optimal scale selection in dynamic multi-scale decision tables based on sequential three-way decisions. Inf. Sci. 2017, 415, 213–232. [Google Scholar] [CrossRef]
  12. Liang, D.; Xu, Z.; Liu, D. Three-way decisions based on decision-theoretic rough sets with dual hesitant fuzzy information. Inf. Sci. 2017, 396, 127–143. [Google Scholar] [CrossRef]
  13. Qu, J.; Bai, X.; Gu, J.; Taghizadeh-Hesary, F.; Lin, J. Assessment of Rough set Theory in Relation to Risks Regarding Hydraulic Engineering Investment Decisions. Mathematics 2020, 8, 1308. [Google Scholar] [CrossRef]
  14. Lei, L. Wavelet neural network prediction method of stock price trend based on rough set attribute reduction. Appl. Soft Comput. 2018, 62, 923–932. [Google Scholar] [CrossRef]
  15. Singh, A.K.; Baranwal, N.; Nandi, G.C. A rough set based reasoning approach for criminal identification. Int. J. Mach. Learn. Cybern. 2019, 10, 413–431. [Google Scholar] [CrossRef]
  16. Lin, Y.; Li, J.; Lin, P.; Lin, G.; Chen, J. Feature selection via neighborhood multi-granulation fusion. Knowl. Based Syst. 2014, 67, 162–168. [Google Scholar] [CrossRef]
  17. Sun, L.; Xu, J. A granular computing approach to gene selection. Bio-Med. Mater. Eng. 2014, 24, 1307–1314. [Google Scholar] [CrossRef]
  18. Fujita, H.; Gaeta, A.; Loia, V.; Orciuoli, F. Resilience Analysis of Critical Infrastructures: A cognitive approach based on Granular Computing. IEEE Trans. Cybern. 2018, 49, 1835–1848. [Google Scholar] [CrossRef]
  19. Qian, H.Y. Granulation Mechanism and Data Modeling of Complex Data. Ph.D. Thesis, Shanxi University, Taiyuan, China, 2011. [Google Scholar]
  20. Dai, J.; Wang, W.; Zhang, C.; Qu, S. Semi-supervised attribute reduction via attribute indiscernibility. Int. J. Mach. Learn. Cybern. 2023, 14, 1445–1464. [Google Scholar] [CrossRef]
  21. Gao, C.; Wang, Z.C.; Zhou, J. Three-way approximate reduct based on information-theoretic measure. Int. J. Approx. Reason. 2022, 142, 324–337. [Google Scholar] [CrossRef]
  22. Yang, L.; Qin, K.; Sang, B.; Fu, C. A novel incremental attribute reduction by using quantitative dominance-based neighborhood self-information. Know. Based Syst. 2023, 261, 110200. [Google Scholar] [CrossRef]
  23. Wang, Z.H.; Zhang, X.P. The granulation attribute reduction of multi-label data. Appl. Intell. 2023, 53, 19266–19284. [Google Scholar] [CrossRef]
  24. Yu, B.; Hu, Y.; Kang, Y.; Cai, M. A novel variable precision rough set attribute reduction algorithm based on local attribute significance. Int. J. Approx. Reason. 2023, 157, 88–104. [Google Scholar] [CrossRef]
  25. Zhang, S.C.; Cheng, D.B.; Hu, R.Y.; Deng, Z. Supervised feature selection algorithm via discriminative ridge regression. World Wide Web 2018, 21, 1545–1562. [Google Scholar] [CrossRef]
  26. Liu, T.; Hu, R.Y.; Zhu, Y.X. Completed sample correlations and feature dependency-based unsupervised feature selection. Multim. Tools Appl. 2023, 82, 15305–15326. [Google Scholar] [CrossRef]
  27. Devi, P.; Kizielewicz, B.; Guleria, A.; Shekhovtsov, A.; Gandotra, N.; Saini, N.; Sałabun, W. Dimensionality reduction technique under picture fuzzy environment and its application in decision making. Int. J. Knowl. Based Intell. Eng. Syst. 2023, 27, 87–104. [Google Scholar] [CrossRef]
  28. Wen, H.T.; Zhao, S.X.; Liang, M.S. Unsupervised attribute reduction algorithm for mixed data based on fuzzy optimal approximation set. Mathematics 2023, 11, 3452. [Google Scholar] [CrossRef]
  29. Li, Z.; Liao, S.; Qu, L.; Song, Y. Attribute selection approaches for incomplete interval-value data. J. Intell. Fuzzy Syst. 2021, 40, 8775–8792. [Google Scholar] [CrossRef]
  30. Liu, X.F.; Dai, J.H.; Chen, J.L.; Zhang, C.C. A fuzzy α-similarity relation-based attribute reduction approach in incomplete interval-valued information systems. Appl. Soft Comput. 2021, 109, 107593. [Google Scholar] [CrossRef]
  31. Dai, J.H.; Wang, Z.Y.; Huang, W.Y. Interval- valued fuzzy discernibility pair approach for attribute reduction in incomplete interval-valued information systems. Inf. Sci. 2023, 642, 119215. [Google Scholar] [CrossRef]
  32. Song, Y.; Luo, D.; Xie, N.; Li, Z. Uncertainty measurement for incomplete set-valued data with application to attribute reduction. Int. J. Mach. Learn. Cybern. 2022, 13, 3031–3069. [Google Scholar] [CrossRef]
  33. Zhou, Y.; Bao, Y.L. A Novel Attribute Reduction Algorithm for Incomplete Information Systems Based on a Binary Similarity Matrix. Symmetry 2023, 15, 674. [Google Scholar] [CrossRef]
  34. Zhang, C.L.; Li, J.J.; Lin, Y.D. Knowledge reduction of pessimistic multigranulation rough sets in incomplete information systems. Soft Comput. 2021, 25, 12825–12838. [Google Scholar] [CrossRef]
  35. He, J.; Qu, L.; Wang, Z.; Chen, Y.; Luo, D.; Wen, C.F. Attribute reduction in an incomplete categorical decision information system based on fuzzy rough sets. Artif. Intell. Rev. 2022, 55, 5313–5348. [Google Scholar] [CrossRef]
  36. Srirekha, B.; Sathish, S.; Devi, R.M. Attribute reduction on SE-ISI Concept Lattice for an Incomplete Context using object ranking. Mathematics 2023, 11, 1585. [Google Scholar] [CrossRef]
  37. Cornelis, C.; Jensen, R.; Martin, G.H.; Slezak, D. Attribute selection with fuzzy decision reducts. Inf. Sci. 2010, 180, 209–224. [Google Scholar] [CrossRef]
  38. Liu, G.L.; Feng, Y.B.; Yang, J.T. A common attribute reduction form for information systems. Know. Based Syst. 2020, 193, 105466. [Google Scholar] [CrossRef]
  39. Nguyen, N.T.; Sartra, W. A novel feature selection Method for High-Dimensional Mixed decision tables. IEEE Trans. Neural Netw. Learn. Syst. 2022, 33, 3024–3037. [Google Scholar]
  40. Zhang, M.; Chen, D.G.; Yang, Y.Y. A new algorithm of attribute reduction based on fuzzy clustering. In Proceedings of the International Conference on Machine Learning and Cybernetics, Tianjin, China, 14–17 July 2013; Volume 7, pp. 14–17. [Google Scholar]
  41. Jia, H.J.; Ding, S.F.; Ma, H.; Xing, W.Q. Spectral Clustering with Neighborhood Attribute Reduction Based on Information. Entropy. J. Comput. 2014, 9, 1316–1324. (In Chinese) [Google Scholar] [CrossRef]
  42. Zhao, R.N.; Gu, L.Z.; Zhu, X.N. Combining Fuzzy C-Means Clustering with Fuzzy Rough Feature Selection. Appl. Sci. 2019, 9, 679. [Google Scholar] [CrossRef]
  43. Jia, X.Y.; Rao, Y.Y.; Shang, L.; Li, T.G. Similarity-based attribute reduction in rough set theory: A clustering perspective. Int. J. Mach. Learn. Cybern. 2020, 11, 1047–1060. [Google Scholar] [CrossRef]
  44. Zhang, Y.; Chen, Z. Information Entropy-Based Attribute Reduction for Incomplete Set-Valued Data. IEEE Access 2021, 10, 8864–8882. [Google Scholar] [CrossRef]
  45. Shu, W.H.; Qian, W.B. A fast approach to attribute reduction from perspective of attribute measures in incomplete decision systems. Know. Based Syst. 2014, 72, 60–71. [Google Scholar] [CrossRef]
  46. Sun, L.; Wang, L.; Ding, W.; Qian, Y.; Xu, J. Neighborhood multi-granulation rough sets-based attribute reduction using Lebesgue and entropy measures in incomplete neighborhood decision systems. Know. Based Syst. 2020, 192, 105373. [Google Scholar] [CrossRef]
Figure 1. The execution process of the IKAR algorithm.
Figure 1. The execution process of the IKAR algorithm.
Mathematics 12 00333 g001
Figure 2. The framework chart of experiments.
Figure 2. The framework chart of experiments.
Mathematics 12 00333 g002
Figure 3. Variation in classification accuracy for different percentages of missing values with classifiers SMO.
Figure 3. Variation in classification accuracy for different percentages of missing values with classifiers SMO.
Mathematics 12 00333 g003
Figure 4. Variation in classification accuracy for different percentages of missing values with classifiers RF.
Figure 4. Variation in classification accuracy for different percentages of missing values with classifiers RF.
Mathematics 12 00333 g004
Figure 5. Trend in classification accuracy of different algorithms under the SMO classifier.
Figure 5. Trend in classification accuracy of different algorithms under the SMO classifier.
Mathematics 12 00333 g005
Figure 6. Trend in classification accuracy of different algorithms under the RF classifier.
Figure 6. Trend in classification accuracy of different algorithms under the RF classifier.
Mathematics 12 00333 g006
Figure 7. Bonferroni–Dunn test with SMO.
Figure 7. Bonferroni–Dunn test with SMO.
Mathematics 12 00333 g007
Figure 8. Bonferroni–Dunn test with RF.
Figure 8. Bonferroni–Dunn test with RF.
Mathematics 12 00333 g008
Table 1. Incomplete information system.
Table 1. Incomplete information system.
Uabced
o100010
o2001*1
o301010
o401*01
o5010*1
o6**111
o710*02
o810102
o9**012
‘*’ represents the missing value about some attributes.
Table 2. Description of twelve datasets.
Table 2. Description of twelve datasets.
IDDatasetsAbbreviation|U||C||D|
1PromotersProm106572
2Heart-statlogHear270132
3HepatitisHepa155192
4HandWrittenHand56206410
5Chess kr-kpChes3196362
6SpliceSpli3190613
7LettersLett20,0001726
8VoteVote435162
9MushroomMush8124222
10QsarQsar1055422
11ShuttleShut43,50097
12SatimageSati6435366
Table 3. The attribute reduction size with the four methods on the twelve UCI datasets.
Table 3. The attribute reduction size with the four methods on the twelve UCI datasets.
DatasetsIKARNGLEIEARPARA
Prom5455
Hear1010910
Hepa910710
Hand12111011
Ches29302930
Spli91099
Lett9989
Vote109810
Mush4555
Qsar31302931
Shut4545
Sati1010911
Ave11.83311.91711.0012.083
Best41111
Table 4. The run time with the four methods on the twelve UCI datasets.
Table 4. The run time with the four methods on the twelve UCI datasets.
DatasetsIKARNGLEIEARPARA
Prom0.982.0423.2311.553
Hear0.030.162.090.33
Hepa0.060.112.130.25
Hand92.9381.613222.8991.28
Ches6.35140.983075.25590.79
Spli35.43205.866557.61309.79
Lett3.986.93117.878.47
Vote0.060.386.091.01
Mush4.45411.094544.04827.16
Qsar3.844.9798.766.61
Shut8.98886.083974.701133.25
Sati25.30125.612322.02154.04
Ave15.20155.491993.89260.38
Best11100
Table 5. Ranking and standard of classification accuracy under SMO classifier.
Table 5. Ranking and standard of classification accuracy under SMO classifier.
DatasetsIKARNGLEIEARPARA
STDRankSTDRankSTDRankSTDRank
Prom±0.0311±1.3622±2.0194±1.2943
Heas±0.0151±1.4553±2.3184±1.5212
Hepa±0.0381±2.4834±1.9854±1.9843
Hand±0.1251±2.7842±2.5374±2.1833
Krkp±0.2472±1.5614±1.7831±3.4203
Spli±0.0351±2.4892±3.0154±3.0193
Lett±0.5121±3.0163±1.9854±2.1842
Vote±0.1871±1.3922±2.4174±2.4783
Mush±0.2611±2.7653±2.5814±2.1042
Qsar±0.3821±2.0543±2.9324±3.0132
Shut±0.0491±1.9373±1.9954±1.3292
Sati±0.5361±2.3493±3.2464±2.5863
AveRank 1.08 2.75 3.58 2.58
Table 6. Ranking and standard of classification accuracy under RF classifier.
Table 6. Ranking and standard of classification accuracy under RF classifier.
DatasetsIKARNGLEIEARPARA
STDRankSTDRankSTDRankSTDRank
Prom±0.1811±1.6732±2.8294±2.6393
Heas±0.2731±2.4423±2.7194±2.5212
Hepa±0.3571±2.8734±2.6832±3.0383
Hand±0.3621±2.3612±3.7314±2.3593
Krkp±0.6521±2.0483±1.9843±2.7632
Spli±0.2481±1.9942±3.5654±3.5583
Lett±0.5831±3.2463±2.7834±3.1602
Vote±0.6521±1.8172±3.0154±2.6293
Mush±0.3931±3.0923±2.9284±1.9722
Qsar±0.4881±2.6362±1.9974±3.2153
Shut±0.3911±2.7434±2.9633±2.3882
Sati±0.4751±1.9824±2.8373±3.0863
AveRank 1 2.83 3.67 2.5
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liang, B.; Jin, E.; Wei, L.; Hu, R. Knowledge Granularity Attribute Reduction Algorithm for Incomplete Systems in a Clustering Context. Mathematics 2024, 12, 333. https://doi.org/10.3390/math12020333

AMA Style

Liang B, Jin E, Wei L, Hu R. Knowledge Granularity Attribute Reduction Algorithm for Incomplete Systems in a Clustering Context. Mathematics. 2024; 12(2):333. https://doi.org/10.3390/math12020333

Chicago/Turabian Style

Liang, Baohua, Erli Jin, Liangfen Wei, and Rongyao Hu. 2024. "Knowledge Granularity Attribute Reduction Algorithm for Incomplete Systems in a Clustering Context" Mathematics 12, no. 2: 333. https://doi.org/10.3390/math12020333

APA Style

Liang, B., Jin, E., Wei, L., & Hu, R. (2024). Knowledge Granularity Attribute Reduction Algorithm for Incomplete Systems in a Clustering Context. Mathematics, 12(2), 333. https://doi.org/10.3390/math12020333

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop