Next Article in Journal
On a Class of Hermite-Obreshkov One-Step Methods with Continuous Spline Extension
Previous Article in Journal
Some Identities for Euler and Bernoulli Polynomials and Their Zeros
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Single-Valued Neutrosophic Clustering Algorithm Based on Tsallis Entropy Maximization

1
School of Science, Xi’an Polytechnic University, Xi’an 710048, China
2
Department of Mathematics, University of New Mexico, Gallup, NM 87301, USA
3
School of Textile and Materials, Xi’an Polytechnic University, Xi’an 710048, China
*
Author to whom correspondence should be addressed.
Axioms 2018, 7(3), 57; https://doi.org/10.3390/axioms7030057
Submission received: 10 May 2018 / Revised: 19 July 2018 / Accepted: 20 July 2018 / Published: 17 August 2018

Abstract

:
Data clustering is an important field in pattern recognition and machine learning. Fuzzy c-means is considered as a useful tool in data clustering. The neutrosophic set, which is an extension of the fuzzy set, has received extensive attention in solving many real-life problems of inaccuracy, incompleteness, inconsistency and uncertainty. In this paper, we propose a new clustering algorithm, the single-valued neutrosophic clustering algorithm, which is inspired by fuzzy c-means, picture fuzzy clustering and the single-valued neutrosophic set. A novel suitable objective function, which is depicted as a constrained minimization problem based on a single-valued neutrosophic set, is built, and the Lagrange multiplier method is used to solve the objective function. We do several experiments with some benchmark datasets, and we also apply the method to image segmentation using the Lena image. The experimental results show that the given algorithm can be considered as a promising tool for data clustering and image processing.

1. Introduction

Data clustering is one of the most important topics in pattern recognition, machine learning and data mining. Generally, data clustering is the task of grouping a set of objects in such a way that objects in the same group (cluster) are more similar to each other than to those in other groups (clusters). In the past few decades, many clustering algorithms have been proposed, such as k-means clustering [1], hierarchical clustering [2], spectral clustering [3], etc. The clustering technique has been used in many fields, including image analysis, bioinformatics, data compression, computer graphics and so on [4,5,6].
The k-means algorithm is one of the typical hard clustering algorithms that is widely used in real applications due to its simplicity and efficiency. Unlike hard clustering, the fuzzy c-means (FCM) algorithm [7] is one of the most popular soft clustering algorithms in which each data point belongs to a cluster to some degree that is specified by membership degrees in [ 0 , 1 ] , and the sum of the clusters for each of the data should be equal to one. In recent years, many improved algorithms for FCM have been proposed. There are three main ways to build the clustering algorithm. First is extensions of the traditional fuzzy sets. In this way, numerous fuzzy clustering algorithms based on the extension fuzzy sets, such as the intuitionistic fuzzy set, the Type-2 fuzzy set, etc., are built. By replacing traditional fuzzy sets with intuitionistic fuzzy set, Chaira introduced the intuitionistic fuzzy clustering (IFC) method in [8], which integrated the intuitionistic fuzzy entropy with the objective function. Hwang and Rhee suggested deploying FCM on (interval) Type-2 fuzzy set sets in [9], which aimed to design and manage uncertainty for fuzzifier m. Thong and Son proposed picture fuzzy clustering based on the picture fuzzy set (PFS) in [10]. Second, the kernel-based method is applied to improve the fuzzy clustering quality. For example, Graves and Pedrycz presented a kernel version of the FCM algorithm, namely KFCM in [11]. Ramathilagam et al. analyzed the Lung Cancer database by incorporating the hyper tangent kernel function [12]. Third, adding regularization terms to the objective function is used to improve the clustering quality. For example, Yasuda proposed an approach to FCM based on entropy maximization in [13]. Of course, we can use these together to obtain better clustering quality.
The neutrosophic set was proposed by Smarandache [14] in order to deal with real-world problems. Now, the neutrosophic set is gaining significant attention in solving many real-life problems that involve uncertainty, impreciseness, incompleteness, inconsistency and indeterminacy. A neutrosophic set has three membership functions, and each membership degree is a real standard or non-standard subset of the nonstandard unit interval ] 0 , 1 + [ = 0 [ 0 , 1 ] 1 + . Wang et al. [15] introduced single-valued neutrosophic sets (SVNSs), which are an extension of intuitionistic fuzzy sets. Moreover, the three membership functions are independent, and their values belong to the unit interval [ 0 , 1 ] . In recent years, the studies of SVNSs have been rapidly developing. For example, Majumdar and Samanta [16] studied the similarity and entropy of SVNSs. Ye [17] proposed correlation coefficients of SVNSs and applied them to single-valued neutrosophic decision-making problems, etc. Zhang et al. in [18] proposed a new definition of the inclusion relation of neutrosophic sets (which is also called the Type-3 inclusion relation), and a new method of ranking neutrosophic sets was given. Zhang et al. in [19] studied neutrosophic duplet sets, neutrosophic duplet semi-groups and cancelable neutrosophic triplet groups.
The clustering methods by the neutrosophic set have been studied deeply. In [20], Ye proposed a single-valued neutrosophic minimum spanning tree (SVNMST) clustering algorithm, and he also introduced single-valued neutrosophic clustering methods based on similarity measures between SVNSs [21]. Guo and Sengur introduced the neutrosophic c-means clustering algorithm [22], which was inspired by FCM and the neutrosophic set framework. Thong and Son did significant work on clustering based on PFS. In [10], a picture fuzzy clustering algorithm, called FC-PFS, was proposed. In order to determine the number of clusters, they built an automatically determined most suitable number of clusters based on particle swarm optimization and picture composite cardinality for a dataset [23]. They also extended the picture fuzzy clustering algorithm for complex data [24]. Unlike the method in [10], Son presented a novel distributed picture fuzzy clustering method on the picture fuzzy set [25]. We can note that the basic ideas of the fuzzy set, the intuitionistic fuzzy set and the SVNS are consistent in the data clustering, but there are differences in the representation of the objects, so that the clustering objective functions are different. Thus, the more adequate description can be better used for clustering. Inspired by FCM, FC-PFS, SVNS and the maximization entropy method, we propose a new clustering algorithm, the single-valued neutrosophic clustering algorithm based on Tsallis entropy maximization (SVNCA-TEM), in this paper, and the experimental results show that the proposed algorithm can be considered as a promising tool for data clustering and image processing.
The rest of paper is organized as follows. Section 2 shows the related work on FCM, IFC and FC-PFS. Section 3 introduces the proposed method, using the Lagrange multiplier method to solve the objective function. In Section 4, the experiments on some benchmark UCI datasets indicate that the proposed algorithm can be considered as a useful tool for data clustering and image processing. The last section draws the conclusions.

2. Related Works

In general, suppose dataset D = { X 1 , X 2 , , X n } includes n data points, each of the data X i = { x i 1 ; x i 2 ; ; x i d } R d is a d-dim feature vector. The aim of clustering is to get k disjoint clusters { C j | , j = 1 , 2 , , k } , satisfying C j j j C j = and D = j = 1 k C j . In the following, we will briefly introduce three fuzzy clustering methods, which are FCM, IFC and FC-PFS.

2.1. Fuzzy c-Means

The FCM was proposed in 1984 [7]. FCM is a data clustering technique wherein each data point belongs to a cluster to some degree that is specified by a membership grade. A data point X i of cluster C j is denoted by the term μ i j , which shows the fuzzy membership degree of the i-th data point in the j-th cluster. We use V = { V 1 , V 2 , , V k } to describe the cluster centroids of the clusters, and V j R d is the cluster centroid of C j . The FCM is based on the minimization of the following objective function:
J = i = 1 n j = 1 k u i j m x i V j 2 ,
where m represents the fuzzy parameter and m 1 . The constraints for (1) are,
l = 1 k μ i l = 1 , μ i j [ 0 , 1 ] , i = 1 , 2 , , n , j = 1 , 2 , , k .
Using the Lagrangian method, the iteration scheme to calculate cluster centroids V j and the fuzzy membership degrees μ i j of the objective function (1) is as follows:
V j = i = 1 n μ i j m X i i = 1 n μ i j m , j = 1 , 2 , , k .
μ i j = ( l = 1 k ( X i V j X i V l ) 2 m 1 ) 1 . i = 1 , 2 , , n . j = 1 , 2 , , k .
The iteration will not stop until it reaches the maximum iterations or | J ( t ) J ( t 1 ) | < ϵ , where J ( t ) and J ( t 1 ) are the objection function value at ( t ) -th and ( t 1 ) -th iterations, and ϵ is a termination criterion between zero and 0.1 . This procedure converges to a local minimum or a saddle point of J. Finally, each data point is assigned to a different cluster according to the fuzzy membership value, that is X i belongs to C l if μ i l = max ( μ i 1 , μ i 2 , , μ i k ) .

2.2. Intuitionistic Fuzzy Clustering

The intuitionistic fuzzy set is an extension of fuzzy sets. Chaira proposed intuitionistic fuzzy clustering (IFC) [8], which integrates the intuitionistic fuzzy entropy with the objective function of FCM. The objective function of IFS is:
J = i = 1 n j = 1 k μ i j m X i V j 2 + j = 1 k π j * e 1 π j * ,
where π j * = 1 n i = 1 n π i j , and π i j is the hesitation degree of X i for C j . The constraints of IFC are similar to (2). Hesitation degree π i k is initially calculated using the following form:
π i j = 1 μ i j ( 1 u i j α ) 1 / α , where α [ 0 , 1 ] ,
and the intuitionistic fuzzy membership values are obtained as follows:
μ i j * = μ i j + π i j ,
where μ i j * ( μ i j ) denotes the intuitionistic (conventional) fuzzy membership of the i-th data in the j-th class. The modified cluster centroid is:
V j = i = 1 n μ i j * m X i i = 1 n μ i j * m , j = 1 , 2 , , k .
The iteration will not stop until it reaches the maximum iterations or the difference between μ i j * ( t ) and μ i j * ( t 1 ) is not larger than a pre-defined threshold ϵ , that is max i , j | μ i j * ( t ) μ i j * ( t 1 ) | < ϵ .

2.3. Picture Fuzzy Clustering

In [26], Cuong introduced the picture fuzzy set (which is also called the standard neutrosophic set [27]), which is defined on a non-empty set S, A ˙ = { x , μ A ˙ ( x ) , η A ˙ ( x ) , γ A ˙ ( x ) | x S } , where μ A ˙ ( x ) is the positive degree of each element x X , η A ˙ ( x ) is the neutral degree and γ A ˙ ( x ) is the negative degree satisfying the constraints,
μ A ˙ ( x ) , η A ˙ ( x ) , γ A ˙ ( x ) [ 0 , 1 ] , x S . μ A ˙ ( x ) + η A ˙ ( x ) + γ A ˙ ( x ) 1 , x S .
The refusal degree of an element is calculated as:
ξ A ˙ ( x ) = 1 ( μ A ˙ ( x ) + η A ˙ ( x ) + γ A ˙ ( x ) ) , x S .
In [10], Thong and Son proposed picture fuzzy clustering (FC-PFS), which is related to neutrosophic clustering. The objective function is:
J = i = 1 n j = 1 k ( μ i j ( 2 ξ i j ) ) m X i V j 2 + i = 1 n j = 1 k η i j ( log η i j + ξ i j ) .
where i = 1 , , n , j = 1 , , k . μ i j , η i j and ξ i j are the positive, neutral and refusal degrees, respectively, for which each data point X i belongs to cluster C j . Denote μ , η and ξ as the matrices whose elements are μ i j , η i j and ξ i j , respectively. The constraints for FC-PFS are defined as follows:
u i j , η i j , ξ i j [ 0 , 1 ] , u i j + η i j + ξ i j 1 , l = 1 k ( u i l ( 2 ξ i l ) ) = 1 , l = 1 k ( η i l + ξ i l / k ) = 1 .
Using the Lagrangian multiplier method, the iteration scheme to calculate μ i j , η i j , ξ i j and V j for the model (11,12) is as the following equations:
ξ i j = 1 ( μ i j + η i j ) ( 1 ( μ i j + η i j ) α ) 1 / α , where α [ 0 , 1 ] , ( i = 1 , , n , j = 1 , , k ) ,
μ i j = 1 l = 1 k ( 2 ξ i j ) ( X i V j X i V l ) 2 m 1 , ( i = 1 , , n , j = 1 , , k ) ,
η i j = e ξ i j l = 1 k e ξ i l ( 1 1 k l = 1 k ξ i l ) , ( i = 1 , , n , j = 1 , , k ) ,
V j = i = 1 n ( μ i j ( 2 ξ i j ) ) m X i i = 1 n ( μ i j ( 2 ξ i j ) ) m , ( j = 1 , , k ) .
The iteration will not stop until it reaches the maximum iterations or μ ( t ) μ ( t 1 ) + η ( t ) η ( t 1 ) + ξ ( t ) ξ ( t 1 ) < ϵ .

3. The Proposed Model and Solutions

Definition 1.
[15] Set U as a space of points (objects), with a generic element in U denoted by u. A SVNS A in U is characterized by three membership functions, a truth membership function T A , an indeterminacy membership function I A and a falsity-membership function F A , where u U ,   T A ( u ) , I A ( u ) , F A ( u ) [ 0 , 1 ] . That is, T A : U [ 0 , 1 ] , I A : U [ 0 , 1 ] and F A : U [ 0 , 1 ] . There is no restriction on the sum of T A ( u ) , I A ( u ) and F A ( u ) ; thus, 0 T A ( u ) + I A ( u ) + F A ( u ) 3 .
Moreover, the hesitate membership function is defined as H A : U [ 0 , 3 ] and u U , T A ( u ) + I A ( u ) + F A ( u ) + H A ( u ) = 3 .
Entropy is a key concept in the uncertainty field. It is a measure of the uncertainty of a system or a piece of information. It is an improvement of information entropy. The Tsallis entropy [28], which is a generalization of the standard Boltzmann–Gibbs entropy, is defined as follows.
Definition 2.
[28] Let X be a finite set and X be a a random variable taking values x X , with distribution p ( x ) . The Tsallis entropy is defined as S m ( X ) = 1 m 1 ( 1 x X p ( x ) m ) , where m > 0 and m 1 .
For FCM, μ i j denotes the fuzzy membership degree of X i to C j , and supports j = 1 k μ i j = 1 . From Definition 2, the Tsallis entropy of μ can be described by S m ( μ ) = i = 1 n 1 m 1 ( 1 j = 1 k μ i j m ) . n being a fixed number, Yasuda [13] used the following formulary to describe the the Tsallis entropy of μ :
S m ( μ ) = 1 m 1 ( i = 1 n j = 1 k μ i j m 1 ) .
The maximum entropy principle has been widely applied in many fields, such as spectral estimation, image restoration, error handling of measurement theory, and so on. In the following, the maximum entropy principle is applied to the single-valued neutrosophic set clustering. After the objection function of clustering is built, the maximum fuzzy entropy is used to regularize variables.
Suppose that there is a dataset D consisting of n data points in d dimensions. Let μ i j , γ i j , η i j and ξ i j be the truth membership degree, falsity-membership degree, indeterminacy membership degree and hesitate membership degree, respectively, that each data point X i belongs to cluster C j . Denote μ ,   γ ,   η and ξ as the matrices, the elements of which are μ i j ,   γ i j ,   η i j and ξ i j , respectively, where ξ i j = 3 μ i j γ i j η i j . The single-valued neutrosophic clustering based on Tsallis entropy maximization (SVNC-TEM) is the minimization of the following objective function:
J = i = 1 n j = 1 k ( μ i j ( 4 ξ i j γ i j ) ) m X i V j 2 + ρ m 1 ( i = 1 n j = 1 k ( u i j ( 4 γ i j ξ i j ) ) m 1 ) + i = 1 n j = 1 k η i j ( log η i j + ξ i j / 3 ) ,
The constraints are given as follows:
μ i j ,   γ i j ,   η i j [ 0 , 1 ] , ξ i j [ 0 , 3 ] , ( i = 1 , 2 , , n , j = 1 , 2 , , k )
l = 1 k ( u i l ( 4 γ i l ξ i l ) ) = 1 , ( i = 1 , 2 , , n ) ,
l = 1 k ( η i l + ξ i l / ( 3 * k ) ) = 1 , ( i = 1 , 2 , , n )
The proposed model in Formulary (18)–(21) is applied to the maximum entropy principle of the SVNS. Now, let us summarize the major points of this model as follows.
  • The first term of the objection function (18) describes the weighted distance sum of each data point X i to the cluster center V j . μ i j being from the positive aspect and ( 4 ξ i j γ i j ) (four is selected in order to guarantee μ i j [ 0 , 1 ] in the iterative calculation) from the negative aspect, denoting the membership degree for X i to V j , we use μ i j ( 4 ξ i j γ i j ) to represent the “integrated true” membership of the i-th data point in the j-th cluster. From the maximum entropy principle, the best to represent the current state of knowledge is the one with largest entropy, so the second term of the objection function (18) describes the negative Tsallis entropy of μ ( 4 γ ξ ) , which means that the minimization of (18) is the maximum Tsallis entropy. ρ is the regularization parameter. If γ = η = ξ = 0 , the proposed model returns the FCM model.
  • Formulary (19) guarantees the definition of the SVNS (Definition 1).
  • Formulary (20) implies that the “integrated true” membership of a data point X i to the cluster center V j satisfies the sum-row constraint of memberships. For convenience, we set T i j = μ i j ( 4 ξ i j γ i j ) , and X i belongs to class C l if T i l = max ( T i 1 , T i 2 , , T i k ) .
  • Equation (21) guarantees the working of the SVNS since at least one of two uncertain factors, namely indeterminacy membership degree and hesitate membership degree, always exists in the model.
Theorem 1.
The optimal solutions of the systems(1821) are:
V j = i = 1 n ( μ i j ( 4 γ i j ξ i j ) ) m X i i = 1 n ( μ i j ( 4 γ i j ξ i j ) ) m ,
μ i j = 1 l = 1 k ( 4 γ i j ξ i j ) ( X i V j 2 + ρ m 1 X i V l 2 + ρ m 1 ) 1 m 1 ,
γ i j = 4 ξ i j 1 u i j l = 1 k ( X i V j 2 + ρ m 1 X i V l 2 + ρ m 1 ) 1 m 1 ,
η i j = ( 1 1 3 k l = 1 k ξ i l ) e ξ i j l = 1 k e ξ i l ,
ξ i j = 3 μ i j γ i j η i j .
Proof. 
The Lagrangian multiplier of the optimization model (1821) is:
J = i = 1 n j = 1 k ( u i j ( 4 γ i j ξ i j ) ) m X i V j 2 + ρ m 1 ( i = 1 n j = 1 k ( u i j ( 4 γ i j ξ i j ) ) m 1 ) + i = 1 n j = 1 k η i j ( log η i j + ξ i j / 3 ) + i = 1 n λ i ( j = 1 C μ i j ( 4 γ i j ξ i j ) m ) 1 ) + i = 1 n χ i ( j = 1 k ( η i j + ξ i j / ( 3 k ) ) 1 ) ,
where λ i and χ i are Lagrangian multipliers.
In order to get V j , taking the derivative of the objective function with respect to V j , we have J V j = i = 1 n ( μ i j ( 4 γ i j ξ i j ) ) m ( 2 X i + 2 V j ) . Since J V j = 0 , so i = 1 n ( μ i j ( 4 η i j ξ i j ) ) m ( 2 X i + 2 V j ) = 0 i = 1 n ( μ i j ( 4 η i j ξ i j ) ) m X i = i = 1 n ( μ i j ( 4 η i j ξ i j ) ) m V j V j = i = 1 n ( μ i j ( 4 η i j ξ i j ) ) m X i i = 1 N ( μ i j ( 4 η i j ξ i j ) ) m .
Similarly, J μ i j = m μ i j m 1 ( 4 ξ i j η i j ) m X i V j 2 + ρ m m 1 μ i j m 1 ( 4 ξ i j η i j ) m ) + λ i ( 4 ξ i j η i j ) = 0 μ i j m 1 ( 4 γ i j ξ i j ) m 1 ( m X i V j 2 + ρ m m 1 ) + λ i = 0 μ i j = 1 4 γ i j ξ i j ( λ i m X k V j 2 + ρ m m 1 ) 1 m 1 .
From (20), we can get l = 1 k ( λ i m X i V l 2 + m ρ m 1 ) 1 m 1 = 1 , that is λ i = ( 1 l = 1 k 1 ( m X i V l 2 + m ρ m 1 ) 1 m 1 ) m 1 , so μ i j = 1 l = 1 k ( 4 ξ i j η i j ) ( X i V j 2 + ρ m 1 X i V l 2 + ρ m 1 ) 1 m 1 ; thus, (23) holds.
From (23), we can also get μ i j ( 4 γ i j ξ i j ) = 1 l = 1 k ( X i V j 2 + ρ m 1 X i V l 2 + ρ m 1 ) 1 m 1 , so γ i j = 4 ξ i j 1 u i j i = 1 C ( X i V j 2 + ρ m 1 X i V k 2 + ρ m 1 ) 1 m 1 ; thus, (24) holds.
Similarly, L η i j = log η i j + 1 χ i + ξ i j = 0 η i j = e ( χ i 1 ξ i j ) . From (21), we have l = 1 k e χ i 1 ξ i l + 1 3 k l = 1 k ξ i l = 1 e χ i 1 l = 1 k e ξ i l = 1 1 3 k l = 1 k ξ i l e χ i 1 = 1 1 3 k l = 1 k ξ i l l = 1 k e ξ i l . Therefore, we have η i j = ( 1 1 3 k l = 1 k ξ i l ) e ξ i j l = 1 k e ξ i l .
Finally, from Definition 1, we can get ξ i j = 3 μ i j γ i j η i j . Thus, (26) holds. ☐
Theorem 1 guarantees the convergence of the proposed method. The detailed descriptions of SVNC-TEM algorithm are presented in the following Algorithm 1:
Algorithm 1: SVNC-TEM
Input: Dataset D = { X 1 , X 2 , , X n } (n elements, d dimensions), number of clusters k, maximal number of iterations (Max-Iter), parameters: m , ϵ , ρ
Output: Cluster result
1: t = 0 ;
2:Initialize μ , γ , ξ , satisfies Constraints (19) and (20);
3:Repeat
4:    t = t + 1 ;
5:   Update V j ( t ) , ( j = 1 , 2 , , k ) using Equation (22);
6:   Update μ i j ( t ) , ( i = 1 , 2 , , n , j = 1 , 2 , , k ) using Equation (23);
7:   Update γ i j ( t ) , ( i = 1 , 2 , , n , j = 1 , 2 , , k ) using Equation (24);
8:   Update η i j ( t ) , ( i = 1 , 2 , , n , j = 1 , 2 , , k ) using Equation (25);
9:   Update ξ i j ( t ) , ( i = 1 , 2 , , n , j = 1 , 2 , , k ) using Equation (26);
10:  Update T i j ( t ) = μ i j ( t ) ( 4 γ i j ( t ) ξ i j ( t ) ) , ( i = 1 , 2 , , n , j = 1 , 2 , , k ) ;
11:  Update J ( t ) using Equation (18);
12:Until | J ( t ) J ( t 1 ) | < ϵ or Max-Iter is reached.
13:Assign X i ( i = 1 , 2 , , n ) to the l-th class if T i l = max ( T i 1 , T i 2 , , T i k ) .
Compared to FCM, the proposed algorithm needs additional time to calculate μ , γ , η and ξ in order to more precisely describe the object and get better performance. If the dimension of the given dataset is d, the number of objects is n, the number of clusters is c and the number of iterations is t, then the computational complexity of the proposed algorithm is O ( d n c t ) . We can see that the computational complexity is very high if d and n are large.

4. Experimental Results

In this section, some experiments are intended to validate the effectiveness of the proposed algorithm SVNC-TEM for data clustering. Firstly, we used an artificial dataset to show that SVNC-TEM can cluster well. Secondly, the proposed clustering method was used in image segmentation using an example. Lastly, we selected five benchmark datasets, and SVNC-TEM was compared to four state-of-the-art clustering algorithms, which were: k-means, FCM, IFC and FS-PFS.
In the experiments, the parameter m was selected as two and ϵ = 10 5 . The maximum iterations (Max-Iter) = 100 . The selected datasets have class labels, so the number of cluster k was known in advance. All the codes in the experiments were implemented in MATLAB R2015b.

4.1. Artificial Data to Cluster by the SVNC-TEM Algorithm

The activities of the SVNC-TEM algorithm are illustrated to cluster artificial data, which was two-dimensional data and had 100 data points, into four classes. We use an example to show the clustering process of the proposed algorithm. The distribution of data points is illustrated in Figure 1a. Figure 1b–e shows the cluster results when the number of iterations was t = 1, 5, 10, 20, respectively. We can see that the clustering result was obtained when t = 20 . Figure 1f shows the final results of the clustering; the number of iterations was 32. We can see that the proposed algorithm gave correct clustering results from Figure 1.

4.2. Image Segmentation by the SVNC-TEM Algorithm

In this subsection, we use the proposed algorithm for image segmentation. As a simple example, the Lena image was used to test the proposed algorithm for image segmentation. Through this example, we wish to show that the proposed algorithm can be applied to image segmentation. Figure 2a is the original Lena image. Figure 2b shows the segmentation images when the number of clusters was k = 2, and we can see that the quality of the image was greatly reduced. Figure 2c–f shows the segmentation images when the number of clusters was k = 5, 8, 11 and 20, respectively. We can see that the quality of segmentation image was improved very much with the increase of the clustering number.
The above two examples demonstrate that the proposed algorithm can be effectively applied to clustering and image processing. Next, we will further compare the given algorithm to other state-of-art clustering algorithms on benchmark datasets.

4.3. Comparison Analysis Experiments

In order to verify the clustering performance, in this subsection, we experiment with five benchmark datasets of the UCI Machine Learning Repository, which are IRIS, CMC, GLASS, BALANCE and BREAST. These datasets were used to test the performance of the clustering algorithm. Table 1 shows the details of the characteristics of the datasets.
In order to compare the performance of the clustering algorithms, three evaluation criteria were introduced as follows.
Given one data point X i , denote p i as the truth class and q i as the predicted clustering class. The clustering accuracy (ACC) measure is evaluated as follows:
A C C = i = 1 n δ ( p i , m a p ( q i ) ) n ,
where n is the total number of data points, δ ( x , y ) = 1 if x = y ; otherwise, δ ( x , y ) = 0 . m a p ( ) is the best permutation mapping function that matches the obtained clustering label to the equivalent label of the dataset. One of the best mapping functions is the Kuhn–Munkres algorithm [29]. The higher the ACC was, the better the clustering performance was.
Given two random variables X and Y, M I ( X ; Y ) is the mutual information of X and Y. H ( X ) and H ( Y ) are the entropies of P and Q, respectively. We use the normalized mutual information (NMI) as follows:
N M I ( X ; Y ) = M I ( X ; Y ) H ( X ) H ( Y ) .
The clustering results C ^ = { C ^ j } j = 1 k ^ and the ground truth classes C = { C j } j = 1 k are regarded as two discrete random variables. Therefore, NMI is specified as follows:
N M I ( C ; C ^ ) = i = 1 k ^ j = 1 k | C ^ i C j | log n | C ^ i C j | | C ^ i | | C j | ( i = 1 k ^ | C ^ i log | C ^ i | n | ) ( j = 1 k | C j | log | C j | n ) .
The higher the NMI was, the better the clustering performance was.
The Rand index is defined as,
R I = 2 ( a + d ) n ( n 1 ) ,
where a is the number of pairs of data points belonging to the same class in C and to the same cluster in C ^ . d is the number of pairs of data points belonging to the different class and to the different cluster. n is the number of data points. The larger the Rand index is, the better the clustering performance is.
We did a series of experiments to indicate the performance of the proposed method for data clustering. In the experiments, we set the parameters of all approaches in the same way to make the experiments fair enough, that is for parameter ρ , we set ρ = { 0.01 , 0.05 , 0.07 , 0.1 , 0.15 , 0.5 , 1 , 2 , 5 , 8 , 9 , 15 , 20 , 50 } . For α , we set α = { 0.1 , 0.2 , 0.3 , 0.4 , 0.5 , 0.6 , 0.7 , 0.8 , 0.9 } . For each parameter, we ran the given method 50 times and selected the best mean value to report. Table 2, Table 3 and Table 4 show the results with the different evaluation measures. In these tables, we use bold font to indicate the best performance.
We analyze the results from the dataset firstly. For IRIS dataset, the proposed method obtained the best performance for ACC, NMI and RI. For the CMC dataset, the proposed method had the best performance for ACC and RI. For the GLASS and BREAST datasets, the proposed method obtained the best performance for ACC and NMI. For the BALANCE dataset, the proposed method had the best performance for NMI and RI. On the other hand, from the three evaluation criteria, for ACC and NMI, the proposed method beat the other methods for four datasets. For RI, SVNC-TEM beat the other methods for three datasets. From the experimental results, we can see that the proposed method had better clustering performance than the other algorithms.

5. Conclusions

In the paper, we consider the truth membership degree, the falsity-membership degree, the indeterminacy membership degree and hesitate membership degree in a comprehensive way for data clustering by the single-valued neutrosophic set. We propose a novel data clustering algorithm, SVNC-TEM, and the experimental results showed that the proposed algorithm can be considered as a promising tool for data clustering and image processing. The proposed algorithm had better clustering performance than the other algorithms such as k-means, FCM, IFC and FC-PFS. Next, we will consider the proposed method to deal with outliers. Moreover, we will consider the clustering algorithm combined with spectral clustering and other clustering methods.

Author Contributions

All authors have contributed equally to this paper.

Funding

This research was funded by National Natural Science Foundation of China (Grant No. 11501435), Instructional Science and Technology Plan Projects of China National Textile and Apparel Council (2016073).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. MacQueen, J.B. Some methods for classification and analysis of multivariate observations. In Proceedings of the 5th Berkeley Symposium on Mathematical Statistics and Probability; University of California Press: Berkeley, CA, USA, 1967; pp. 281–297. [Google Scholar]
  2. Johnson, S.C. Hierarchical clustering schemes. Psychometrika 1967, 2, 241–254. [Google Scholar] [CrossRef]
  3. Ng, A.Y.; Jordan, M.I.; Weiss, Y. On spectral clustering: Analysis and an algorithm. Adv. Neural Inf. Process. Syst. 2002, 14, 849–856. [Google Scholar]
  4. Andenberg, M.R. Cluster Analysis for Applications; Academic Press: New York, NY, USA, 1973. [Google Scholar]
  5. Ménard, M.; Demko, C.; Loonis, P. The fuzzy c + 2 means: Solving the ambiguity rejection in clustering. Pattern Recognit. 2000, 33, 1219–1237. [Google Scholar]
  6. Aggarwal, C.C.; Reddy, C.K. Data Clustering: Algorithms and Applications; CRC Press: Boca Raton, FL, USA, 2013. [Google Scholar]
  7. Bezdek, J.C.; Ehrlich, R.; Full, W. FCM: The fuzzy c-means clustering algorithm. Comput. Geosci. 1984, 10, 191–203. [Google Scholar] [CrossRef]
  8. Chaira, T. A novel intuitionistic fuzzy C means clustering algorithm and its application to medical images. Appl. Soft Comput. 2011, 11, 1711–1717. [Google Scholar] [CrossRef]
  9. Hwang, C.; Rhee, F.C.H. Uncertain fuzzy clustering: interval type-2 fuzzy approach to c-means. IEEE Trans. Fuzzy Syst. 2007, 15, 107–120. [Google Scholar] [CrossRef]
  10. Thong, P.H.; Son, L.H. Picture fuzzy clustering: A new computational intelligence method. Soft Comput. 2016, 20, 3549–3562. [Google Scholar] [CrossRef]
  11. Graves, D.; Pedrycz, W. Kernel-based fuzzy clustering and fuzzy clustering: A comparative experimental study. Fuzzy Sets Syst. 2010, 161, 522–543. [Google Scholar] [CrossRef]
  12. Ramathilagam, S.; Devi, R.; Kannan, S.R. Extended fuzzy c-means: An analyzing data clustering problems. Clust. Comput. 2013, 16, 389–406. [Google Scholar] [CrossRef]
  13. Yasuda, M. Deterministic Annealing Approach to Fuzzy C-Means Clustering Based on EntropyMaximization. Adv. Fuzzy Syst. 2011, 2011, 960635. [Google Scholar] [CrossRef]
  14. Smarandache, F. A Unifying Field in Logics. Neutrosophy: Neutrosophic Probability, Set and Logic; American Research Press: Rehoboth, TX, USA, 1998. [Google Scholar]
  15. Wang, H.; Smarandache, F.; Zhang, Y.Q. Sunderraman, R. Single Valued Neutrosophic Sets. Multispace Multistruct 2010, 4, 410–413. [Google Scholar]
  16. Majumdar, P.; Samant, S.K. On similarity and entropy of neutrosophic sets. J. Intell. Fuzzy Syst. 2014, 26, 1245–1252. [Google Scholar]
  17. Ye, J. Improved correlation coefficients of single valued neutrosophic sets and interval neutrosophic sets for multiple attribute decision making. J. Intell. Fuzzy Syst. 2014, 27, 2453–2462. [Google Scholar]
  18. Zhang, X.; Bo, C.; Smarandache, F.; Dai, J. New Inclusion Relation of Neutrosophic Sets with Applications and Related Lattice Structure. J. Mach. Learn. Cybern. 2018, 2018, 1–11. [Google Scholar] [CrossRef]
  19. Zhang, X.H.; Smarandache, F.; Liang, X.L. Neutrosophic duplet semi-group and cancelable neutrosophic triplet groups. Symmetry 2017, 9, 275. [Google Scholar] [CrossRef]
  20. Ye, J. Single-Valued Neutrosophic Minimum Spanning Tree and Its Clustering Method. J. Intell. Syst. 2014, 23, 311–324. [Google Scholar] [CrossRef]
  21. Ye, J. Clustering Methods Using Distance-Based Similarity Measures of Single-Valued Neutrosophic Sets. J. Intell. Syst. 2014, 23, 379–389. [Google Scholar] [CrossRef]
  22. Guo, Y.; Sengur, A. NCM: Neutrosophic c-means clustering algorithm. Pattern Recognit. 2015, 48, 2710–2724. [Google Scholar] [CrossRef]
  23. Thong, P.H.; Son, L.H. A novel automatic picture fuzzy clustering method based on particle swarm optimization and picture composite cardinality. Knowl. Based Syst. 2016, 109, 48–60. [Google Scholar] [CrossRef]
  24. Thong, P.H.; Son, L.H. Picturefuzzyclusteringforcomplexdata. Eng. Appl. Artif. Intell. 2016, 56, 121–130. [Google Scholar] [CrossRef]
  25. Son, L.H. DPFCM: A novel distributed picture fuzzy clustering method on picture fuzzy sets. Expert Syst. Appl. 2015, 42, 51–66. [Google Scholar] [CrossRef]
  26. Cuong, B.C. Picture Fuzzy Sets. J. Comput. Sci. Cybern. 2014, 30, 409–420. [Google Scholar]
  27. Cuong, B.C.; Phong, P.H.; Smarandache, F. Standard Neutrosophic Soft Theory: Some First Results. Neutrosophic Sets Syst. 2016, 12, 80–91. [Google Scholar]
  28. Tsallis, C. Possible generalization of Boltzmann-Gibbs statistics. J. Stat. Phys. 1988, 52, 479–487. [Google Scholar] [CrossRef]
  29. Lovasz, L.; Plummer, M.D. Matching Theory; AMS Chelsea Publishing: Providence, RI, USA, 2009. [Google Scholar]
Figure 1. The demonstration figure of the clustering process for artificial data. (a) The original data. (be) The clustering figures when the number of iterations t = 1, 5, 10, 20, respectively. (f) The final clustering result.
Figure 1. The demonstration figure of the clustering process for artificial data. (a) The original data. (be) The clustering figures when the number of iterations t = 1, 5, 10, 20, respectively. (f) The final clustering result.
Axioms 07 00057 g001
Figure 2. The image segmentation for the Lena image. (a) The original Lena image. (bf) The clustering images when the number of clusters k = 2, 5, 8, 11 and 20, respectively.
Figure 2. The image segmentation for the Lena image. (a) The original Lena image. (bf) The clustering images when the number of clusters k = 2, 5, 8, 11 and 20, respectively.
Axioms 07 00057 g002
Table 1. Description of the experimental datasets.
Table 1. Description of the experimental datasets.
DatasetNo. of ElementsNo. of AttributesNo. of ClassesElements in Each Classes
IRIS15043[50, 50, 50]
CMC147393[629, 333, 511]
GLASS21496[29, 76, 70, 17, 13, 9]
BALANCE62543[49, 288, 288]
BREAST27792[81, 196]
Table 2. The ACC for different algorithms on different datasets.
Table 2. The ACC for different algorithms on different datasets.
Datasetk-MeansFCMIFCFC-PFSSVNC-TEM
IRIS0.88030.89330.90000.89330.9000
CMC0.39650.39170.39580.39170.3985
GLASS0.32190.25700.36360.29350.3681
BALANCE0.53000.52600.54130.52060.5149
BREAST0.66760.57650.65950.65850.6686
Bold format: the best performance.
Table 3. The NMI for different algorithms on different datasets.
Table 3. The NMI for different algorithms on different datasets.
Datasetk-MeansFCMIFCFC-PFSSVNC-TEM
IRIS0.75140.74960.71020.75010.7578
CMC0.03200.03300.03220.03340.0266
GLASS0.04880.03870.06730.04190.0682
BALANCE0.13560.13360.12320.12130.1437
BREAST0.06230.03090.02850.06100.0797
Bold format: the best performance.
Table 4. The RI for different algorithms on different datasets.
Table 4. The RI for different algorithms on different datasets.
Datasetk-MeansFCMIFCFC-PFSSVNC-TEM
IRIS0.87330.87970.88270.87970.8859
CMC0.55760.55820.55890.55820.5605
GLASS0.53730.62940.46170.58740.4590
BALANCE0.59400.59280.58990.59040.5999
BREAST0.57080.51590.57320.56560.5567
Bold format: the best performance.

Share and Cite

MDPI and ACS Style

Li, Q.; Ma, Y.; Smarandache, F.; Zhu, S. Single-Valued Neutrosophic Clustering Algorithm Based on Tsallis Entropy Maximization. Axioms 2018, 7, 57. https://doi.org/10.3390/axioms7030057

AMA Style

Li Q, Ma Y, Smarandache F, Zhu S. Single-Valued Neutrosophic Clustering Algorithm Based on Tsallis Entropy Maximization. Axioms. 2018; 7(3):57. https://doi.org/10.3390/axioms7030057

Chicago/Turabian Style

Li, Qiaoyan, Yingcang Ma, Florentin Smarandache, and Shuangwu Zhu. 2018. "Single-Valued Neutrosophic Clustering Algorithm Based on Tsallis Entropy Maximization" Axioms 7, no. 3: 57. https://doi.org/10.3390/axioms7030057

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop