Next Article in Journal
Learned Design of a Compressive Hyperspectral Imager for Remote Sensing by a Physics-Constrained Autoencoder
Previous Article in Journal
A Novel Double Ensemble Algorithm for the Classification of Multi-Class Imbalanced Hyperspectral Data
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Efficient Supervised Image Clustering Based on Density Division and Graph Neural Networks

1
College of Computer Science and Technology, Harbin Engineering University, Harbin 150001, China
2
ICT Cluster, Singapore Institute of Technology, Singapore 138683, Singapore
3
S R Nathan School of Human Development, Singapore University of Social Sciences, Singapore 599494, Singapore
*
Author to whom correspondence should be addressed.
Remote Sens. 2022, 14(15), 3768; https://doi.org/10.3390/rs14153768
Submission received: 15 July 2022 / Accepted: 24 July 2022 / Published: 5 August 2022
(This article belongs to the Topic Big Data and Artificial Intelligence)

Abstract

:
In recent research, supervised image clustering based on Graph Neural Networks (GNN) connectivity prediction has demonstrated considerable improvements over traditional clustering algorithms. However, existing supervised image clustering algorithms are usually time-consuming and limit their applications. In order to infer the connectivity between image instances, they usually created a subgraph for each image instance. Due to the creation and process of a large number of subgraphs as the input of GNN, the computation overheads are enormous. To address the high computation overhead problem in the GNN connectivity prediction, we present a time-efficient and effective GNN-based supervised clustering framework based on density division namely DDC-GNN. DDC-GNN divides all image instances into high-density parts and low-density parts, and only performs GNN subgraph connectivity prediction on the low-density parts, resulting in a significant reduction in redundant calculations. We test two typical models in the GNN connectivity prediction module in the DDC-GNN framework, which are the graph convolutional networks (GCN)-based model and the graph auto-encoder (GAE)-based model. Meanwhile, adaptive subgraphs are generated to ensure sufficient contextual information extraction for low-density parts instead of the fixed-size subgraphs. According to the experiments on different datasets, DDC-GNN achieves higher accuracy and is almost five times quicker than those without the density division strategy.

1. Introduction

Due to the prevalence of cameras, a lot of photographs are created every day in the modern world (e.g., smartphones, monitor systems, etc.). Unfortunately, this has created significant difficulties for image management systems [1,2,3,4]. For instance, to better manage face images, face clustering has emerged as a highly popular method to handle face photos for tasks such as face recognition, face labeling, etc. [5,6,7,8].
Conventional clustering techniques (e.g., K-means [9], DBSCAN [10], HAC [11], Spectral Clustering [12], etc.) rely on certain assumptions and are hard to handle face images with complicated data distributions. For example, K-Means requires the datasets to be convex-shaped, and DBSCAN requires different clusters with similar densities. Supervised clustering techniques have been developed to learn cluster patterns based on Graph Neural Networks (GNNs) link prediction [13] and have demonstrated significant improvements over conventional approaches in terms of accuracy [14,15]. These link prediction-based algorithms can learn the real similarity between instances and have less strict constraints for the data distribution which can enhance the effectiveness of face clustering. These kinds of methods share similar operations. The first is the construction of subgraphs for each instance as the central node first, followed by the prediction of the linkage probability between the central node and its neighbors. In the end, all instances are grouped into clusters according to the predicted connectivity. However, these methods have produced large computational overheads despite being sufficient to identify more precise clusters. The main computation of this type of method is a large number of subgraphs generations for each image instance and inferring the similarity between the instance and its neighbor. A basic observation is that a central node can be a neighbor node in other subgraphs and the produced subgraphs are usually highly overlapped. This increases the cost of duplicated calculations excessively and slows down the clustering speed.
To resolve the shortcomings in supervised clustering, we use density division clustering to improve the supervised clustering and propose the GNN-based supervised clustering framework DDC-GNN, which is both time-efficient and effective. In the first step, density division is performed to divide the image instances (or referred to as nodes) into high-density parts and low-density parts based on their density. The high-density parts are the data instances which locate at the cluster center. Then a subgraph is generated for each node in low-density parts, and a GNN model is used to perform link prediction for each subgraph. Finally, the pseudo-label propagation approach merges the nodes in low-density parts and high-density parts by a predicted linkage similarity. After several iterations in which the unqualified edges are deleted, the final clusters are formed.
A preliminary version of this paper has been published in [16]. In this paper, we have incorporated numerous extensions, including a general framework of GNN-based supervised clustering based on density division, evaluating two different types of GNN models (i.e., the GCN-based model and the GAE-based model) under the framework, a more comprehensive experiment evaluation and analysis together with numerous expansion information thorough the paper. In particular, we leverage two different types of GNN models in the connectivity learning and prediction phase, which are the GCN-based model and the GAE-based model. The difference between the two models is obvious. The GCN-based model deals with a node classification task to neighbor nodes. A neighbor node may have the same label as the center node or have a different label from the center node. The encoding manner of the center node feature in GCN is to build the edge between the center node and the neighbor node, which means that the feature of the center node is encoded in the GCN model by message passing/propagation. While the GAE-based model is a link prediction/classification task. A linkage probability between the center node and neighbor node will be predicted to be from 0 to 1, which is the output of the decoders in the GAE-based model. It is the multiplication of the hidden center node feature and the hidden neighbor node feature. The encoding manner of the center node feature in the GAE-based model is the inner product of the hidden feature.
The contributions of our paper are summarized as follows:
  • We propose a novel time-efficient and effective GNN-based supervised clustering framework based on density division DDC-GNN. The main phrases are the density division phase, the GNN model inferring phase, and the cluster requiring phase. In GNN model inferring phase, any Plug-and-Play model for image similarity inferring can be applied. In this paper, we have tested GCN-based and GAE-based models.
  • To reduce the high computation overhead in existing supervised image clustering algorithms, we present a novel density division method to split the face image instances into high-density parts and low-density parts, where only the low-density parts of the nodes need to perform the connectivity prediction. The density division strategy drastically reduces the computing overhead and unnecessary inferring.
  • Extensive experiments are implemented to compare the DDC-GNN framework with other approaches using the MS-Celeb-1M and IJB-B face datasets. The results demonstrate that DDC-GNN is five times quicker than existing approaches with comparable accuracy in a fair comparison.
This paper is organized as follows. Section 2 introduces the related work about the graph neural network and face clustering. Section 3 shows the details of the proposed DDC-GNN framework. Section 4, Section 5 and Section 6 provide the experimental evaluations, discussion, conclusion and future work respectively.

2. Related Work

2.1. Graph Neural Networks

In general, convolutional neural networks (eg. Mask R-CNN [17], etc.) cannot process non-grid spatial structure data while graph convolutional networks can solve this problem perfectly. Andriyanov et al. were to develop an algorithm that could help increase segmentation accuracy when used together with the known algorithms [18]. In semi-supervised node classification tasks, Kipf et al. [19] proposed the graph convolutional networks, which significantly enhanced the performance of node classification. Schlichtkrull et al. [20] solved link prediction and entity classification in knowledge graph mapping tasks by graph convolutional networks. Liu et al. [21] applied graph convolutional networks to capture the global information to solve text classification. Graph auto-encoder also used GCN layers to predict linkage [22]. GraphSAGE extended the graph neural network to inductive learning by fusing the neighbor nodes of the current node to obtain the feature representation [23]. Each neighbor node is treated equally in GraphSAGE. But in actual scenarios, different neighbor nodes may play different roles. Attention was a good mechanism and widely used [24]. The Graph Attention Network (GAT) was introduced to aggregate the neighbor nodes through the self-attention mechanic, which realized the adaptive matching of the weights of different neighbors, thereby improving the accuracy of the model [25]. In this paper, we leverage the powerful information aggregation of the graph neural network to learn the potential relationship between nodes and infer connectivity between nodes.

2.2. Unsupervised Face Clustering

Face clustering is a typical task that organizes images of faces into groups according to their identities, ensuring that faces in one group are all from the same identity and faces in separate groups have different identities. It has received much attention as a basic data mining learning task. The conventional clustering techniques, such as K-means [9], DBSCAN [10], mean shift [26], rely on certain assumptions, making it challenging to cluster complicated face images because of high dimension features, face position, light, and other factors. To organize face image instances into clusters, Lin et al. [27] presented a hierarchical clustering algorithm based on linear support vectors. In order to handle data with an uneven distribution of density, a density-aware cluster-level affinity measure that used singular value decomposition was also created [28]. Zhu et al. [29] converted the face clustering problem into a multi-image cutting problem by the gradient flow method. For face clustering, Tapasw et al. [30] suggested a ball clustering learning method that could automatically count the number of clusters in the video.

2.3. Supervised Face Clustering

Supervised learning has been used for face clustering in a recent study. Zhan et al. [31] tried to bring input data into several models for each pair of nodes and including numerous commentator networks and then trained a multi-layer perception classifier to gather the information. Wang et al. [2] combined the subgraphs information using the GCN’s aggregation capabilities, which significantly enhanced the prediction of inter-node linkages. Yang et al. [14] developed an approach to produce multi-scale subgraphs as input for a graph convolutional neural network to detect cluster proposals. Zhao et al. [32] analyzed data distribution with supervised learning and used an effective framework to cluster new data in an incremental context. Qi et al. [33] used ResNet-GCNs to discover the connectivity between nodes. Yang et al. [15] employed a Reverse-Imbalance Weighted Sampling (RIWS) technique to improve graph representations in unbalanced labeled environments.
Despite the fact that supervised face clustering has been shown to be more efficient, it has a significant computational cost because they all need to produce a large number of subgraphs for connection learning. Moreover, these subgraphs are usually overlapped. For instance, L-GCN [2] needs a lot of subgraphs to detect the cluster pattern (linkage between nodes) properly. There are many neighbors between each subgraph, especially close to the cluster center, which are overlapping. However, our proposed framework focuses on predicting the connection of the low-density part after dividing face images by the density. Since the face image instances from the high-density parts are naturally connected in the cluster center, fewer computing resources are required and the clustering proceeds more quickly and accurately.

3. Supervised Clustering Based on Density Division

3.1. Overview

To better appreciate the proposed solution, we first provide a framework overview of DDC-GNN. The DDC-GNN framework overview is presented in Figure 1, which consists of three main stages.
The first stage of the framework is the node division by density. In general supervised clustering algorithms, most of the approaches need to predict the connectivity/linkage of edges between all the nodes. Unlike the previous work, the DDC-GNN framework does not need to predict the linkage for the entire face image sets. Instead, DDC-GNN divides the entire face image sets into high-density parts and low-density parts. The nodes in the high-density parts are close to the center of the cluster, which tends to be in the same cluster with strong connectivity and do not need to make predictions.
In the second stage of the framework, for each node from the low-density parts, a customized adaptive subgraph (ASG) is built and learned by a GNN model (GCN or GAE model) for linkage/connectivity prediction. These ASGs can be used as an input by the GNN model to predict connectivity that reflects real image instance similarity.
The third stage of the framework is linkage merging to form the clusters. A pseudo label propagation approach organizes the nodes from high-density and low-density parts into clusters.

3.2. Notations and Descriptions

The notations we used are shown in Table 1. The formal expression of our image clustering task is C = D D C - G N N ( V ) , where C is the cluster results and V is the node-set with extracted feature F from a pre-trained model.

3.3. Connected Region Division

As the existing linkage-based face clustering approaches need to form subgraphs for each face (node), there are a large number of subgraphs generated that result in high computation overhead. To reduce the overhead, our strategy is to divide the nodes into high-density parts and low-density parts. The high-density parts refer to the nodes which are strongly connected with each other and likely in the same cluster in the end. Otherwise, the nodes belong to the low-density parts which need to predict their connectivities using the GNN.
To do so, we first calculate the distance between the nodes. The distance can be easily calculated based on the existing measures like cosine similarity and so on. Then, we calculate the density of each node based on their distances to other nodes. We define the density of node v as the number of nodes that are within E distance, where E is a predefined distance threshold for density calculation. After calculating the distance, we divide the nodes into high-density parts and low-density parts. For instance, we can define the proportion ρ of entire nodes as the high-density nodes, where ρ can be obtained from the training dataset. Next, among all the high-density nodes, we further form them into high-density components by linking those high-density nodes. For any given high-density nodes m and n, if m is one of the M nearest nodes of n, we link m and n with one edge, where M is another parameter obtained from the dataset. In other words, if the distance between m and n is smaller than other M nodes with n, m will not be linked to n. This is to make sure that the high-density nodes linked together are strongly connected to each other. Thus, they can be safely linked together as a strongly connected component. Those nodes in a strongly connected component are likely in the same cluster. Thus, the edge connectivity weights between these nodes in the same high-density component can be set as 1, as they will be in the same cluster.
By so doing, the high-density nodes are grouped into various strongly connected components with assigned weights.
For example, we set E = 0.5 , ρ = 50 % , M = 2 and there are four nodes. The distances between nodes are shown in the following matrix.
n o d e 1 n o d e 2 n o d e 3 n o d e 4 n o d e 1 0 0.1 0.3 0.5 n o d e 2 0.1 0 0.8 0.4 n o d e 3 0.3 0.8 0 0.6 n o d e 4 0.5 0.4 0.6 0
The densities of four nodes are 2,2,1,1 and d = { 2 , 2 , 1 , 1 } . When we choose node 1 and node 2 are two high-density nodes and node 3 and node 4 are low-density nodes.

3.4. Node Similarity Prediction of Adaptive Subgraph in GNN

The calculation of similarity/connectivity/linkage between nodes is usually based on non-learnable Euclidean distance or non-learnable cosine similarity. In recent works, supervised similarity learning has been applied to GCN and obtains promising results, such as the node classification model in L-GCN [2] and GCN-VE. Our paper summarizes the mainstream methods into a unified connectivity prediction framework, namely: 1. Locate nodes of subgraphs; 2. Build edges of subgraphs; 3. Apply a GNN model on ASGs. Under the connectivity prediction framework, we generate adaptive subgraphs instead of previously fixed subgraphs and design two types of models (i.e., node classification model and link prediction model). Based on the method, when a GCN node classification model is used for connectivity prediction in the framework, it is named DDC-GCN. When a GAE link prediction model is used for connectivity prediction, we call it DDC-GAE.

3.4.1. Node Classification Model for Connectivity Prediction

According to the neighborhood context information of two face images (nodes), their connectivity can be learned by ASGs, which are constructed to extract the real similarity information of neighboring nodes. The ASGs generation is shown in Figure 2.
The first phase is locating nodes in a certain ASG. For a given node as the central node p, we locate second-hop neighbor nodes of p to form an ASG. We denote ASGs as G ( V s , E s ) . For each hop of neighbor nodes, we construct a subgraph by locating h 1 one-hop direct neighbors and h 2 two-hop indirect neighbors of each central node. h 1 one-hop neighbors are the h 1 nearest neighbors of p and h 2 two-hop neighbors are the h 2 nearest neighbors of one-hop neighbors.
The second phase is building edges for all nodes in a certain ASG. We calculate an adaptive average distance d i s , iteratively obtain the maximum Euclidean distance between each direct neighbor (one-hop) and its h 2 indirect neighbors (two-hop), and then calculate the average of these maximum values to obtain the final adaptive distance. For a node q V s , we look for potential edges in the k nearest neighbors of q. If a node r appears in V s and the distance between q and r is less than d i s , the edge (q, r) is added to the set E s . This process can ensure that sufficient structure of the subgraph. Then we get the subgraphs with central node p, the node-set V s , and the feature set F p .
F p = [ , f s , ] T , s V s
Finally, we finish the construction of the adaptive subgraph, obtaining the characteristic matrix F p and adjacency matrix A s R ( | V s | x | V s | ) .
The context information in ASGs can be used to predict the connectivities of low-density data instances. A four-layer GCN model is used to perform node classification tasks on ASGs which reflects the similarities between nodes and the formal expression of each layer is provided as follows.
F ( l + 1 ) = σ ( [ F l G F l ] W )
where F l R ( N × D i n ) , F ( L + 1 ) R ( N × D o u t ) , N is the number of nodes, D i n is the input dimension of the node feature, D o u t is the output dimension of the node feature, F l is the output characteristic of the l layer, W is a learnable weight matrix for the convolution layer, which has a size of 2 d i n × d o u t . G = Λ ( 1 / 2 ) A Λ ( 1 / 2 ) is an aggregation matrix of size N × N , each row of which adds up to 1, where A is the adjacency matrix and Λ is a diagonal matrix. The diagonal element is Λ I I = Σ j A i j . The operator ‖ means that the matrix is joined along with the feature dimension. σ ( · ) is a nonlinear activation function. There are two steps in the graph convolution process. In the first step, the potential neighbor information of the node is gathered by the operation that F l multiply G, and then the input node features F l and G · F l are concatenated along the feature dimension. In the second step, the feature matrix is transformed into the input of the next layer F ( L + 1 ) through a set of learnable parameters W.
As shown in Figure 3, A is the adjacency matrix and X is the feature matrix. The node classification model GCN of DDC-GNN is a stack of four convolutional layers of graphs, and the R e L U nonlinear activation function of each GCN layer is used for activation. To predict the existence of edges between nodes in subgraphs, we use a two-layer MLP classifier to minimize the cross-entropy loss between the predicted connectivity and the ground-truth edge label. We use a pair of node features corresponding to each edge in the subgraph as the input of MLP to obtain the two-dimensional connectivity.
In the training phase, for a neighbor node that is in the same cluster as the central node, the connectivity between this neighbor and the central node is set to 1 (with the same label) and the connectivity is set to 0 when the neighbor node is in a different cluster with the central node (with a different label). For the inferring phase, the output of GCN is from 0 to 1 built by the softmax function. The architecture of the node classification model in Table 2.
The cross-entropy loss function is defined as the node-wise cross-entropy loss overall labeled instances.
c r o s s - e n t r o p y   l o s s = v V C d i m = 1 C d i m L ( v ) [ f ] · log F h i d d e n [ f ]
V is a set of nodes in the graph, and v is a node belonging to V. C d i m is the output dimension, which is equal to the number of classes. There are positive instances and negative ones in our node classification model. Positive instances are the neighbor nodes with the same label as certain center nodes and negative instances are the neighbor nodes with different labels, so C d i m is 2 in the node classification model. Given an input training set in the model, we can set a two-dimensional vector label by the label indicator matrix L ( v ) .
L ( v ) = ( 1 , 0 ) , l v = l c e n t e r   n o d e ( 0 , 1 ) , l v l c e n t e r   n o d e
When the label of a neighbor node is in the same class as the center node, the label vector is (1, 0). While a neighbor node is in a different class from the center node, the label vector is (0, 1).
After we divide the data set, we construct an adaptive subgraph for each node in the low-density nodes and predict the possibility of a connection between each node in the subgraph. Therefore, we get a set of edges with the predicted connectivity which indicates the similarity between a pair of nodes.

3.4.2. Link Prediction Model for Connectivity Prediction

The node classification task predicts the probability that the neighbor node shares the same class label with the central node. Different from the node classification model, the link prediction model is to predict the edge connectivity probability between two nodes in a network. Graph auto-encoder is a typical model which is used to perform link prediction tasks. The ASGs generation in our designed link prediction model is the same as the node classification model, while the supervised information is different. In a node classification task, a one-hop neighbor node has a label with 0 or 1. Differently, an edge is assigned a label with 1. If two nodes (the central node and its one-hop neighbor node) have no edge, the position of the adjacent matrix is set to be 0.
As the construction of the link prediction model is shown in Figure 4, the GAE model has an encoder and decoder structure. The encoder has n GCN-layers and the decoder is a matrix multiplied to reconstruct the adjacent matrix.
The formal expression of the encoder is as follows.
Z = G C N l a y e r ( X , A )
The formal expression of the decoder is:
A ^ = σ ( Z Z T )
where G C N l a y e r ( X , A ) = A ˜ R e L U ( A ˜ X W 0 ) W 1 and A ˜ = Λ ( 1 / 2 ) A Λ ( 1 / 2 ) . Λ is a diagonal matrix. Z is the hidden variable of the hidden layer and X is the feature matrix and A is the adjacent matrix. A ^ is the reconstruction adjacent matrix. We minimize the cross-entropy reconstruction loss of A ^ and A. The architecture of the link prediction model is shown in Table 3.
In the training phase of the link prediction model GAE, as a neighbor node is in the same cluster with the central node, the connectivity between this neighbor and the central node is set to 1 (with the same label) and the connectivity is set to 0 when the neighbor node is in a different cluster with the central node (with a different label).
The loss function is defined as the node-wise cross-entropy loss overall labeled instances.
c r o s s - e n t r o p y   l o s s = i , j V C d i m = 1 C d i m L ( i , j ) [ f ] · log A i , j ^ [ f ]
where ( i , j ) is the set of node pairs in the graph belonging to V. C d i m is the output dimension of the linkage, which is equal to the number of classes. There are two classes in our graph auto-encoder model, which is whether there should be a linkage between the node pairs, so C d i m is 2 in the GAE model. We can set a two-dimensional vector label by the label indicator matrix L ( v ) when we get the input training set into the model.
L ( v ) = ( 1 , 0 ) , c o n n e c t i v i t y = 1 ( 0 , 1 ) , c o n n e c t i v i t y = 0
When there should be a linkage and the connectivity is 1, the label vector in the indicator matrix is (1, 0). While connectivity between two nodes is 0, the label vector in the indicator matrix is (0, 1).

3.5. Linkage Merging

To obtain clusters, a simple method is to set a threshold, cut out all edges with less weight than the threshold, and propagate the pseudo-labels using Breadth First Search. However, this method is sensitive to the threshold value. Therefore, we adopt the pseudo-label propagation strategy proposed in [31]. In each iteration, edges with weights less than the threshold are cut off, and the remaining edge instances are connected to form clusters. If a cluster in these clusters is larger than the predefined maximum size of the cluster, it will be added to the queue for processing in the next iteration. The threshold which is used to cut edges increases gradually during iteration, and then stops iterating when the queue is empty, which means the clustering is complete.

3.6. Efficiency Analysis

Three components of the algorithm-density division, GNN connectivity prediction, and link merging-contribute mostly to its complexity. Approximate Nearest Neighbor Search ( O ( n l o g n ) ) is more efficient than the brutal search ( O ( n × n ) ) in neighbor node locating. The density value is calculated linearly in the density division. The complexity of the density division is O ( n × M ) because M nearest neighbor nodes of each node with high density must be found to create the edge in the high-density parts. The creation of ASGs in the low-density parts is the main component of connectivity prediction. The complexity of O ( n l o g n ) can be much lower than the ANN search in the ASG creation process. Link merging requires fewer computations and a lower complexity O ( n × h 1 ) . Compared with n, M and h 1 are two small number. As a result, the time complexity of DDC-overall GNN is O ( n l o g n ) . The DDC-GNN separates the data into two portions with differing densities and only applies GCN or GAE inference to the low-density parts. Because the computation required for ASGs inferring reduces greatly, clustering of the DDC-GNN framework is more effective.
Algorithm 1 is the proposed supervised clustering based on the density division algorithm.
Algorithm 1: supervised clustering based on density division
Input: the number of neighbors M, feature set F, distance threshold E, descending density set of all nodes d, m o d e l _ t y p e = G A E / G C N
Output: Clusters C
1:
procedure  DDC-GNN
2:
    Edge, D l = Density Partition( E , M , d )
3:
    Edge = Edge ∪ Connectivity Prediction(F, D l , m o d e l _ t y p e )
4:
    C = Use pseudo-label propagation strategy proposed in [31] to find connected graphs on Edge
5:
     return C
6:
    
7:
    function Density Partition( E , M , d )
8:
        Divide d into two sets according to the Ratio: high-density D h and low-density D l
9:
        for  i = 1 , 2 , 3 , n in D h  do
10:
           Find M nearest neighbors v i N ( v i )
11:
           if dis( v i , v j ) <E and v j N ( v i )  then
12:
               Edge = Edge ∪ Edge{ v i , v j }
13:
        return Edge, D l
14:
    
15:
    function Similarity Prediction(F, D l , m o d e l _ t y p e )
16:
        if  m o d e l t y p e = = G A E  then
27:
           Connectivity=GAE(F, D l )
18:
        if  m o d e l t y p e = = G C N  then
19:
           Connectivity= GCN(F, D l )
20:
        return Connectivity

4. Experiments and Evaluations

4.1. Experiments Setting

4.1.1. Dataset

DDC-GNN is evaluated by the publicly available benchmarks for large-scale face recognition, the MS-Celeb-1M [34] and IJB-B [35]. In total, there are 100k identities that constitute MS-Celeb-1M. The number of images varies from 10 to 600 in one identity in the CT images [36] and the remote sensing images [37]. In order to divide them into a training set and a testing set, we apply the same parameters as in [14]. The numbers of identities included in the three subtasks are 512, 1024, and 1845, respectively. And the numbers of samples included are 18,171, 36,575, and 68,195, respectively. We use the model trained on a random subset of CASIA [38] datasets for testing clustering.

4.1.2. Metrics

In order to evaluate the performance of the proposed face clustering method in a comparable way, we use the most common evaluation metrics in image clustering task in the same experimental environment: B-cubed F-score (denoted as F), Normalized Mutual Information (denoted as NMI) [39], Adjusted Rand Index (denoted as ARI) and Silhouette Coefficient (denoted as SC).
NMI: Given the true label Ω of a cluster, C is the predicted cluster label, and the definition of NMI is as follows:
N M I ( Ω , C ) = I ( Ω , C ) H ( Ω ) H ( C )
I ( Ω , C ) = H ( Ω ) + H ( C ) H ( Ω , C )
H ( Ω ) = ω P Ω ( ω ) log P Ω ( ω )
H ( C ) = c P C ( c ) log P C ( c )
H ( Ω , C ) = Ω , C P Ω , C ( ω , c ) log P Ω , C ( ω , c )
Among them, H(·) is the entropy function, respectively, and I( Ω ,C) represents mutual information. P Ω ( ω ) and P C ( c ) are the distribution of Ω and C. P Ω , C ( ω , c ) is the joint distribution of Ω and C.
F-score: F-score (denoted as F in the following paper) is the harmonic average of precision and recall. C ( i ) and L ( i ) are to represent the cluster label and true label of node i, respectively. And we define the accuracy of two nodes i and j as follows.
Correctness ( i , j ) = 1 , if L ( i ) = L ( j ) and C ( i ) = C ( j ) 0 , if otherwise
The definitions of precision and recall are as follows.
Precision = 1 N i = 1 N j C ( i ) Correctness ( i , j ) | C ( i ) |
Recall = 1 N i = 1 N j L ( i ) Correctness ( i , j ) | L ( i ) |
where | C ( i ) | and | L ( i ) | represent the size of sets C ( i ) and L ( i ) , respectively. The definition of F- s c o r e is as follows.
B C u b e d _ F s c o r e = 2 × Precision × Recall Precision + Recall
ARI: ARI measures the coincidence degree of two data distributions. The higher the value is, the more the clustering result matches the real situation. A R I 1 , 1 , and it can be expressed as:
A R I = R I E ( R I ) m a x ( R I ) E ( R I )
where R I is Rand Index which:
R I = a + d C N 2
C N 2 is the data pair in all data and N is the total sample number. a is the logarithm of data pairs in the same category in true clusters and also in the same category in predicted clusters at the same time. d is the logarithm of data pairs not in the same category in true clusters and not in the same category in predicted clusters.
SC:  s ( i ) of a sample i can be expressed as follows:
s ( i ) = 1 a ( i ) b ( i ) , a ( i ) < b ( i ) 0 , a ( i ) = b ( i ) a ( i ) b ( i ) 1 , a ( i ) > b ( i )
where a ( i ) is the average distance from sample i to other samples in the same cluster and b ( i ) is the average distance from sample i to all samples of other clusters. Silhouette coefficient of sample i is defined according to the intra-cluster dissimilarity a ( i ) and inter-cluster dissimilarity b ( i ) of sample i. The silhouette coefficient of the predicted cluster is the average silhouette coefficient of all samples.

4.1.3. Parameter Setting

We investigate several values for the parameters M and E and discover that the clustering effect is best when M is set to the number around 0.1 percent of the data volume and the distance E for cosine distance is set to 0.8. In addition, several experiments lead to the density selection threshold being at 0.4. The weight decay value is 1 × 10 5 and the initial learning rate is 0.1.

4.1.4. Benchmarking

Our method is compared with these clustering methods, which are described briefly as follows.
  • K-Means [9]: K-Means divides a data set into several clusters according to how far between data instances. The cluster’s sample distances are reduced as much as possible via K-Means.
  • Affinity Propagation (AP) [40]: All data points should be treated as potential clustering centers, and the information transmission between the data instances is taken into consideration during clustering, which is appropriate for high-dimensional, multi-class data clustering algorithms.
  • DBSCAN [10]: By using appropriate density criteria, DBSCAN maintains the sparse background as noise.
  • Approximate Rank-Order Clustering (ARO) [41]: ARO clusters data by an enhanced distance metric and approximative closest neighbor.
  • Hierarchical Agglomerative Clustering (HAC) [42]: HAC uses a bottom-up approach, combines clusters, and stops clustering under predetermined rules.
  • Spectral Clustering [12]: Spectral Clustering separates the data into several related components based on the similarity matrix between the data.
  • Deep Density Clustering (DDC) [28]: DDC is the clustering of local neighborhoods based on their similarity in terms of density in the feature space.
  • Consensus-driven Propagation Clustering (CDP) [31]: CDP is a learnable clustering algorithm that uses committee networks to improve robustness.
  • L-GCN [2]: L-GCN is a learnable clustering technique that makes use of GCN to extract contextual data from the network for linkage prediction.
  • Non-density division-GCN Clustering (NDD-GCN): A method that constructs an adaptive graph for all nodes as context without density division parts, then applies GCN for reasoning on it.
  • Density Division Clustering based on GAE (DDC-GAE): A method that constructs an adaptive graph for all nodes as context with density division parts, then applies GAE for reasoning on it.

4.2. Experimental Results

4.2.1. Results and Performance Analysis

The comparative results between the DDC-GNN framework and other methods are shown in Table 4. The best results are highlighted in bold, while a “-” in the “Time” column indicates that the accuracy and time are not given in the original publications. We reproduce the experiments but lead to an “out of memory” result. The results demonstrate that our DDC-GNN framework outperforms other clustering approaches on IJB-B.
Since the predetermined number of clusters has a significant impact on K-Means performance, it might be challenging to achieve satisfactory results when the number of clusters is uncertain. Dealing with imbalanced data is out of the capability of spectral clustering and the data distribution has a significant impact on DBSCAN since it assumes that the density of various clusters is identical. Although HAC does not require a predetermined certain number of clusters, the iterative clustering process is time-consuming. The DDC-GNN framework performs better than ARO and DDC because the learnable connection is more accurate than unlearnable cosine similarity.
In the supervised methods comparison of the IJB-B datasets, the DDC-GNN framework outperforms other approaches. It is worth focusing that the results on IJB-B-512 are not the SOTA on the NMI metric, possibly because the amount of data is too small and the high-density and low-density parts cannot be separated well by tuning parameters E and M. We also keep track of how long DDC-GAE/GCN and L-GCN take. L-GCN is less efficient than DDC-GAE and DDC-GNN because we have shortened the time required for a portion of the GNN subgraphs inference.
As seen in the results of MS-Celeb-1M datasets in Table 5, the DDC-GNN framework outperforms conventional clustering methods in MS-Celeb-1M. It outperforms L-GCN by 2%, which performs better than CDP, but slower than CDP since L-GCN is required to extract context information for building and inferring subgraphs. It can be seen that utilizing adaptive subgraphs as input can better forecast the connectivity of edges since the approach with non-density division (NDD-GCN) is superior to L-GCN. Calculations of L-GCN will take a lot of time because subgraphs are highly overlapped. However, compared with other supervised approaches, DDC-GAE with density division achieves a computation time reduction and DDC-GCN achieves a large calculation time reduction with comparable (or even slightly higher) performance. This demonstrates the accuracy and time-efficiency of the DDC-GNN framework.

4.2.2. Ablation Study

In the framework of DDC-GNN, we divide the data into high-density and low-density parts in the linked region division stage. Our experimental results will be directly impacted by how to form the high-density components of the data. Important factors for density division are E and M. We also examine how E and M affect performance.
Because a bigger E makes the estimated density of nodes more discriminative, Figure 5 illustrates that F- s c o r e increases as E grows from 0.5 to 0.8, while F- s c o r e begins to decrease after 0.8. When E is small, that means nodes might have few nodes around them and they are discriminative to each other. When E is big, that means nodes might include many nodes from different clusters and this will affect the performance. Figure 6 illustrates that a greater M also results in a higher F- s c o r e from 15 to 18. The F- s c o r e states to decrease after 18. This is because a too big M value makes DDC-GNN create big strong connected components in the high-density parts, which may result in wrong clusters. When M decreases, DDC-GNN generates small strong connected components, which rely on the last clustering step to link them back.
When E and M rise to a certain amount, the performance achieves saturation. It is challenging to identify the nodes by density when distance E is set too low since all densities tend to be similar tiny numbers. When E is set too large, it is also easy to make the densities to be too smooth to be distinguished. For parameter M in building high-density parts, a large M will wrongly connect different cluster centers into one cluster while this type of inaccuracy occurs when setting a small M.
Considering the clustering efficiency and clustering accuracy, the values of E and M should not be too large, and when E is 0.8 and M is 18, the best performance can be obtained. As long as the data sources of face images are the same, the same E and M can be easily obtained, because of the same data distribution. While the parameter K in K-means is a sensitive parameter to the clustering results and is hard to choose. Therefore, E and M are two insensitive parameters.
We implement experiments with various ratios of the high-density nodes/parts to the overall nodes. Figure 7 demonstrates that as ρ grows, so does the clustering performance. The best performance achieves when ρ is 0.4. The inference time becomes lower as ρ rises, as seen in Figure 8. After a certain value, the increasing ρ will result in a decrease in performance. This is because high-density nodes/parts are often located around the cluster center. If ρ is too high, it will affect the accuracy of density division, which might result in wrong cluster central parts and worse performance.
The procedure of locating the nodes in the adaptive subgraph is influenced by two hyperparameters k 1 and k 2 . We examine how different values for k 1 and k 2 affect the adaptive subgraphs. We select a higher value k 1 = 100 since we anticipate that more supervised context information will be backpropagated during the training phase. At the same time, we set k 2 = 10 to avoid the adaptive subgraph being too big. In the testing step, we run a comparative test on IJB-B-512 to examine the effects of k 1 and k 2 on performance. Different k 1 and k 2 are employed in experiments and the results are shown in Figure 9.
It is obvious that the performance improves greatly with a bigger value of k 1 in the early stages. More additional neighbor nodes may be introduced by a bigger k 1 , which will also increase the number of candidate node pairs (center node and its neighbors). As the value of k 2 increases, the overall performance shows an upward trend. The contextual information collected by the subgraph becomes increasingly adequate as k 2 increases in value. The performance achieves saturation at a specific level of k 1 and k 2 . When k 1 and k 2 are in a specific range, it may be concluded that it can forecast the connectivity accurately.

5. Discussion

Existing face clustering methods based on graph neural networks leverage a large number of subgraphs to organize the data when clustering faces. These subgraphs are usually highly overlapping and incur redundant computational overheads which will lead to the inference speed being reduced. As the accuracy increases, more and more attention is focused on the time-efficiency.
We propose a more accurate face clustering method based on a density division framework with different connectivity predictions by graph convolutional neural networks, which extracts features from faces images and treats them as different nodes, and divides and builds the nodes into high-density parts and low-density parts based on local density values. The high-density parts have strong connectivity and form the core of different clusters. So we pay more attention to low-density parts. For each node in the low-density parts, an adaptive subgraph is generated based on the context, and the link prediction model and node classification model are used to reason about each subgraph to obtain a set of edge connectivities. The final connectivities predictions are obtained based on the weights of the edges connectivities to determine the connectivity possibility between the pair of nodes. The the high-density parts are merged into the low-density parts, and the pseudo-label propagation strategy is adopted to eliminate the unqualified edges by multiple iterations to obtain the final clustering.
Although the proposed graph neural network-based clustering method constructs adaptive subgraphs for nodes as input, it can contain more contextual information. The overall information is still local structural information, and it is difficult to grasp the structural information of the whole graph. Further research is needed to mine the global information to obtain better performance.

6. Conclusions and Future Work

We propose a supervised image clustering framework DDC-GNN based on density division and graph neural networks, which solves the problem of slow clustering speed caused by redundant subgraph calculation in the process of existing supervised face clustering. By a density division strategy that segments the face images into high-density and low-density parts, the DDC-GNN framework only infers the connectivities for images instance in the low-density parts, which improves the efficiency of face clustering significantly. In addition, DDC-GNN proposes constructing an adaptive subgraph in connectivity learning to capture the context information of a specific node. Experiments on MS-Celeb-1M and IJB-B show that the DDC-GNN framework is both time-efficient and effective.
Current clustering technology still needs a supervised training process and manual labeling. While manual labeling brings much effort when a large number of images are massive, clustering without manual labeling is a future research field, such as clustering by self-supervised learning or unsupervised learning.

Author Contributions

Q.Z. and L.L. wrote the manuscript and designed the comparative experiments; Y.C. supervised the study and revised the manuscript; Z.Y. gave suggestions to and revised the manuscript; Z.W. and W.S. contributed to the research design and manuscript writing. All authors have read and agreed to the published version of the manuscript.

Funding

This research is partially funded by the China National Key Laboratory Foundation of Underwater Measurement and Control Technology and Fundamental Research Funds for the Central Universities under Grant 3072022TS0601, and Singapore’s National Research Foundation (NRF’s) National Cybersecurity R&D Grant GC2018-NCR-0008.

Data Availability Statement

The links to public datasets are as follows, https://github.com/Zhongdao/gcn_clustering (accessed on 2 July 2022) and https://github.com/yl-1993/learn-to-cluster/blob/master/DATASET.md (accessed on 2 July 2022).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Yang, L.; Huang, Q.; Huang, H.; Xu, L.; Lin, D. Learn to propagate reliably on noisy affinity graphs. In Proceedings of the Computer Vision—ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Springer: Berlin/Heidelberg, Germany, 2020; pp. 447–464, Part XV 16. [Google Scholar]
  2. Wang, Z.; Zheng, L.; Li, Y.; Wang, S. Linkage based face clustering via graph convolution network. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 1117–1125. [Google Scholar]
  3. Tian, Y.; Liu, W.; Xiao, R.; Wen, F.; Tang, X. A face annotation framework with partial clustering and interactive labeling. In Proceedings of the 2007 IEEE Conference on Computer Vision and Pattern Recognition, Minneapolis, MN, USA, 17–22 June 2007; pp. 1–8. [Google Scholar]
  4. Xue, J.; Qu, S.; Li, J.; Chu, Y.; Wang, Z. TSC-GCN: A Face Clustering Method Based on GCN. In Proceedings of the 15th International Conference on Knowledge Science, Engineering and Management, Singapore, 6–8 August 2022; pp. 260–271. [Google Scholar]
  5. Li, P.; Zhao, H.; Liu, H. Deep fair clustering for visual learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 9070–9079. [Google Scholar]
  6. Guo, S.; Xu, J.; Chen, D.; Zhang, C.; Wang, X.; Zhao, R. Density-aware feature embedding for face clustering. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 6698–6706. [Google Scholar]
  7. Siddiqui, A.; Mishra, N.; Verma, J.S. A survey on automatic image annotation and retrieval. Int. J. Comput. Appl. 2015, 118, 27–32. [Google Scholar] [CrossRef]
  8. Phillips, P.; Flynn, P.; Scruggs, T.; Bowyer, K.; Chang, J.; Hoffman, K.; Marques, J.; Min, J.; Worek, W. Overview of the face recognition grand challenge. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Diego, CA, USA, 20–25 June 2005; Volume 1, pp. 947–954. [Google Scholar] [CrossRef]
  9. Lloyd, S. Least squares quantization in PCM. IEEE Trans. Inf. Theory 1982, 28, 129–137. [Google Scholar] [CrossRef] [Green Version]
  10. Ester, M.; Kriegel, H.P.; Sander, J.; Xu, X. A density-based algorithm for discovering clusters in large spatial databases with noise. In Proceedings of the Second International Conference on Knowledge Discovery and Data Mining, KDD’96, Portland, OR, USA, 2–4 August 1996; Volume 96, pp. 226–231. [Google Scholar]
  11. Müllner, D. Modern hierarchical, agglomerative clustering algorithms. arXiv 2011, arXiv:1109.2378. [Google Scholar]
  12. Von Luxburg, U. A tutorial on spectral clustering. Stat. Comput. 2007, 17, 395–416. [Google Scholar] [CrossRef]
  13. Zhang, S.; Tong, H.; Xu, J.; Maciejewski, R. Graph convolutional networks: A comprehensive review. Comput. Soc. Netw. 2019, 6, 1–23. [Google Scholar] [CrossRef] [Green Version]
  14. Yang, L.; Zhan, X.; Chen, D.; Yan, J.; Loy, C.C.; Lin, D. Learning to cluster faces on an affinity graph. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 2298–2306. [Google Scholar]
  15. Yang, H.; Chen, X.; Zhang, F.; Hei, G.; Wang, Y.; Du, R. GCN-Based Linkage Prediction for Face Clustering on Imbalanced Datasets: An Empirical Study. arXiv 2021, arXiv:2107.02477. [Google Scholar]
  16. Zhao, Q.; Li, L.; Chu, Y.; Wang, Z.; Shan, W. Density Division Face Clustering Based on Graph Convolutional Networks. In Proceedings of the International Conference on Pattern Recognition, Montréal, QC, Canada, 21–25 August 2022; Springer: Berlin/Heidelberg, Germany, 2022; pp. 1–7. [Google Scholar]
  17. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R.B. Mask R-CNN. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017. [Google Scholar]
  18. Andriyanov, N.; Dementiev, V. Developing and studying the algorithm for segmentation of simple images using detectors based on doubly stochastic random fields. Pattern Recognit. Image Anal. 2019, 29, 1–9. [Google Scholar] [CrossRef]
  19. Kipf, T.N.; Welling, M. Semi-supervised classification with graph convolutional networks. arXiv 2016, arXiv:1609.02907. [Google Scholar]
  20. Schlichtkrull, M.; Kipf, T.N.; Bloem, P.; Van Den Berg, R.; Titov, I.; Welling, M. Modeling relational data with graph convolutional networks. In Proceedings of the European Semantic Web Conference, Monterey, CA, USA, 8–12 October 2018; Springer: Berlin/Heidelberg, Germany, 2018; pp. 593–607. [Google Scholar]
  21. Liu, X.; You, X.; Zhang, X.; Wu, J.; Lv, P. Tensor graph convolutional networks for text classification. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 8409–8416. [Google Scholar]
  22. Kipf, T.; Welling, M. Variational Graph Auto-Encoders. arXiv 2016, arXiv:1611.07308. [Google Scholar]
  23. Hamilton, W.L.; Ying, R.; Leskovec, J. Inductive Representation Learning on Large Graphs. In Advances in Neural Information Processing Systems 30 (NIPS 2017); Curran Associates Inc.: Red Hook, NY, USA, 2017; Volume 30. [Google Scholar]
  24. Chu, Y.; Yue, X.; Yu, L.; Sergei, M.; Wang, Z. Automatic image captioning based on ResNet50 and LSTM with soft attention. Wirel. Commun. Mob. Comput. 2020, 2020, 8909458. [Google Scholar] [CrossRef]
  25. Veličković, P.; Cucurull, G.; Casanova, A.; Romero, A.; Liò, P.; Bengio, Y. Graph Attention Networks. In Proceedings of the International Conference on Learning Representations 2018, Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
  26. Comaniciu, D.; Meer, P. Mean shift: A robust approach toward feature space analysis. IEEE Trans. Pattern Anal. Mach. Intell. 2002, 24, 603–619. [Google Scholar] [CrossRef] [Green Version]
  27. Lin, W.A.; Chen, J.C.; Chellappa, R. A proximity-aware hierarchical clustering of faces. In Proceedings of the 2017 12th IEEE International Conference on Automatic Face & Gesture Recognition, Washington, DC, USA, 30 May–3 June 2017; pp. 294–301. [Google Scholar]
  28. Lin, W.A.; Chen, J.C.; Castillo, C.D.; Chellappa, R. Deep density clustering of unconstrained faces. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 8128–8137. [Google Scholar]
  29. Zhu, H.; Chen, C.; Liao, L.Z.; Ng, M.K. Multiple graphs clustering by gradient flow method. J. Frankl. Inst. 2018, 355, 1819–1845. [Google Scholar] [CrossRef]
  30. Tapaswi, M.; Law, M.T.; Fidler, S. Video face clustering with unknown number of clusters. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019; pp. 5027–5036. [Google Scholar]
  31. Zhan, X.; Liu, Z.; Yan, J.; Lin, D.; Loy, C.C. Consensus-driven propagation in massive unlabeled data for face recognition. In Proceedings of the European Conference on Computer Vision, Munich, Germany, 8–14 September 2018; pp. 568–583. [Google Scholar]
  32. Zhao, X.; Wang, Z.; Gao, L.; Li, Y.; Wang, S. Incremental Face Clustering with Optimal Summary Learning Via Graph Convolutional Network. Tsinghua Sci. Technol. 2021, 26, 536–547. [Google Scholar] [CrossRef]
  33. Qi, C.; Zhang, J.; Jia, H.; Mao, Q.; Song, H. Deep face clustering using residual graph convolutional network. Knowl.-Based Syst. 2021, 211, 106561. [Google Scholar] [CrossRef]
  34. Guo, Y.; Zhang, L.; Hu, Y.; He, X.; Gao, J. Ms-celeb-1m: A dataset and benchmark for large-scale face recognition. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; Springer: Berlin/Heidelberg, Germany, 2016; pp. 87–102. [Google Scholar]
  35. Whitelam, C.; Taborsky, E.; Blanton, A.; Maze, B.; Adams, J.; Miller, T.; Kalka, N.; Jain, A.K.; Duncan, J.A.; Allen, K.; et al. Iarpa janus benchmark-b face dataset. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Honolulu, HI, USA, 21–26 July 2017; pp. 90–98. [Google Scholar]
  36. Yin, S.; Deng, H.; Xu, Z.; Zhu, Q.; Cheng, J. SD-UNet: A Novel Segmentation Framework for CT Images of Lung Infections. Electronics 2022, 11, 130. [Google Scholar] [CrossRef]
  37. Wu, Q.; Feng, D.; Cao, C.; Zeng, X.; Feng, Z.; Wu, J.; Huang, Z. Improved Mask R-CNN for Aircraft Detection in Remote Sensing Images. Sensors 2021, 21, 2618. [Google Scholar] [CrossRef]
  38. Yi, D.; Lei, Z.; Liao, S.; Li, S.Z. Learning face representation from scratch. arXiv 2014, arXiv:1411.7923. [Google Scholar]
  39. Amigó, E.; Gonzalo, J.; Artiles, J.; Verdejo, F. A comparison of extrinsic clustering evaluation metrics based on formal constraints. Inf. Retr. 2009, 12, 461–486. [Google Scholar] [CrossRef] [Green Version]
  40. Frey, B.J.; Dueck, D. Clustering by passing messages between data points. Science 2007, 315, 972–976. [Google Scholar] [CrossRef] [Green Version]
  41. Otto, C.; Wang, D.; Jain, A.K. Clustering millions of faces by identity. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 40, 289–303. [Google Scholar] [CrossRef] [Green Version]
  42. Sibson, R. SLINK: An optimally efficient algorithm for the single-link cluster method. Comput. J. 1973, 16, 30–34. [Google Scholar] [CrossRef]
  43. Shi, J.; Malik, J. Normalized cuts and image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2000, 22, 888–905. [Google Scholar]
  44. Bo, D.; Wang, X.; Shi, C.; Zhu, M.; Lu, E.; Cui, P. Structural Deep Clustering Network. In Proceedings of the Web Conference 2020, Taipei, Taiwan, 20–24 April 2020; ACM: New York, NY, USA, 2020; pp. 1400–1410. [Google Scholar] [CrossRef]
Figure 1. The framework of DDC-GNN. Firstly, the face data are divided into high-density and low-density parts. The high-density parts are located at the center of the cluster and have strong connectivity. Secondly, we use each data point of the low-density parts as a node, and construct an ASG for each node only in low-density parts, then apply GNN on ASGs to predict the connectivities among the nodes. After that, all the nodes are connected into a big graph with their linkage and clusters are obtained by iteratively deleting unnecessary links.
Figure 1. The framework of DDC-GNN. Firstly, the face data are divided into high-density and low-density parts. The high-density parts are located at the center of the cluster and have strong connectivity. Secondly, we use each data point of the low-density parts as a node, and construct an ASG for each node only in low-density parts, then apply GNN on ASGs to predict the connectivities among the nodes. After that, all the nodes are connected into a big graph with their linkage and clusters are obtained by iteratively deleting unnecessary links.
Remotesensing 14 03768 g001
Figure 2. Construction of adaptive subgraph. (a) Finding nodes; (b) Adding edges.
Figure 2. Construction of adaptive subgraph. (a) Finding nodes; (b) Adding edges.
Remotesensing 14 03768 g002
Figure 3. The structure of node classification model GCN.
Figure 3. The structure of node classification model GCN.
Remotesensing 14 03768 g003
Figure 4. The structure of link prediction model GAE.
Figure 4. The structure of link prediction model GAE.
Remotesensing 14 03768 g004
Figure 5. The impact of E.
Figure 5. The impact of E.
Remotesensing 14 03768 g005
Figure 6. The impact of M.
Figure 6. The impact of M.
Remotesensing 14 03768 g006
Figure 7. The impact of different high-density ratios on performance.
Figure 7. The impact of different high-density ratios on performance.
Remotesensing 14 03768 g007
Figure 8. The impact of different high-density ratios on Runtimes.
Figure 8. The impact of different high-density ratios on Runtimes.
Remotesensing 14 03768 g008
Figure 9. Sensitivity of performance with parameters k 1 and k 2 .
Figure 9. Sensitivity of performance with parameters k 1 and k 2 .
Remotesensing 14 03768 g009
Table 1. Notations and Descriptions.
Table 1. Notations and Descriptions.
NotationsDescriptions
V, vV is a set of nodes in the graph, and v is an instance that belongs to V.
AA is the adjacency matrix of a graph.
FF is the feature set of extracted features from the pre-trained model.
CC is the cluster result.
DD is the dimension of the features.
NN is the number of face images.
Λ Λ is a diagonal matrix.
dd is a set of node density values.
WW is a learnable weight matrix.
ZZ is the hidden layer feature.
V s V s is the node set of the subgraph.
E s E s is the edge set of the subgraph.
G ( V s , E s ) G ( V s , E s ) is the input subgraph of GNN.
F p F p is the feature set of the subgraph which is built for node p.
Table 2. The architecture of the node classification model GCN.
Table 2. The architecture of the node classification model GCN.
MethodLayers
GCN-based4 Layers GCN
2 layers MLP
Softmax Activate Function
cross-entropy Loss Function
Table 3. The architecture of the link prediction model GAE.
Table 3. The architecture of the link prediction model GAE.
MethodLayers
GAE-basedEncoder: 2 Layers GCN
Decoder: Inner Product
Sigmoid Activate Function
cross-entropy Loss Function
Table 4. Comparison on IJB-B datasets.
Table 4. Comparison on IJB-B datasets.
MethodIJB-B-512IJB-B-1024IJB-B-1845
FNMIARISC(0.166)TimeFNMIARISC(0.170)TimeFNMIARISC(0.194)Time
K-Means [9]0.6390.8560.4470.181175 s0.6030.8650.4150.166699 s0.6000.8680.3400.1642307 s
AP [40]0.4940.854---0.4840.864---0.4770.869---
DBSCAN [10]0.7680.9040.7760.18711 s0.7650.8770.1520.14142 s0.7370.8510.0650.116150 s
Spectral Clustering [43]0.5380.7740.1770.143285 s0.5220.7800.1260.1331692 s0.5510.8400.2960.1363252 s
ARO [41]0.6850.8870.7470.077267 s0.5990.8800.6340.063551 s0.5230.8740.4480.0551062 s
SDCN [44]0.4740.836---0.4520.830--------
DDC [28]0.8020.921---0.8050.926---0.8000.929---
L-GCN [2]0.8210.9200.8260.206108 s0.8190.9280.8460.110198 s0.8100.9260.6760.187362 s
NDD-GCN0.8260.9260.8300.20997 s0.8190.9310.8460.111192 s0.8120.9300.6760.181350 s
DDC-GAE0.8230.9060.8290.22646 s0.8170.9330.8290.10870 s0.8090.9190.6750.16798 s
DDC-GCN0.8290.9170.8330.20732 s0.8220.9350.8480.11851 s0.8160.9310.6770.17575 s
Table 5. Comparison on MS-Celeb-1M datasets.
Table 5. Comparison on MS-Celeb-1M datasets.
MethodsPrecisionRecallF-ScoreNMITime
K-Means [9]52.5270.4560.1894.5613 h
DBSCAN [10]72.8842.4653.5092.31100 s
HAC [42]66.8470.0168.3992.9218 h
ARO [41]81.107.3013.3484.68250 s
CDP [31]80.1970.4775.0194.69350 s
L-GCN [2]84.4072.3678.8694.762500 s
NDD-GCN87.2474.9880.6595.472042 s
DDC-GAE82.7679.9680.8595.49650 s
DDC-GCN84.7877.5680.9595.53420 s
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zhao, Q.; Li, L.; Chu, Y.; Yang, Z.; Wang, Z.; Shan, W. Efficient Supervised Image Clustering Based on Density Division and Graph Neural Networks. Remote Sens. 2022, 14, 3768. https://doi.org/10.3390/rs14153768

AMA Style

Zhao Q, Li L, Chu Y, Yang Z, Wang Z, Shan W. Efficient Supervised Image Clustering Based on Density Division and Graph Neural Networks. Remote Sensing. 2022; 14(15):3768. https://doi.org/10.3390/rs14153768

Chicago/Turabian Style

Zhao, Qingchao, Long Li, Yan Chu, Zhen Yang, Zhengkui Wang, and Wen Shan. 2022. "Efficient Supervised Image Clustering Based on Density Division and Graph Neural Networks" Remote Sensing 14, no. 15: 3768. https://doi.org/10.3390/rs14153768

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop