Next Article in Journal
Hilbert Space Delocalization under Random Unitary Circuits
Previous Article in Journal
TransConv: Transformer Meets Contextual Convolution for Unsupervised Domain Adaptation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

View-Driven Multi-View Clustering via Contrastive Double-Learning

Information Engineering College, Shanghai Maritime University, Shanghai 201306, China
*
Author to whom correspondence should be addressed.
Entropy 2024, 26(6), 470; https://doi.org/10.3390/e26060470
Submission received: 22 April 2024 / Revised: 17 May 2024 / Accepted: 27 May 2024 / Published: 29 May 2024
(This article belongs to the Section Multidisciplinary Applications)

Abstract

:
Multi-view clustering requires simultaneous attention to both consistency and the diversity of information between views. Deep learning techniques have shown impressive abilities to learn complex features when working with extensive datasets; however, existing deep multi-view clustering methods often focus only on either consistency information or diversity information, making it difficult to balance both aspects. Therefore, this paper proposes a view-driven multi-view clustering using the contrastive double-learning method (VMC-CD), aiming to generate better clustering results. This method first adopts a view-driven approach to consider information from other views to encourage diversity, thus guiding feature learning. Additionally, it presents the idea of dual contrastive learning to enhance the alignment of views at both the clustering and feature levels. The VMC-CD method’s superiority over various cutting-edge methods is substantiated by experimental findings across three datasets, affirming its effectiveness.

1. Introduction

Multi-view data usually include representations from diverse features or sources, where each view contains shared semantic information inherent in the multi-view data. The insights derived from multiple views tend to complement each other [1,2]. Visual information, for instance, can be characterized through diverse techniques like SIFT, HOG, and LBP. Likewise, environmental data like temperature and humidity can be gathered using several sensors positioned across a specified region, while the details of these data might vary, there are overarching similarities in the cluster patterns when viewed from a broader perspective. Multi-view clustering seeks to categorize data into various groups by leveraging insights from all accessible viewpoints [3,4,5,6]. However, acquiring knowledge from multiple sources simultaneously is challenging [7].
Multi-view learning (MVC) has gained significant interest across various machine learning applications, such as feature selection [2], scene recognition [8], and information retrieval [9,10,11]. Traditional machine learning approaches can be broadly classified into four categories: subspace learning methods [12], non-negative matrix factorization (NMF) methods [13,14], graph-based techniques [1,15], and a range of kernel-based methods [16]. However, traditional shallow models often struggle to effectively learn feature representations from large datasets [17].
To tackle these challenges, a plethora of deep learning techniques have been introduced [18,19,20,21,22]. Deep multi-view clustering endeavours to improve performance by harnessing the feature representation capabilities inherent in deep models. In essence, these methods strive to enhance distinct consensus representations by utilizing specialized encoder networks for each viewpoint to transform the data.
Recently, contrastive learning has been integrated into deep learning frameworks to obtain unique representations from various viewpoints [23,24]. Most contrastive learning techniques focus on maximizing the common information found in the distributions of multiple perspectives [25]. As an illustration, Yang et al. [26] utilized existing data as positive samples and randomly chose certain cross-view samples as negative samples in deep multi-view clustering (MVC).
Although these multi-view learning techniques have shown impressive outcomes, they still encounter common challenges. Initially, prior algorithms often map diverse views into a unified space under the assumption that there exists a substantial correlation between views. However, in practice, both correlation (consistency) and independence (complementarity) coexist, making it difficult to balance them. Consequently, current algorithms are either focused on maximizing correlation for coherence [27] or on achieving independence for diversity [28], thereby merging coherence and diversity into an intricate and unified challenge [29,30].
In contrast to previous methods, this paper introduces a VMC-CD methodology. Specifically, in the process of learning different view representations, it considers the interesting information from other views and employs two levels of contrastive learning to restrict them for potential clustering and further guide feature learning. Clustering-level and feature-level contrastive learning are two essential aspects that contribute to distinct stages of feature learning and synergistically strengthen each other. In clustering-level contrastive learning, the invariant representations of multiple views are aligned and feature information obtained from a single view is normalized from the perspective of the clustering assignment. Feature-level contrastive learning endeavours to align encoded feature representations specific to each view, thereby mitigating heterogeneity among different views to some extent. This paper introduces an attention-driven approach to the generation of a discriminative feature representation, which aids in guiding feature learning. This method boosts attention towards information from all views while diminishing the influence of information specific to irrelevant subsets of views regarding clustering. Moreover, this paper illustrates that aligning latent feature distributions across different views using contrastive learning can achieve robust view invariance. Additionally, this approach outperforms existing deep view clustering methods in terms of clustering results, even without the requirement of a well-initialized autoencoder. To summarize, the key contributions of this paper include the following:
  • Introduction of a VMC-CD technique, which incorporates valuable information from alternate views while learning feature representations across diverse viewpoints. It provides guidance information in an attention-driven manner, effectively integrating multiple views into a discriminative common representation to guide feature learning.
  • Introduction of dual contrastive learning, conducting contrastive learning at both the clustering and feature levels, encouraging consistency in clustering across multiple views while preserving their feature diversity.
  • Experiments on three multi-view datasets, demonstrating the effectiveness of the VMC-CD method.

2. Related Work

This section delves into recent advancements in pertinent areas, specifically focusing on multi-view clustering and contrastive learning.

2.1. Multi-View Clustering

Multi-view clustering and classification techniques can generally be categorized into two main types: conventional methods and deep learning-based approaches. Conventional multi-view clustering methods can be subdivided into five distinct categories. Firstly, some methods are achieved through non-negative matrix factorization techniques, such as that in Liu et al. [31], who explored common latent factors between multiple views and established a deep structure [32] to find more consistent shared features.
Approaches in the second category utilize self-representation to illustrate the relationships among samples [33]. In research conducted by [5], a self-representation layer was employed to hierarchically reconstruct view-specific subspaces and encoding layers, thereby enhancing the consistency of cross-view subspaces.
Approaches in the third category employ dimensionality reduction methods to convert multi-view data into a common, low-dimensional space, enabling a uniform representation. Subsequently, clustering outcomes are derived using established clustering methods [34]. Canonical Correlation Analysis (CCA) [35] is a notable technique within this branch. In a recent study [36], a versatile framework was introduced for reducing the dimensionality of multi-view data, enabling the handling of multi-view feature representations within kernel space.
Methods in the fourth category employ graph models for multi-view clustering [37,38]. The central concept of this method is to identify a common graph among various perspectives and then utilize spectral graph techniques (like spectral clustering) on this shared graph to derive clustering outcomes. Moreover, a study by [39] introduced graph autoencoders for learning multi-view representations. The study [40] focused on extracting valuable insights from complex multi-view data dispersed across various high-dimensional spaces. Through graph learning, the fundamental correlations between different views are uncovered, thereby addressing the issue of effective multi-view collaboration.
The last category of methods tackles this issue using kernel function strategies [41,42], frequently utilizing predefined kernel functions like Gaussian kernels to handle diverse views. Subsequently, these methods linearly or nonlinearly blend these kernel functions to establish a uniform kernel. Yet, the primary challenge with this method is the identification of appropriate kernel functions.
These statistical models face a common limitation in their ability to capture intricate structures within the data. As a result, deep multi-view clustering has garnered considerable attention within the community and has demonstrated effectiveness across various practical scenarios.
In early research, Wang et al. [18] employed a deep autoencoder design to acquire a consolidated representation of multi-view data, yielding commendable results in speech and visual analysis tasks. Subsequently, Andrew et al. [27] introduced an enhancement to Deep Canonical Correlation Analysis (DCCA). Their work centred on creating a unified representation of multi-view data by maximizing the correlation between extracted deep features and CCA. Subsequently, Abavisani et al. [43] introduced a deep multi-view subspace clustering network aimed at revealing a unified affinity matrix across all viewpoints. Moreover, Zhu et al. [44] utilized deep autoencoders for self-representation learning and incorporated diversity and ubiquitous regularization to capture meaningful interconnections among different viewpoints.
While existing algorithms typically prioritize either maximizing view correlation for consistency or maximizing view independence for complementarity, this paper advocates for emphasizing diversity alongside maintaining consistency. This balanced approach aims to achieve improved results by striking a harmonious equilibrium between diversity and consistency.

2.2. Contrastive Learning

Contrastive learning has significantly progressed in the realm of self-supervised representation learning [24]. Fundamentally, contrastive learning strives to enhance the feature space of raw data by amplifying similarities among positive pairs (similar instances) while reducing the similarities among negative pairs (dissimilar instances) [45]. Positive pairs generally consist of data from the identical instance, while negative pairs consist of data from different instances.
For instance, Chen et al. [24] introduced a visualization representation framework within contrastive learning. This framework seeks to optimize the agreement between diverse augmented views of a singular example within the latent feature space.
Lately, an approach named Contrastive Prediction (COMPLETER) [46] has advanced significantly by combining reconstruction, cross-view contrastive learning, and cross-view dual prediction methodologies. This method stands out not just for its effectiveness in incomplete multi-view clustering but also for its ability to simultaneously handle data recovery and consistency learning in incomplete multi-view datasets.
These methods contribute to learning high-quality representations based on data. However, determining invariant representations across multiple views remains a challenging problem.

3. Methods

In this section, we initially present a clear formulation of the problem and delineate its particulars. Next, we propose a network framework to address this problem. We then delve into each module of the proposed network, including the deep autoencoder module, dual contrastive learning module, and attention weight learning module, in detail.

3.1. Problem Formulation

Given a set of multi-view data X = X ( ν ) R d v × N ν = 1 n v with n v views and N samples, let us denote the v-th view of the multi-view data as X ( v ) . Each view X ( ν ) = X 1 ( ν ) , X 2 ( ν ) , , X N ( ν ) has N samples, where X i ( ν ) ( 1 i N ) dimensions are d v ; this represents a certain sample dimension of a particular view, and it is important to note that different samples in each view may have different dimensions. Given K as the cluster count, instances with identical semantic labels can be grouped together into a shared cluster. Hence, there is a requirement to partition N samples into K distinct clusters.

3.2. Overview of the Network Architecture

According to Figure 1, the VMC-CD method aims to directly extract semantic labels for end-to-end clustering from raw data instances spanning multiple perspectives. We achieve this by applying the dual contrastive learning module to feature representation learning, introducing an end-to-end deep clustering network structure. Additionally, we have given special treatment to the encoder by integrating a view-driven attention mechanism. As shown in the diagram below, the proposed VMC-CD network architecture consists of three main modules: the deep autoencoder module, the dual contrastive learning module, and the attention weight learning module (AT BLOCK). The core of the entire architecture is the deep autoencoder module, which learns features conducive to clustering across multiple perspectives through unsupervised representation learning. The dual contrastive learning module is divided into two parts: one part performs contrastive learning on the discriminative feature representation learned by the aforementioned encoder–decoder, known as the feature-level contrastive learning part, and the other part optimizes parameters through contrastive soft clustering assignment, known as the clustering-level contrastive learning part. The attention weight learning module primarily enhances the clustering level of the discriminative feature representation by leveraging information from other views.

3.3. Deep Autoencoder Module

Our network architecture primarily relies on a deep autoencoder module comprising multi-view feature encoders and multi-view feature decoders. When learning feature representations, our attention mechanism incorporates information from other views. Thus, during the feature encoding phase, we take into account relevant information from alternate views. Illustrated in the diagram below, the multi-view feature encoder utilized in this study comprises two components: a view-specific autoencoder module and an attention module influenced by other views, specifically the attention weight learning module (discussed in detail in Section 3.5).
The view-specific encoder module comprises three initial blocks: a linear layer, a batch normalization layer, and an activation layer (ReLU). The feature encoder module primarily aims to convert view-specific data into a discriminative feature representation. This is achieved by integrating the output of the view-specific autoencoder module with the output of the attention module, which is influenced by information from other perspectives. Subsequently, a softmax function is applied to generate the feature representation. On the other hand, the feature decoder module performs the opposite operation, converting the discriminative feature representation back to the original view information. The construction of each decoder block is the same as that of the encoder block.
The overview steps of the autoencoder module are as follows: First, based on input feature data X i ( ν ) , the encoder component acquires a compact representation Z i ( v ) , where X i ( ν ) represents the i-th sample of the v-th view, and Z i ( v ) represents the low-dimensional representation of the i-th sample of the v-th view. This is the simplified formula:
Z i ( v ) = f E X i ( ν )
Here, f E ( . ) broadly refers to a series of operations by the encoder on the input data X i ( ν ) . For the v-th view, where v = m, the specific encoder is as follows:
r m i ( e ) = f B N W r ( e ) T X i ( m ) + b r ( e ) e = 1 f B N W r ( e ) T r m i ( e 1 ) + b r ( e ) e = 2 , 3 f B N W r ( e ) T r m i ( e 1 ) + b r ( e ) e = 4
r m i ( e ) represents the latent representation of the i-th sample of the m-th view after passing through a certain layer of the encoder. W r ( e ) denotes the weights, while b r ( e ) denotes the biases of the encoder segment. BN represents the batch normalization operation. f ( · ) represents the ReLU activation function. The attention module driven by other views obtains attention weights through the sigmoid function, denoted by o m i ( 6 ) (details on how to calculate this will be provided in Section 3.5). This module embeds the interesting information into the attention weights, which are then element-wise multiplied with the view-specific feature representation, resulting in a view-specific discriminative representation. The specific formula is as follows:
Z i ( m ) = ζ r m i ( 4 ) o m i ( 6 )
Here, ξ ( . ) represents the result after passing through the sigmoid function. Subsequently, the decoder transforms the reconstructed features Z i ( m ) back into the original input data by extracting hidden representations. The Equation is as follows:
X i ( m ) = f D Z i ( m )
The specific decoder is as follows:
r m i ( d ) = f B N W r ( d ) T Z i ( m ) + b r ( d ) d = 1 f B N W r ( d ) T r m i ( d 1 ) + b r ( d ) d = 2 , 3 , 4
r m i ( d ) represents the latent representation of the i-th sample of the m-th view after passing through a certain layer of the decoder. W r ( d ) and b r ( d ) represent the weights and biases of the encoder part. Let X i ( m ) ^ = r m i ( 4 ) . X i ( m ) ^ represents the reconstructed data of the i-th sample of the m-th view. This allows us to construct the reconstruction loss function. In this study, the autoencoder network’s objective function is attained through the minimization of the reconstruction error. After extrapolating the loss of the m-th view to all views, the total reconstruction loss is as follows:
r e c = v = 1 n v i = 1 N X i ( v ) X i ( v ) ^

3.4. Dual Contrastive Learning Module

The dual contrastive learning module is divided into two parts. One part performs contrastive learning on the discriminative feature representation learned by the encoder–decoder, referred to as feature-level contrastive learning in this paper. The other part optimizes parameters through contrastive clustering assignment, referred to as clustering-level contrastive learning.
Feature-level contrastive learning is performed within the latent space of the autoencoder representation to explore the common information representation across various views. This process focuses on learning the alignment between different views by maximizing their mutual information. The loss function for feature-level contrastive learning is denoted by:
c h = 1 m n n v i = 1 N I Z i ( m ) , Z i ( n ) + H Z i ( m ) + H Z i ( n )
Here, I represents mutual information, H represents entropy, and the parameter is used to regularize entropy. According to information theory, entropy represents information content. Hence, a higher entropy H Z i ( m ) signifies a larger information content within X i ( m ) , ensuring diversity of information across different views. Additionally, maximizing mutual information I Z i ( m ) , Z i ( n ) between H Z i ( m ) and H Z i ( n ) will maintain information coherence across diverse perspectives during feature acquisition.
The contrastive clustering assignment used here is a soft clustering assignment method, unlike hard clustering, which allows data points to be assigned to multiple categories with different probabilities or membership degrees. Soft clustering assigns each data point a membership value for every category, denoting the degree to which the data point pertains to that specific category. These membership values form a membership matrix, where data points are rows and categories are columns, reflecting the membership of data points to each category. In contrast, hard clustering assignment requires each data point to be explicitly and uniquely assigned to a single category, without allowing for sharing or ambiguity. The specific application in this paper is as follows: For any view V = v, after obtaining r v i ( 4 ) , a separate branch is opened for all views to undergo further processing. The specific operational process is as follows:
r v i ( e ) = W r ( e ) r v i ( e 1 ) + b γ ( e ) e = 5 , 6
As shown in the above equation, we first pass it through two linear layers with the purpose of dimensionality reduction, making the dimension of this vector equal to the number of clusters K, in order to proceed with the next step of soft clustering assignment. Let r ν i ( 6 ) = H i ( ν ) . If all samples are processed uniformly according to the above, matrix H ( v ) R N × K v = 1 n v can be obtained. Here, H i ( v ) denotes the i-th row of H ( v ) and H i j ( v ) represents the j-th element of the i-th row of matrix H ( v ) , indicating the likelihood that sample i in view vs. belongs to cluster j.
To enhance the diversity between cluster assignments and thus strengthen the effectiveness of soft clustering, Q ( v ) R N × K v = 1 n v is used to reinforce the results of H ( v ) R N × K v = 1 n v , improving the performance of soft clustering. The calculation process for each element in Q ( v ) is as follows:
Q i j ( v ) = H i j ( v ) 2 / i = 1 N H i j ( v ) k = 1 K H i k ( v ) 2 / i = 1 N H i k ( v )
Let Q j ( ν ) be the j-th column of Q ( v ) . Each element in Q j ( ν ) , denoted by Q i j ( v ) , represents the soft clustering assignment of sample i to cluster j. Therefore, Q j ( ν ) denotes the clustering assignment of samples belonging to the same semantic cluster. Samples that are the same across different views share the same semantic information. The similarity between two clustering assignments Q j v 1 and Q j v 2 for cluster j can be measured by the following equation:
s Q j v 1 , Q j v 2 = Q j v 1 T Q j v 2
The symbols v 1 and v 2 represent two different views, but the clustering assignment probabilities of instances between different views are similar because these instances represent the same samples. Additionally, if instances from multiple views are used to describe different samples, they are uncorrelated with each other. The similarity between cluster assignments within clusters should be maximized, while the similarity between cluster assignments across clusters should be minimized. We perform clustering on samples concurrently, ensuring coherence in the clustering assignments. The cross-view contrastive loss between Q k v 1 and Q k v 2 is defined as follows:
ν 1 , ν 2 = 1 K k = 1 K log e s Q k v 1 , Q k v 2 / τ T T = j = 1 , j k K e s Q j v 1 , Q k v 2 / τ + j = 1 K e s Q j v 1 , Q k v 2 / τ
The symbol τ represents a temperature parameter, Q k v 1 , Q k ν 2 denotes positive clustering assignments pairs between two views, while Q j v 1 , Q k v 1 and Q j v 1 , Q k ν 2 ( j k ) represent negative clustering assignment pairs between two views. The cross-view contrastive loss induced across multiple views is designed as follows:
c = 1 2 v 1 = 1 n v v 2 = 1 , v 2 v 1 n v v 1 , v 2
The cross-view contrastive loss explicitly compares clustering assignment pairs across multiple views. It pulls together pairs from the same cluster assignment and pushes apart pairs from different cluster assignments. To avoid a scenario where all instances are grouped into a single sub-cluster, we introduce a regularization term as follows:
a = ν = 1 n v j = 1 K P j ( ν ) log P j ( ν )
The term P j ( ν ) = i = 1 N Q i j ( ν ) N represents the loss defined as the cross-view consistency loss, which prevents all instances from belonging to the same cluster ‘j’. Therefore, the total loss for the contrastive clustering level is as follows:
c l = c + μ a

3.5. The Attention Weight Learning Module (AT BLOCK)

As shown in Figure 2, in learning the feature representations of multiple views, attention is generated from other views, incorporating interest information from these views during the feature encoding process.
We constructed the AT BLOCK using fully connected layers with ReLU and a transformer, connecting the transformer with two ReLU fully connected layers’ inputs and outputs through skip connections. The primary purpose of AT BLOCK is to map complex data into spaces corresponding to different views, obtaining attention weights for different views to guide feature learning.
The attention module is structured with a sequence of fully connected layers followed by ReLU activation. Through the sigmoid function, the attention module calculates attention weights, which encapsulate relevant information within the dataset.
In the multi-view feature encoder input, with two views, the feature learning procedure integrates input from the other view’s data into the attention-driven module to support feature learning. This is symbolized as A 1 = X 2 and A 2 = X 1 . The specific process of the attention module is as follows:
O m i ( e ) = f W o ( e ) T A i ( m ) + b o ( e ) e = 1 f W o ( e ) T O m i ( e 1 ) + b o ( e ) e = 2 , 4 , 5 t O m i ( e 1 ) e = 3 O m i ( 2 ) + O m i ( 5 ) e = 6
Here, A i ( m ) represents the input to the attention module for the i-th sample of the m-th view, O m i ( e ) represents the feature representation after the e-th layer of the attention module, and t ( · ) represents the result after the transformer block. W o ( e ) and b o ( e ) denote the weights and biases of the linear layer in the e-th layer.

3.6. Total Loss Function

After introducing all the losses and their computation methods, we can obtain the total loss function of VMC-CD:
v m c = c l + λ 1 c h + λ 2 r e c
Here, λ 1 and λ 2 are weighting hyperparameters. c l represents the loss function for cluster-level contrastive learning. c h represents the loss function for feature-level contrastive learning. r e c represents the loss function for reconstruction. The weighted sum of these three constitutes the total loss function v m c in this context.

3.7. Complexity Analysis

Let α and β represent the mini-batch size and maximum number of neurons in the proposed network architecture’s hidden layer, respectively. Let d z denote the dimensionality of the view feature representation. The overall complexity of the model is denoted by O α β n v d v , while the complexities of the reconstruction loss, feature-level contrastive learning loss, and cluster-level contrastive learning loss are represented by O α n v d v , O α d z n v and O α 2 K n v n v 1 + n v ( K 1 ) + n v K , respectively. Therefore, the overall complexity of the proposed method is denoted by O T α β n v d v + α d z n v + α 2 K 2 n v 2 , where T represents the maximum number of iterations during training.

3.8. Algorithm Flow

This algorithm flow (Algorithm 1) is shown below.
Algorithm 1 View-driven dual-contrastive learning in multi-view clustering
  • Requirements: multi-view data samples X = X ( ν ) R d v × N ν = 1 n v , maximum number of iterations T max
  • 1: Initialize the parameters of the autoencoder network and set t = 0
  • 2: While t < T max and loss function v m c is not converged do
  • 3: Compute the loss and update the parameters of the entire network
  • 4: t = t + 1
  • 5: Obtain discriminative feature representations for all views
  • 6: Concatenate different view feature representations of the same sample to form Z 1 : : Z n v , pass it through the k-means clustering algorithm, yielding the clustering outcome denoted as Q
  • 7: Output: Clustering result Q

4. Experiment

In this section, comprehensive experiments were conducted to evaluate the efficacy of the VMC-CD method proposed in this study. We performed experiments on five commonly utilized multi-view datasets, evaluating the performance of our method against other established multi-view clustering techniques. The source code of VMC-CD is implemented in Python 3.7. All experiments were carried out using a system that includes a GeForce RTX 3080 Ti GPU with 16 GB of memory, a 12th Gen Intel Core i9-12900H CPU, and 32 GB of RAM.

4.1. Experimental Setup

4.1.1. Datasets

In this study, we utilized five commonly used datasets. These datasets are Caltech101-20 [47], Scene-15 [48], LandUse-21 [49], MNIST-USPS [50,51] and BDGP [52]. The Caltech101-20 dataset comprises 2386 images representing 20 subjects and incorporates HOG and GIST features as distinct perspectives. The LandUse-21 dataset includes 2100 satellite images across 21 classes, utilizing PHOG, LBP, and GIST features. The Scene-15 dataset comprises 485 images showcasing 15 scenes and incorporates PHOG, LBP, and GIST features. The MNIST-USPS dataset is a handwritten digit image dataset with two different styles, each view containing 10 categories with 500 examples per category. The BDGP dataset has 2500 drosophila embryo images, five categories, each image with 1750-dimensional visual and 79-dimensional textual features for clustering.

4.1.2. Evaluation Metrics

In this study, we utilize accuracy (ACC), normalized mutual information (NMI), and the adjusted Rand index (ARI) as the primary metrics to assess clustering performance. Improved clustering outcomes are indicated by higher values on these metrics.

4.1.3. Network Architecture and Parameter Settings

The VMC-CD model was trained using the Adam optimizer with an initial learning rate of 0.0001. The batch size was fixed at 256, and the number of training iterations varied depending on the dataset: 200 iterations for Caltech101-20, MNIST-USPS, and BDGP, 700 iterations for LandUse-21, and 500 iterations for Scene-15. For all datasets, the entropy parameter in the feature-level contrastive learning, denoted as , is set to 9, while the temperature coefficient, denoted as τ , is set to 1. The hyperparameters λ 1 , λ 2 , and μ are chosen from the range [0.05, 0.1, 0.2, 0.5, 1] based on different datasets. For cluster-level contrastive learning, two linear layers are established. The dimension of the first linear layer is selected from the range of [32, 64] depending on the dataset, while the dimension of the second linear layer is configured to match the number of clusters in the dataset.

4.2. Performance Evaluation

As shown in Table 1, for Caltech101-20, Scene-15, and LandUse-21 datasets, this study contrasted the proposed approach with 11 other multi-view clustering methods, including IMG (Incomplete Multi-View Grouping for Visual Data) [53], EERIMVC (Efficient and Effective Regularized Incomplete Multi-View Clustering) [16], DAIMC (Dual-Aligned Incomplete Multi-View Clustering) [54], UEAF (Unified Embedding Alignment Framework) [55], DCCAE (Deep Canonical Correlation Autoencoder) [56], PVC (Partial Multi-View Clustering) [57], AE2-Nets (Autoencoders in Autoencoder Networks) [58], DCCA (Deep Canonical Correlation Analysis) [27], PICCAE (Probabilistic Incomplete Canonical Correlation Autoencoder) [59], COMPLETER (Contrastive Prediction) [46], and ATTENTION (Attention-Driven Deep Multi-View Clustering) [60].
In comparison with ATTENTION, the proposed method achieved relative improvements in accuracy of 2.82%, 2.84%, and 1.08% on the Caltech101-20 dataset, Scene-15 dataset, and LandUse-21 dataset, respectively.
As shown in Table 2, we also conducted experiments on two additional datasets and compared our model with other models that perform well on these datasets, achieving excellent results. The comparison methods include: Deep Embedded Clustering (DEC) [22], Improved Deep Embedded Clustering (IDEC) [61], Binary Multi-View Clustering (BMVC) [62], Multi-View Clustering via Late Fusion Alignment Maximization (MVC-LFA) [63], Deep Adversarial Multi-View Clustering Network (DAMC) [21], Self-Paced and Auto-Weighted Multi-View Clustering (SAMVC) [64], Cognitive Deep Incomplete Multi-View Clustering Network (CDIMC-net) [65], End-to-End Adversarial-Attention Network for Multi-Modal Clustering (EAMC) [6], and Reconsidering Representation Alignment for Multi-View Clustering (SiMVC and CoMVC) [54]. On the MNIST-USPS dataset, our model outperforms the second-best method, CoMVC [66], by 0.99%, 1.11%, and 0.79% in terms of ACC, NMI, and ARI, respectively. On the BDGP dataset, our model surpasses the second-best model, DAMC [21], by 0.78%, 2.21%, and 3.16% in ACC, NMI, and ARI, respectively.
Therefore, our model was evaluated on a total of five datasets. In previous studies, models typically demonstrated excellent performance on only a limited number of datasets. In contrast, the VMC model exhibited outstanding performance across all five datasets, showcasing its exceptional generalization capabilities. This broad applicability highlights the robustness and versatility of the VMC model.

4.3. Ablation Studies

According to the overall loss equation, which includes three different loss components; the first part is the reconstruction loss component for obtaining consensus representation, the second part is the feature-level contrastive learning component, and the third part is the cluster-level contrastive learning component. To validate the importance of components in VMC-CD, we conducted ablation studies using the same experimental settings to isolate external interference factors. Specifically, we considered two special cases: one where only cluster-level loss is considered during end-to-end training, without considering feature-level loss and another where only feature-level loss is considered during end-to-end training, without considering cluster-level loss. The table displays the results of these two special cases along with the three metrics of our model. The clustering outcomes presented in the initial two rows of the table correspond to the two distinct scenarios. As anticipated, optimal performance is attained when both feature-level contrastive learning and cluster-level contrastive learning are simultaneously incorporated.
In terms of accuracy, VMC-CD with dual contrastive learning outperformed the model without feature-level loss by 9.22%, 0.98%, and 0.09% on the three datasets and the NMI and ARI metrics also showed improvements. VMC-CD with dual contrastive learning also performed better compared to the model without cluster-level loss, with improvements of 24.85%, 3.41%, and 2.38% on the three datasets. Therefore, dual contrastive learning plays a crucial role in learning invariant representations across views and is indispensable. The specific experimental results are shown in Table 3, Table 4 and Table 5.

4.4. Parameter Sensitivity Analysis

As shown in Figure 3, we conducted experiments on the Caltech101-20 dataset to study the sensitivity of the parameters λ 1 and λ 2 in the proposed VMC-CD method. The λ 1 parameter is selected from {0.1, 0.5, 1, 2}, and the λ 2 parameter is selected from {0.01, 0.05, 0.1, 0.5, 1}. The chart showcases how the VMC-CD method performs in clustering, measured by ACC, NMI, and ARI scores, across various combinations of λ 1 and λ 2 . The clustering performance of the VMC-CD method on the Caltech101-20 dataset varies with different combinations of λ 1 and λ 2 . The results exhibit relative consistency in ACC and NMI scores but sensitivity to the λ 1 parameter concerning ARI. Specifically, an increase in the λ 1 parameter correlates with a notable decrease in ARI.

4.5. Training Analysis

Figure 4 displays the curves depicting the evolution of each clustering metric with respect to the iteration count on the Scene-15 dataset. The illustrated curves showcase the exceptional stability of the method proposed in this paper, consistently delivering robust clustering performance.
Apart from the previously discussed visualizations, we also include t-SNE [67] visualizations showcasing the learning of a unified representation on the Caltech101-20 dataset. As shown in Figure 5, with an increase in the number of epochs, the learned representation becomes more condensed and distinctive.
Additionally, the following Table 6 presents the number of iterations and runtime of the model on five datasets, further illustrating the speed and efficiency of the proposed model in this paper.

5. Discussion

The VMC-CD method effectively addresses the challenge of multi-view clustering by balancing consistency and diversity of information. It not only advances multi-view clustering techniques but also aligns with trends in deep learning and data clustering research. Emphasizing a view-driven approach and dual contrastive learning, it improves clustering performance and feature alignment. Future directions may include exploring dynamic dataset handling and high-dimensional data applications. VMC-CD represents significant progress in multi-view clustering, inspiring research in deep learning and data clustering.

6. Conclusions

This paper introduces a view-driven dual-contrastive learning approach for multi-view clustering. This method involves incorporating relevant information from other views during the feature representation learning phase, promoting view diversity, and facilitating consensus feature learning. The concept of dual-contrastive learning is introduced, which promotes view consistency from both the clustering level and the feature level, complementing each other.

Author Contributions

Conceptualization, S.L.; Methodology, S.L.; Software, S.L.; Validation, S.L.; Formal analysis, S.L.; Investigation, S.L.; Resources, S.L.; Data curation, S.L.; Writing—original draft, S.L.; Writing—review & editing, S.L.; Visualization, S.L.; Supervision, C.Z., Z.L., Z.Y. and W.G.; Project administration, S.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by National Natural Science Foundation of China (CN) [62276164, 61602296], ‘Science and technology innovation action plan’ Natural Science Foundation of Shanghai [22ZR1427000], and Shanghai Oriental Talent Program-Youth Program. The authors would like to thank their supports.

Data Availability Statement

The data presented in this study are available on request from the author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Chen, J.; Yang, S.; Peng, X.; Peng, D.; Wang, Z. Augmented sparse representation for incomplete multiview clustering. IEEE Trans. Neural Netw. Learn. Syst. 2022, 35, 4058–4071. [Google Scholar] [CrossRef] [PubMed]
  2. Xu, J.; Ren, Y.; Tang, H.; Yang, Z.; Pan, L.; Yang, Y.; Pu, X.; Philip, S.Y.; He, L. Self-supervised discriminative feature learning for deep multi-view clustering. IEEE Trans. Knowl. Data Eng. 2022, 35, 7470–7482. [Google Scholar] [CrossRef]
  3. Hu, P.; Peng, D.; Sang, Y.; Xiang, Y. Multi-view linear discriminant analysis network. IEEE Trans. Image Process. 2019, 28, 5352–5365. [Google Scholar] [CrossRef] [PubMed]
  4. Kang, Z.; Zhao, X.; Peng, C.; Zhu, H.; Zhou, J.T.; Peng, X.; Chen, W.; Xu, Z. Partition level multiview subspace clustering. Neural Netw. 2020, 122, 279–288. [Google Scholar] [CrossRef] [PubMed]
  5. Li, R.; Zhang, C.; Fu, H.; Peng, X.; Zhou, T.; Hu, Q. Reciprocal multi-layer subspace learning for multi-view clustering. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Repulic of Korea, 27 October–2 November 2019; pp. 8172–8180. [Google Scholar]
  6. Zhou, R.; Shen, Y.D. End-to-end adversarial-attention network for multi-modal clustering. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 14619–14628. [Google Scholar]
  7. Baltrušaitis, T.; Ahuja, C.; Morency, L.P. Multimodal machine learning: A survey and taxonomy. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 41, 423–443. [Google Scholar] [CrossRef] [PubMed]
  8. Tang, C.; Li, Z.; Wang, J.; Liu, X.; Zhang, W.; Zhu, E. Unified one-step multi-view spectral clustering. IEEE Trans. Knowl. Data Eng. 2022, 35, 6449–6460. [Google Scholar] [CrossRef]
  9. Han, Z.; Zhang, C.; Fu, H.; Zhou, J.T. Trusted multi-view classification with dynamic evidential fusion. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 2551–2566. [Google Scholar] [CrossRef]
  10. Zhang, P.; Liu, X.; Xiong, J.; Zhou, S.; Zhao, W.; Zhu, E.; Cai, Z. Consensus one-step multi-view subspace clustering. IEEE Trans. Knowl. Data Eng. 2020, 34, 4676–4689. [Google Scholar] [CrossRef]
  11. Chen, J.; Yang, S.; Mao, H.; Fahy, C. Multiview subspace clustering using low-rank representation. IEEE Trans. Cybern. 2021, 52, 12364–12378. [Google Scholar] [CrossRef] [PubMed]
  12. Tao, Z.; Li, J.; Fu, H.; Kong, Y.; Fu, Y. From ensemble clustering to subspace clustering: Cluster structure encoding. IEEE Trans. Neural Netw. Learn. Syst. 2021, 34, 2670–2681. [Google Scholar] [CrossRef]
  13. Zhao, W.; Xu, C.; Guan, Z.; Liu, Y. Multiview concept learning via deep matrix factorization. IEEE Trans. Neural Netw. Learn. Syst. 2020, 32, 814–825. [Google Scholar] [CrossRef] [PubMed]
  14. Hu, M.; Chen, S. One-pass incomplete multi-view clustering. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; Volume 33, pp. 3838–3845. [Google Scholar]
  15. Li, L.; Wan, Z.; He, H. Incomplete multi-view clustering with joint partition and graph learning. IEEE Trans. Knowl. Data Eng. 2021, 35, 589–602. [Google Scholar] [CrossRef]
  16. Liu, X.; Li, M.; Tang, C.; Xia, J.; Xiong, J.; Liu, L.; Kloft, M.; Zhu, E. Efficient and effective regularized incomplete multi-view clustering. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 43, 2634–2646. [Google Scholar] [CrossRef] [PubMed]
  17. Wang, Y.; Chang, D.; Fu, Z.; Wen, J.; Zhao, Y. Graph contrastive partial multi-view clustering. IEEE Trans. Multimed. 2022, 25, 6551–6562. [Google Scholar] [CrossRef]
  18. Wang, Q.; Tao, Z.; Xia, W.; Gao, Q.; Cao, X.; Jiao, L. Adversarial multiview clustering networks with adaptive fusion. IEEE Trans. Neural Netw. Learn. Syst. 2022, 34, 7635–7647. [Google Scholar] [CrossRef] [PubMed]
  19. Peng, X.; Li, Y.; Tsang, I.W.; Zhu, H.; Lv, J.; Zhou, J.T. XAI beyond classification: Interpretable neural clustering. J. Mach. Learn. Res. 2022, 23, 1–28. [Google Scholar]
  20. Yang, M.; Li, Y.; Hu, P.; Bai, J.; Lv, J.; Peng, X. Robust multi-view clustering with incomplete information. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 1055–1069. [Google Scholar] [CrossRef] [PubMed]
  21. Li, Z.; Wang, Q.; Tao, Z.; Gao, Q.; Yang, Z. Deep adversarial multi-view clustering network. Proc. Ijcai 2019, 2, 4. [Google Scholar]
  22. Xie, J.; Girshick, R.; Farhadi, A. Unsupervised deep embedding for clustering analysis. In Proceedings of the International Conference on Machine Learning, PMLR, New York, NY, USA, 19–24 June 2016; pp. 478–487. [Google Scholar]
  23. Li, Y.; Hu, P.; Liu, Z.; Peng, D.; Zhou, J.T.; Peng, X. Contrastive clustering. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtually, 2–9 February 2021; Volume 35, pp. 8547–8555. [Google Scholar]
  24. Chen, T.; Kornblith, S.; Norouzi, M.; Hinton, G. A simple framework for contrastive learning of visual representations. In Proceedings of the International Conference on Machine Learning, PMLR, Virtual Event, 13–18 July 2020; pp. 1597–1607. [Google Scholar]
  25. Xu, J.; Tang, H.; Ren, Y.; Peng, L.; Zhu, X.; He, L. Multi-level feature learning for contrastive multi-view clustering. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 16051–16060. [Google Scholar]
  26. Yang, M.; Huang, Z.; Hu, P.; Li, T.; Lv, J.; Peng, X. Learning with twin noisy labels for visible-infrared person re-identification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 14308–14317. [Google Scholar]
  27. Andrew, G.; Arora, R.; Bilmes, J.; Livescu, K. Deep canonical correlation analysis. In Proceedings of the International Conference on Machine Learning, PMLR, Atlanta, GA, USA, 17–19 June 2013; pp. 1247–1255. [Google Scholar]
  28. Cao, X.; Zhang, C.; Fu, H.; Liu, S.; Zhang, H. Diversity-induced multi-view subspace clustering. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 586–594. [Google Scholar]
  29. Jiang, G.; Peng, J.; Wang, H.; Mi, Z.; Fu, X. Tensorial multi-view clustering via low-rank constrained high-order graph learning. IEEE Trans. Circuits Syst. Video Technol. 2022, 32, 5307–5318. [Google Scholar] [CrossRef]
  30. Wang, H.; Yao, M.; Jiang, G.; Mi, Z.; Fu, X. Graph-collaborated auto-encoder hashing for multiview binary clustering. IEEE Trans. Neural Netw. Learn. Syst. 2023, 1–13. [Google Scholar] [CrossRef] [PubMed]
  31. Liu, J.; Wang, C.; Gao, J.; Han, J. Multi-view clustering via joint nonnegative matrix factorization. In Proceedings of the 2013 SIAM International Conference on Data Mining, SIAM, Austin, TX, USA, 2–4 May 2013; pp. 252–260. [Google Scholar]
  32. Zhao, H.; Ding, Z.; Fu, Y. Multi-view clustering via deep matrix factorization. In Proceedings of the AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017; Volume 31. [Google Scholar]
  33. Yang, Z.; Xu, Q.; Zhang, W.; Cao, X.; Huang, Q. Split multiplicative multi-view subspace clustering. IEEE Trans. Image Process. 2019, 28, 5147–5160. [Google Scholar] [CrossRef] [PubMed]
  34. Blaschko, M.B.; Lampert, C.H. Correlational spectral clustering. In Proceedings of the 2008 IEEE Conference on Computer Vision and Pattern Recognition, Anchorage, AK, USA, 23–28 June 2008; pp. 1–8. [Google Scholar]
  35. Chaudhuri, K.; Kakade, S.M.; Livescu, K.; Sridharan, K. Multi-view clustering via canonical correlation analysis. In Proceedings of the 26th Annual International Conference on Machine Learning, Montreal, QC, Canada, 14–18 June 2009; pp. 129–136. [Google Scholar]
  36. Wang, H.; Wang, Y.; Zhang, Z.; Fu, X.; Zhuo, L.; Xu, M.; Wang, M. Kernelized multiview subspace analysis by self-weighted learning. IEEE Trans. Multimed. 2020, 23, 3828–3840. [Google Scholar] [CrossRef]
  37. Nie, F.; Li, J.; Li, X. Self-weighted multiview clustering with multiple graphs. In Proceedings of the IJCAI, Melbourne, Australia, 19–25 August 2017; pp. 2564–2570. [Google Scholar]
  38. Tao, Z.; Liu, H.; Li, S.; Ding, Z.; Fu, Y. Marginalized multiview ensemble clustering. IEEE Trans. Neural Netw. Learn. Syst. 2019, 31, 600–611. [Google Scholar] [CrossRef] [PubMed]
  39. Fan, S.; Wang, X.; Shi, C.; Lu, E.; Lin, K.; Wang, B. One2multi graph autoencoder for multi-view graph clustering. In Proceedings of the Web Conference 2020, Taipei, Taiwan, 20–24 April 2020; pp. 3070–3076. [Google Scholar]
  40. Wang, H.; Jiang, G.; Peng, J.; Deng, R.; Fu, X. Towards adaptive consensus graph: Multi-view clustering via graph collaboration. IEEE Trans. Multimed. 2022, 25, 6629–6641. [Google Scholar] [CrossRef]
  41. Li, M.; Liu, X.; Wang, L.; Dou, Y.; Yin, J.; Zhu, E. Multiple Kernel Clustering with Local Kernel Alignment Maximization; AAAI Press: Washington, DC, USA, 2016. [Google Scholar]
  42. Wang, Y.; Liu, X.; Dou, Y.; Lv, Q.; Lu, Y. Multiple kernel learning with hybrid kernel alignment maximization. Pattern Recognit. 2017, 70, 104–111. [Google Scholar] [CrossRef]
  43. Abavisani, M.; Patel, V.M. Deep multimodal subspace clustering networks. IEEE J. Sel. Top. Signal Process. 2018, 12, 1601–1614. [Google Scholar] [CrossRef]
  44. Zhu, P.; Hui, B.; Zhang, C.; Du, D.; Wen, L.; Hu, Q. Multi-view deep subspace clustering networks. arXiv 2019, arXiv:1908.01978. [Google Scholar]
  45. Hadsell, R.; Chopra, S.; LeCun, Y. Dimensionality reduction by learning an invariant mapping. In Proceedings of the 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’06), New York, NY, USA, 17–22 June 2006; Volume 2, pp. 1735–1742. [Google Scholar]
  46. Lin, Y.; Gou, Y.; Liu, Z.; Li, B.; Lv, J.; Peng, X. Completer: Incomplete multi-view clustering via contrastive prediction. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 11174–11183. [Google Scholar]
  47. Li, Y.; Nie, F.; Huang, H.; Huang, J. Large-scale multi-view spectral clustering via bipartite graph. In Proceedings of the AAAI Conference on Artificial Intelligence, Austin, TX, USA, 25–30 January 2015; Volume 29. [Google Scholar]
  48. Fei-Fei, L.; Perona, P. A bayesian hierarchical model for learning natural scene categories. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05), San Diego, CA, USA, 20–25 June 2005; Volume 2, pp. 524–531. [Google Scholar]
  49. Yang, Y.; Newsam, S. Bag-of-visual-words and spatial extensions for land-use classification. In Proceedings of the 18th SIGSPATIAL International Conference on Advances in Geographic Information Systems, San Jose, CA, USA, 2–5 November 2010; pp. 270–279. [Google Scholar]
  50. LeCun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef]
  51. Hull, J.J. A database for handwritten text recognition research. IEEE Trans. Pattern Anal. Mach. Intell. 1994, 16, 550–554. [Google Scholar] [CrossRef]
  52. Cai, X.; Wang, H.; Huang, H.; Ding, C. Joint stage recognition and anatomical annotation of drosophila gene expression patterns. Bioinformatics 2012, 28, i16–i24. [Google Scholar] [CrossRef]
  53. Zhao, H.; Liu, H.; Fu, Y. Incomplete multi-modal visual data grouping. In Proceedings of the IJCAI, New York, NY, USA, 9–16 July 2016; pp. 2392–2398. [Google Scholar]
  54. Hu, M.; Chen, S. Doubly aligned incomplete multi-view clustering. arXiv 2019, arXiv:1903.02785. [Google Scholar]
  55. Wen, J.; Zhang, Z.; Xu, Y.; Zhang, B.; Fei, L.; Liu, H. Unified embedding alignment with missing views inferring for incomplete multi-view clustering. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; Volume 33, pp. 5393–5400. [Google Scholar]
  56. Wang, W.; Arora, R.; Livescu, K.; Bilmes, J. On deep multi-view representation learning. In Proceedings of the International Conference on Machine Learning, PMLR, Lille, France, 7–9 July 2015; pp. 1083–1092. [Google Scholar]
  57. Li, S.Y.; Jiang, Y.; Zhou, Z.H. Partial multi-view clustering. In Proceedings of the AAAI Conference on Artificial Intelligence, Québec, QC, Canada, 27–31 July 2014; Volume 28. [Google Scholar]
  58. Zhang, C.; Liu, Y.; Fu, H. Ae2-nets: Autoencoder in autoencoder networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 2577–2585. [Google Scholar]
  59. Wang, H.; Zong, L.; Liu, B.; Yang, Y.; Zhou, W. Spectral perturbation meets incomplete multi-view data. arXiv 2019, arXiv:1906.00098. [Google Scholar]
  60. Ma, Z.; Yu, J.; Wang, L.; Chen, H.; Zhao, Y.; He, X.; Wang, Y.; Song, Y. Multi-view clustering based on view-attention driven. Int. J. Mach. Learn. Cybern. 2023, 14, 2621–2631. [Google Scholar] [CrossRef]
  61. Guo, X.; Gao, L.; Liu, X.; Yin, J. Improved deep embedded clustering with local structure preservation. Proc. IJCAI 2017, 17, 1753–1759. [Google Scholar]
  62. Zhang, Z.; Liu, L.; Shen, F.; Shen, H.T.; Shao, L. Binary multi-view clustering. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 41, 1774–1782. [Google Scholar] [CrossRef] [PubMed]
  63. Wang, S.; Liu, X.; Zhu, E.; Tang, C.; Liu, J.; Hu, J.; Xia, J.; Yin, J. Multi-view Clustering via Late Fusion Alignment Maximization. In Proceedings of the IJCAI, Macao, China, 10–16 August 2019; pp. 3778–3784. [Google Scholar]
  64. Ren, Y.; Huang, S.; Zhao, P.; Han, M.; Xu, Z. Self-paced and auto-weighted multi-view clustering. Neurocomputing 2020, 383, 248–256. [Google Scholar] [CrossRef]
  65. Wen, J.; Wu, Z.; Zhang, Z.; Fei, L.; Zhang, B.; Xu, Y. Structural deep incomplete multi-view clustering network. In Proceedings of the 30th ACM International Conference on Information & Knowledge Management, Queensland, Australia, 1–5 November 2021; pp. 3538–3542. [Google Scholar]
  66. Trosten, D.J.; Lokse, S.; Jenssen, R.; Kampffmeyer, M. Reconsidering representation alignment for multi-view clustering. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 1255–1265. [Google Scholar]
  67. Van der Maaten, L.; Hinton, G. Visualizing data using t-SNE. J. Mach. Learn. Res. 2008, 9, 2579–2605. [Google Scholar]
Figure 1. Architecture of the view-driven dual contrastive learning for multi-view clustering.
Figure 1. Architecture of the view-driven dual contrastive learning for multi-view clustering.
Entropy 26 00470 g001
Figure 2. Architecture of the attention weight learning module.
Figure 2. Architecture of the attention weight learning module.
Entropy 26 00470 g002
Figure 3. Sensitivity analysis of parameters for Caltech101-20.
Figure 3. Sensitivity analysis of parameters for Caltech101-20.
Entropy 26 00470 g003
Figure 4. Change curve of clustering metrics with iteration count on Scene-15 dataset.
Figure 4. Change curve of clustering metrics with iteration count on Scene-15 dataset.
Entropy 26 00470 g004
Figure 5. Sensitivity analysis of parameters for Caltech101-20.
Figure 5. Sensitivity analysis of parameters for Caltech101-20.
Entropy 26 00470 g005
Table 1. Clustering performance on Caltech101-20, Scene-15, and LandUse-21 datasets.
Table 1. Clustering performance on Caltech101-20, Scene-15, and LandUse-21 datasets.
MethodsCaltech101-20Scene-15LandUse-21
ACC NMI ARI ACC NMI ARI ACC NMI ARI
IMG44.5161.3535.7424.2025.649.5716.4027.115.10
EERIMVC43.2855.0430.4239.6038.9922.0624.9229.5712.24
DAIMC45.4861.7932.4032.0933.5517.4224.3529.3510.26
UEAF47.4057.9038.9834.3736.6918.5223.0027.058.79
DCCAE44.0559.1234.5636.4439.7821.4715.6224.414.42
PVC44.9162.1335.7730.8331.0514.9825.2230.4511.72
AE2-Nets49.1065.3835.6636.1040.3922.0824.7930.3610.35
DCCA41.8959.1433.3936.1838.9220.8715.5123.154.43
PICCAE62.2767.9351.5638.7240.4622.1224.8629.7410.48
COMPLETER70.1868.0677.8841.0744.6824.7825.6331.7313.05
ATTENTION74.8871.2586.4541.9344.0825.1026.6831.8913.64
Ours77.7073.1192.0444.7745.6626.9127.7633.9313.88
Table 2. Clustering performance on BDGP and MNIST-USPS datasets.
Table 2. Clustering performance on BDGP and MNIST-USPS datasets.
MethodsMNIST-USPSBDGP
ACC NMI ARI ACC NMI ARI
DEC73.1071.4663.2394.7886.6287.02
IDEC76.5876.8968.0195.9689.4090.25
BMVC88.0289.4584.4834.9212.028.33
MVC-LFA76.7867.4960.9254.6833.4528.81
DAMC71.7280.8569.8098.2294.6194.37
SAMVC69.6574.5860.9053.8646.2520.99
CDIMC-net62.0367.6363.3888.2778.9381.94
EAMC73.0483.5372.1567.5647.0239.31
SiMVC97.7496.3095.2869.7253.2644.55
CoMVC98.4797.3598.0180.6867.3959.28
Ours99.4698.4698.8099.0096.8297.53
Table 3. Ablative study of main components of the proposed VMC-CD method on Caltech101-20 datasets.
Table 3. Ablative study of main components of the proposed VMC-CD method on Caltech101-20 datasets.
c l c h r e c ACC (%)NMI (%)ARI (%)
-68.4868.1185.20
-52.8556.7651.62
77.7073.1192.04
Table 4. Ablative study of main components of the proposed VMC-CD method on Scene-15 datasets.
Table 4. Ablative study of main components of the proposed VMC-CD method on Scene-15 datasets.
c l c h r e c ACC (%)NMI (%)ARI (%)
-43.7945.0926.64
-41.3640.0623.46
44.7745.6626.91
Table 5. Ablative study of main components of the proposed VMC-CD method on LandUse-21 datasets.
Table 5. Ablative study of main components of the proposed VMC-CD method on LandUse-21 datasets.
c l c h r e c ACC (%)NMI (%)ARI (%)
-27.6731.1213.51
-25.3829.1811.97
27.7633.9313.88
Table 6. Number of iterations and runtime on the aforementioned five datasets.
Table 6. Number of iterations and runtime on the aforementioned five datasets.
DatasetIterations (epochs)Running Time (s)
Caltech101-2020049.69
Scene-15500206.85
LandUse-21700148.73
MNIST-UPS20071.39
BDGP20038.97
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liu, S.; Zhu, C.; Li, Z.; Yang, Z.; Gu, W. View-Driven Multi-View Clustering via Contrastive Double-Learning. Entropy 2024, 26, 470. https://doi.org/10.3390/e26060470

AMA Style

Liu S, Zhu C, Li Z, Yang Z, Gu W. View-Driven Multi-View Clustering via Contrastive Double-Learning. Entropy. 2024; 26(6):470. https://doi.org/10.3390/e26060470

Chicago/Turabian Style

Liu, Shengcheng, Changming Zhu, Zishi Li, Zhiyuan Yang, and Wenjie Gu. 2024. "View-Driven Multi-View Clustering via Contrastive Double-Learning" Entropy 26, no. 6: 470. https://doi.org/10.3390/e26060470

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop