Next Article in Journal
A Review of Model Predictive Control for Grid-Connected PV Applications
Previous Article in Journal
DeepGuard: Identification and Attribution of AI-Generated Synthetic Images
Previous Article in Special Issue
How Self-Regulated Learning Is Affected by Feedback Based on Large Language Models: Data-Driven Sustainable Development in Computer Programming Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Optimal-Transport-Based Multimodal Big Data Clustering

1
College of Information, Shenyang Institute of Engineering, Shenyang 110136, China
2
School of Computer Science, Beijing Institute of Technology, Beijing 100081, China
*
Author to whom correspondence should be addressed.
Electronics 2025, 14(4), 666; https://doi.org/10.3390/electronics14040666
Submission received: 24 December 2024 / Revised: 28 January 2025 / Accepted: 29 January 2025 / Published: 8 February 2025
(This article belongs to the Special Issue Advances in Data-Driven Artificial Intelligence)

Abstract

:
Multimodal clustering achieves outstanding performance in various applications by aggregating information from heterogeneous devices. However, previous methods rely on strong-notion distances to fuse crossmodal complementary knowledge, established on a fragile assumption about the existence of a ubiquitous non-negligible intersection between heterogeneous manifolds of modalities. Due to this unstable theoretical basis, previous methods are essentially challenged by limited performance on general multimodal data. To address this challenge, an optimal-transport-based multimodal clustering (OTMC) method is defined as the optimal transport (OT) from multimodal data distributions to clustering distributions, which leverages a weak-topology measure to capture complementary knowledge with clear discriminative structures. OTMC consists of a modality-specific OT delivering private structures and a modality-common OT delivering shared structures, which transports category structures scattered in manifolds of each modality and all modalities to common prototypes, respectively. Furthermore, variational solutions to OTMC are derived by matching the data-prototype joint distribution, which induces the multimodal OT clustering network, to capture discriminative structures. Finally, the experimental results from four real-world datasets demonstrate the superiority of OTMC, helped by never relying on the phantom of heterogeneous manifold intersections. In particular, OTMC obtains 92.15% ACC, 84.96% NMI, and 83.35% ARI on Handwritten, improving by 2.25%, 2.82%, and 3.28%, respectively.

1. Introduction

With the rapid development and widespread applications of emerging computing technologies, data produced by human beings are experiencing explosive growth [1]. For instance, a smart power plant acquires real-time data through various devices, such as flow sensors, current transducers, and power meters, to monitor the working status of power equipment, in which an enormous amount of data with heterogeneous distributions are collected. In smart cities, smart car-detection technology utilizes multiple cameras and microphones to verify the categories of cars, amassing a great volume of information about cars. These extensive data collected from multiple devices are often referred to as multimodal data, in which the data of each modality contain consistent information that describes shared properties and complementary information about different properties among modalities [2,3,4]. The fusion of complementary and consistent information in multimodal data can further facilitate the understanding of patterns hidden in complex phenomena of the real world [5]. Thus, it is of great importance to design innovative computing paradigms to integrate information from multimodal data.
Multimodal clustering (MC), as a fundamental unsupervised task, aims to aggregate knowledge scattered in heterogeneous modalities to recognize intrinsic patterns of data [6,7,8]. Recently, some pioneering deep multimodal clustering methods have been proposed to leverage hierarchical nonlinear embedding features to explore patterns from fusion information, which can be roughly separated into two classes, i.e., deep fusion methods [9,10] and deep clustering methods [11,12]. The former utilizes self-supervision information of data to extract generalized fusion representations to characterize fusion information, and then performs vanilla single-modality clustering methods to mine patterns, whose two-stage learning strategy may disconnect fusion representation learning and clustering partition, causing suboptimal fusion representations for clustering partitioning. To close the gap between the two processes, deep clustering methods combine fusion representation learning and clustering partition into a joint optimization strategy. They utilize the divergences of data structures to guide the fusion of complementary information among modalities, as well as mining clustering structures in an end-to-end manner.
Existing MC methods have made significant progress in the pattern mining of multimodal data. However, most of them ignore the manifold topology of multimodal data, and focus on distance metrics with strong notions that assume strict overlaps between heterogeneous distributions of data in various modalities, such as Kullback–Leibler (KL) divergence and Jensen–Shannon (JS) divergence, leading to the degradation of performance on pattern mining when they are faced with the challenge of general multimodal data applications [13]. Specifically, general multimodal data collected from diverse sources obey heterogeneous distributions and are embedded in their corresponding manifolds, and the manifolds of data in each modality almost have no non-negligible intersection [14]. In that case, the assumption above is not valid everywhere, which leads to failure in measuring the differences between heterogeneous manifolds when integrating complementary information from multimodal data. In other words, distance metrics with strong notions provide no useful gradient information in training deep clustering models, which blurs discriminative structures when fusing representations of multimodal data. Thus, previous methods may not exhibit desirable performance in pattern mining in real-world applications.
To address this challenge, an optimal-transport-based multimodal clustering method (OTMC) is proposed to conduct multimodal clustering structure mining based on the optimal-transport (OT) theory, which ensures clustering-friendly discriminative structures when fusing complementary modality information with heterogeneous distributions. Specifically, multimodal clustering is defined as a map from multimodal data measures to clustering prototype set measures in OTMC, which utilizes the Wasserstein distance to obtain clear discriminative structures from heterogeneous information with no intersection of data distributions in different modalities. Furthermore, OTMC is decomposed into a modality-specific OT (MsOT) and a modality-common OT (McOT) within a consistent constraint, to disentangle the transportation of private and shared information of each modality. In addition, a generative transport plan is designed to derive the variational solution to OT, which can effectively search the map with a minimal transport cost for mining intrinsic patterns. Finally, although designed for all data with heterogeneous distributions in different modalities, e.g., text and video, OTMC is conducted on four real-world benchmark image datasets in extensive experiments to validate its superiority on high-dimension data.
The main contributions of this paper are as follows:
  • An innovative weak-notion distance metric-based method is designed to measure differences between the manifold structures of data collected from diverse devices, which ensures the full fusion of complementary information from data with heterogeneous distributions.
  • Multimodal clustering is innovatively modeled using an optimal-transport-based multimodal clustering method (OTMC), which can capture fusion information with clear discriminative structures from heterogeneous modalities for mining intrinsic patterns.
  • A variational solution is derived to solve OT based on a generative transport plan, which can precisely match the transport map for transporting the multimodal data to clustering prototypes in OTMC.
  • Extensive experiments are conducted on four real-world benchmark datasets, which verify the superiority of OTMC compared with other methods in multimodal clustering, helped by never relying on the phantom of heterogeneous manifold intersections. In particular, OTMC obtains 92.15% ACC, 84.96% NMI, and 83.35% ARI on Handwritten, improving by 2.25%, 2.82%, and 3.28%, respectively.
In the rest of this paper, Section 2 gives the related deep fusion methods and deep clustering methods; Section 3 introduces the theoretical framework of OTMC; Section 4 derives the variational generative solution to OTMC and network implementations; Section 5 showcases the experimental results of OTMC in comparison with deep fusion methods and deep clustering methods; and Section 6 concludes this work and points out some possible future research directions for subsequent exploration.

2. Related Work

Clustering is a classical unsupervised learning task. Various effective clustering algorithms like k-means have been proposed and widely utilized as cornerstones of the solutions to problems such as pavement deterioration [15], facility location [16], and the energy management of wireless sensor networks [17]. OTMC learns intrinsic patterns of multimodal data via neural networks, and is closely related to deep fusion methods and deep clustering methods in multimodal clustering.
Deep fusion methods, as two-stage methods, focus on extracting generalized fusion representations of multimodal data, and then leveraging a single-modality clustering algorithm to mine clustering patterns. For example, DCCA [18,19] learned fusion representations of multimodal data via hierarchical nonlinear mappings along with crossmodal correlation maximizing. DMF [20] mined the consensus semantics of multimodal data by leveraging multi-layer transformations subject to non-negative constraints to explore informative spatial bases. AE2-Nets [9] modeled a degradation process from multimodal data to fusion representations within a nested autoencoder architecture to capture consistent and complementary information among modalities. CMIB [10] utilized the rate-distortion principle to explore robust fusion representations by preserving complementary and consistent information and discarding the superfluous information of modalities. MIFN-ML [21] conducted intermediate fusion of biometric and facial representations after manifold-learning-based dimension reduction techniques, which then made a clustering decision for stress detection via a nonlinear mapping. MCBVAD [22] introduced attentions from other views to suppress information from irrelevant views on the basis of maximizing information entropy and mutual information of views, which produced information-fruitful representations to improve k-means clustering results. Further, with the help of a Transformer-based attention learning block, VMC-CD [23] applied cross-view contrastive learning and maximum entropy of categories to obtain discriminative representations for k-means clustering. OSCAMC [24] jointly learned and optimized view-collective matrices, including a combined non-negative embedding matrix, a collective similarity matrix, a joint coefficient matrix, a unified spectral projection matrix, and so on, and then obtained the clustering results from those matrices.
Deep clustering methods integrate complementary information fusion with clustering pattern recognition in an end-to-end architecture, where divergences of data structures are used to supervise model training. MVaDE [25] modeled the generation process from hidden categories to visible data, where the Gaussian distribution was used to constrain the information fusion of each modality. CoMVC [26] utilized crossmodal contrastive learning followed by adaptive linear combination to fuse modality information, and then conducted Cauchy–Schwarz divergence in the fusion representation space for clustering. SDMVC [27] utilized KL-divergence to mine clustering patterns from concatenated complementary information, which uses consistent semantics to constrain the alignments of representation distributions in each modality. CMSC [28] fused modality information on the basis of a consensus subspace which maximized the correlations of modalities, and then performed spectral clustering to mine clustering patterns. DMIM [29] fused complementary information by maximizing the dependence between modalities and mined multimodal data patterns with the help of an over-clustering strategy. DMAC-SI [5] combined invariant semantics representation learning and the Markov decision process of multimodal data clustering into an end-to-end clustering framework, ensuring the flexibility of clustering structure exploration. EMVFC [30] integrated low-dimensional representation learning and cooperative learning in clustering to improve performance, which reduced redundancy features and noise, enhancing the complementary information and correlation.
The deep fusion methods above extracted generalized fusion representations and then performed vanilla single-modality clustering methods to mine patterns, which may disconnect fusion representation learning and clustering partition, causing suboptimal fusion representations for clustering partition. The deep clustering methods above combined fusion representation learning and clustering partition into a joint optimization strategy for mining clustering structures in an end-to-end manner. However, they still relied on strong-notion distances to fuse crossmodal complementary knowledge, which limited their clustering performance. In contrast, optimal transport can define multimodal clustering as a map from multimodal data measures to clustering prototype set measures, which avoids representation-partition disconnections and dependency on the intersection of data distributions in different modalities.
Through the comparison between optimal transport and clustering, it can be seen that a difference between optimal transport and clustering is that optimal transport ex-changes the operation form for the lowest cost. That is, for any two mappings that map data points to cluster centers, optimal transport chooses the one with lowest cost and does not consider how it works. In a sense, optimal transport is more flexible and difficult than clustering. Thus, the solution to optimal transport can be used in clustering. In multimodal clustering, dividing the data into modality-specific and modality-common components is a standard approach. OTMC further distinguishes itself by incorporating the concept of optimal transport into this framework. Specifically, OTMC extends the rationale for optimal transport by partitioning it into modality-common and modality-specific components as well. This allows OTMC to better capture the intrinsic relationships both within and across modalities, offering a more nuanced representation of the data. In addition, there are diverse opinions about how to evaluate the quality of a clustering algorithm so that many metrics are proposed. If the objective of clustering is just minimizing the sum of distances in each cluster in k-means, no more problems appear. So, in this context, our method is a beneficial step toward providing a new direction for clustering, i.e., evaluating clustering quality based on transport cost.

3. OT for Multimodal Clustering

In this section, multimodal clustering is redefined within the optimal-transport architecture, to mine patterns of multimodal data by adaptively learning an optimal transport map between multimodal data and clustering prototypes. Then, rooted in private and common information hidden in multimodal data, OTMC is decomposed into modality-specific OT and modality-common OT, as shown in Figure 1. In addition, the most frequently used notations are listed in Table 1.

3.1. Modeling Multimodal Clustering Based on OT

Optimal transport utilizes a weak topology measure to estimate the similarity between two manifolds defined over a space by computing the cost of moving one point to another within the space. Thus, optimal-transport-based multimodal clustering (OTMC) is proposed to aggregate complementary and consistent information in modalities with heterogeneous manifolds, which provides representations with clear discriminate structures without the help of intersections between data manifolds of different modalities. Specifically, let X = { X m } m = 1 M denote the measurable space of a multimodal instance with measure μ , where X m is the m-th modality space in M modalities, and let C = { c j } j = 1 k denote the prototype set with cluster number k and prototype measure ν , in which c j denotes the j -th clustering prototype. Multimodal clustering T groups data into corresponding clustering prototypes with minimal pre-defined criteria as follows:
T arg min W j = 1 k W j d ( x , c j ) d μ ( x ) ,
where W = { ( W j , c j ) j = 1 k } stands for a clustering scheme and W j stands for the data partition belonging to the prototype c j .
Multimodal clustering groups data into corresponding clusters by minimizing discrepancies between data and prototypes, which is natural and consistent with the optimal transport that finds an optimal-transport map from data to the corresponding clusters, as follows:
T arg min W j = 1 k W j d ( x , c j ) d μ ( x ) arg min T # μ = ν X d ( x , T ( x ) ) d μ ( x ) ,
where d , measures the transport map cost between data and prototypes, and ν = j = 1 k ν j δ ( c c j ) stands for the prototype measure based on the Dirac function δ ( ) , in which ν j stands for the j -th prototype mass. Meanwhile, T # μ = ν indicates that the overall data mass transported to the prototype c j is the same as the c j mass, that is, μ ( T 1 ( c j ) ) = ν j . The cluster partition W is equivalent to the map T when the prototype measure ν is fixed. That is, the map T projects instances of cluster W j to the prototype c j , i.e., W j = { x i | T ( x i ) = c j , x i X } .

3.2. Decomposition of OTMC

In multimodal data, each modality of the same instance utilizes heterogeneous structures to deliver the information from different aspects. In other words, the data of each modality contain both private and common information. Thus, the optimal-transport-based multimodal clustering method is decomposed into the modality-specific OT (MsOT) that captures the private information, and the modality-common OT (McOT) that catches the common information.
Modality-specific OT T s m ( m = 1 , , M ) aims to capture the m -th modality private information that contributes to recognizing inherent patterns in each modality. Specifically, T s m represents the map from the data of the m -th modality space X m to the k prototype set C s m with measure-preserving T s m μ s m # = ν s m , which pushes forward μ s m to ν s m , i.e., T s m : X m C s m . μ s m and ν s m denote the measure of the space X m and the set C s m , respectively. T s m can be obtained by minimizing the following total cost based on Equation (2):
MsOT = arg min T s m μ s m # = ν s m X m d s m ( x , T s m ( x ) ) d μ s m ( x ) , m = 1 , , M ,
where d s m : X m × C s m + is the transport map cost between data and prototypes in the m -th modality. The modality-specific OT explores manifold differences between modality-specific data and clustering prototypes by utilizing a measure with weak topology, to learn an optimal transform map from a local perspective, which effectively captures the private information for mining patterns in each modality.
Modality-common OT T c catches common information that reveals the consistent semantics of instances among modalities. To be specific, T c maps data in multimodal space X to the modality-common prototype set C c with measure ν c , which means the map T c pushes forward μ to ν c . T c can be obtained by minimizing the cost based on Equation (2):
McOT = arg min T c# μ = ν c X d c ( x , T c ( x ) ) d μ ( x ) ,
where d c : X × C c + is the transport map cost of modality-common OT T c . Modality-common OT explores manifold differences between multimodal data and modality-common clustering prototypes by utilizing a measure with weak topology to learn an optimal transform map from a global perspective, which effectively captures the common semantic information for mining patterns in multimodal data.

3.3. The Consistent Constraint

Since both modality-specific OT and modality-common OT are used to mine the intrinsic clustering patterns of multimodal data, their transport plan T s m and T c should be consistent. To this end, a consistent constraint L c o n between ν s m and ν c is introduced to achieve the above objective by keeping the clustering prototypes of modality-specific OT and modality-common OT subject to the same measure:
L c o n = min m = 1 M D ( ν s m , ν c ) ,
where D ( ) denotes a metric function that evaluates differences between the measures ν s m and ν c . Finally, modality-common OT T c is used for the cluster assignment result.

4. The Variational Generative Solution Network Implementation

In this section, a variational generative solution to OT is derived, and then, it is generalized to multimodal scenarios for optimizing data partitioning within an optimal-transport clustering network.

4.1. Variational Generative Solution

The variational generative solution learns the generative relationship between data and prototypes by constructing the joint distribution p x , c , which provides a theoretically accurate analytical solution to OT. Specifically, p x , c = p c | x p x , where p x stands for the data distribution and p c | x denotes the cluster conditional distribution. Since the OT that transports the multimodal data to the clustering prototypes is actually finding the optimal prototype c j for each data point x i , which is fitting the cluster conditional distribution p c | x , the joint distribution has the same solution as OT.
To fit the joint distribution over the data measure μ and the clustering prototype measure ν , the solution is to find an optimal coupling measure π between μ and ν such that π ( , c ) = μ and π ( x , ) = ν are satisfied for any corresponding marginal measures. The optimal coupling π is obtained by finding the infimum of the coupling cost in the set of coupling measures Π ( μ , ν ) as follows:
π = arg inf π Π ( μ , ν ) X × C d ( x , c ) d π ( x , c ) ,
where d x , c stands for the coupling cost between x and c . The optimal coupling solution π is obtained by reaching the infimum of the total coupling cost over the product space Π ( X × C ) .
Since p x , c = p x | c p c , the fitting of p x , c can be transformed into the learning of the categorical prior distribution p c and the generative conditional distribution p x | c . The former models prior knowledge of the data’s categorical distribution, which can be represented by any distribution C a t . The latter learns the generation mechanism of the data to better reflect the probabilistic properties of the data, which can be achieved through the generative transport plan G from the clustering prototypes to the data. Then, the categorical distribution C a t and the generative transport plan G can be updated by minimizing the transport cost between the generated data and the prototypes:
arg inf G , C a t c ~ C a t d ( G ( c ) , c ) d ν ( c ) ,
where d ( ) is the ground transport cost between x and c . After fitting the categorical prior distribution p c and the generative conditional distribution p x | c , p x , c can be obtained via the Bayesian rule.

4.2. The Variational Generative Solution to OTMC

Based on the variational generative solution to OT, the modality-specific OT and the modality-common OT decomposed from OTMC can easily be solved. Specifically, the solution to modality-specific OT can be achieved by solving the generative transport plan G s m from the categorical prior distribution C a t s m to the data of the m -th modality space X m , minimizing the following ground transport cost:
MsOT = arg min G s m , C a t s m c ~ C a t s m d s m ( G s m ( c ) , c ) d ν s m ( c ) , m = 1 , , M ,
where G s m : C a t s m X m is the generative transport map of the m -th modality, and G s m generates instances from the categorical distribution, which can retain the real underlying structure information under the supervision of multimodal data. The generative transport plan G s m and the categorical distribution C a t s m can jointly learn the private data structure information and clustering information of the m -th modality.
In practice, the integral over the measure ν s m is approximated by Monte Carlo sampling, i.e., the integral of the measure ν s m over the categorical distribution C a t s m is equivalent to the expectation of all clustering prototypes over C a t s m . Thus, Equation (8) can induce the following objective function of modality-specific OT L s m :
argmin G s m , C a t s m E X m E C a t s m [ ν s m ( c ) d s m ( G s m ( c ) , c ) ] ,
where m = 1 , , M indicates the m -th modality, and the objective function L s m fully captures inherent information of the m -th modality by jointly learning structure differences between data and semantic information within data.
Similarly, the optimization of modality-common OT is equivalent to solving the generative transport plan G c from the categorical distribution to the data distribution for all modalities:
McOT = argmin G c , C a t c c ~ C a t c d c ( G c ( c ) , c ) d ν c ( c ) ,
where G c : C a t c X is the generative transport map for all modalities, which can generate data of all modalities from the common categorical distribution C a t c to capture common semantics of data.
The joint optimization of the generative transport plan G c and the categorical distribution C a t c can simultaneously capture the data structure information of all modalities and crossmodal clustering information. The integral over the measure ν c is approximated by the expectation of all prototypes over C a t c based on the Monte Carlo sampling, so the objective function of modality-common OT L c can be written as:
argmin G c , C a t c E X E C a t c [ ν c ( c ) d c ( G c ( c ) , c ) ] .
The objective function L c effectively preserves common information in multimodal data by jointly optimizing the structure differences of all modalities and crossmodal semantic information.
Thus, OTMC can be optimized via the following variational solution:
L = m = 1 M argmin G s m , C s m E X m E C a t s m [ ν s m ( c ) d s m ( G s m ( c ) , c ) ] + argmin G c , C a t c E X E C a t c [ ν c ( c ) d c ( G c ( c ) , c ) ] + m = 1 M D K L ( C a t s m , C a t c ) ,
where D K L denotes the consistent constraint of categorical prior distributions C a t s m and C a t c , which enables MsOT and McOT to have the same transport target, so that the results of the two transport schemes tend to be the same. It is implemented by the KL-divergence penalty between C a t s m and C a t c .

4.3. Multimodal OT Clustering Network

Although they attract much attention from researchers, generative adversarial networks (GANs) involve performance-unstable and computation-expensive adversarial training. Hence, to implement OTMC, a multimodal OT clustering network is designed within a deep neural network architecture which contains a generative network and a clustering network, instead of a GAN. The generative network implements the generative transport plan G together with the categorical distribution C a t ( ϖ ) , and the clustering network produces final assignment results from multimodal data.
The clustering network aims to learn the modality-specific and modality-common soft-clustering assignments that are included in MsOT and McOT. Specifically, in modality-specific OT, the clustering network E s m ( ; θ s m ) encodes modality-specific data into corresponding latent features, and utilizes Student‘s t -distribution in the last layer of E s m ( ; θ s m ) to calculate the similarity between latent features z i m and clustering prototypes c j m to obtain the soft-clustering-assignment matrix P m as follows:
z i m = E s m ( x i m ; θ s m ) p i j m = ( 1 + | | z i m c j m | | 2 / α ) α + 1 2 j ( 1 + | | z i m c j m | | 2 / α ) α + 1 2 ,
where x i m X m is the i -th instance from the m -th modality; c j m C a t m ( ϖ m ) is the j -th clustering prototype of the m -th modality; and p i j m can be interpreted as the possibility that the i -th instance is divided into the j -th cluster. To guide the learning of networks, a target distribution Q m is introduced, and the KL-divergence between Q m and the soft-assignment matrix P m is used as the objective function:
L c l u s t e r m = D K L ( Q m , P m ) ,
where q i j = p i j 2   / i p i j j p i j 2 / i p i j is the target assignment possibility of the i -th instance to the j -th clustering prototype. This objective function guides network learning by aligning the soft assignment to the target probability distribution.
For McOT, the objective of the clustering network E c ( ; θ c ) is D K L ( Q , P ) , where P and Q are the soft-assignment matrix and target probability distribution over all modalities, respectively. According to the consistent constraint L c o n in Equation (5), each D ( P m , P ) is also regarded as an objective component. Hence, the total objective of the clustering networks is
L c l u s t e r = m = 1 M ( D ( Q m , P m ) + D ( P m , P ) ) + D ( Q , P ) ,
where D ( , ) denotes the KL-divergence between two assignment matrices. The first term guides the network to learn the clustering information within each modality, the second term is the consistency constraint of modality-specific and modality-common clustering information, and the last term guides the network to capture the clustering information over all modalities. Finally, the soft-assignment matrix P is used as the cluster assignment matrix.
The generative network is designed to learn the categorical distribution C a t ( ϖ ) and the generative transport plan G included in MsOT and McOT. Specifically, for MsOT, the generative network G s m ( ; φ s m ) decodes latent features z ˜ i m sampled from the categorical distribution C a t ( ) to corresponding data:
z i m = s a m p l e ( C a t ( ϖ m ) ) x ˜ i m = G s m ( z i m ; φ s m ) ,
where C a t ( ϖ m ) is the categorical distribution of the m -th modality with parameter ϖ m . To update the parameter ϖ s m of the generative network and the categorical distribution parameter ϖ m , the topological reconstruction loss between the generative data and the original data is used as a guide, namely,
L r e c m = i = 1 N D w ( x i m , G s m ( s a m p l e ( C a t ( ϖ m ) ) ) ) ,
where D w ( , ) represents the topological reconstruction loss based on optimal transport to measure the differences in manifold structures between original and reconstructed data. The network structure and loss of the generative network G c ( ; φ c ) of McOT are similar to those of MsOT. So, the total objection function of generative networks can be written as follows:
L r e c = m = 1 M i = 1 N D w ( x i m , G s m ( z i m ) ) + D w ( x i , G c ( z i ) ) ,
where the purpose of the first term is to optimize the categorical distribution and generative network within the modality, and the purpose of the second term is to optimize over all modalities, capturing the data structure information within and between modalities, respectively.

4.4. The Overall Loss

The overall loss function L of the multimodal OT clustering network is defined as follows:
L = L r e c + α L c l u s t e r ,
where α denotes the trade-off parameter for balancing the topological reconstruction loss L r e c and clustering pattern mining loss L c l u s t e r , and L is used to optimize the multimodal OT clustering network via the stochastic gradient descent strategy, which effectively boosts the performance of multimodal clustering:
θ t + 1 = θ t η t θ L ( θ t ) ,
where θ t and η t denote the parameter and the learning rate of the multimodal OT clustering network at time step t , respectively, and θ L ( θ t ) denotes the gradient at time step t . The overall optimization algorithm is shown in Algorithm 1.
Algorithm 1. Multimodal OT clustering network.
Input: Multimodal dataset X X and convergence criteria thr.
Output: Soft-clustering-assignment matrix P and the parameters of clustering network E s m ( ; θ s m ) ,   E c ( ; θ c ) , generative network G s m ( ; φ s m ) ,   G c ( ; φ c ) categorical distribution C a t ( ϖ s m ) ,   C a t ( ϖ c ) .
Initialize: modality-specific categorical distributions C a t ( ϖ s m ) modality-common categorical distribution C a t ( ϖ c ) parameters of modality-specific clustering networks E s m ( ; θ s m ) , parameters of modality-common clustering network E c ( ; θ c ) , parameters of modality-specific generative networks G s m ( ; φ s m ) , and parameters of modality-common generative network G c ( ; φ c )
while convergence criteria thr is not reached by Equation (19) do
for each m = 1 , , M do
  Sample data in the m -th modality X m .
  Generate the m -th modality soft-assignment matrix P m according to Equation (13).
  Update E s m ( ; θ s m ) by minimizing Equation (14).
  Sample modality-specific clustering prototypes from C a t ( ϖ s m ) .
  Generate the m -th modality reconstructed data x ˜ i m according to Equation (16).
  Update G s m ( ; φ s m ) and C a t ( ϖ s m ) by minimizing Equation (17).
end
 Sample multimodal data X .
 Generate modality-common soft-assignment matrix P .
 Update modality-common clustering network E c ( ; θ c ) by minimizing Equation (15).
 Sample modality-common clustering prototypes from C a t ( ϖ c ) .
 Generate multimodal reconstructed data.
 Update modality-common generative network G c ( ; φ c ) by minimizing Equation (18).
 Update all network parameters by minimizing Equation (19) according to Equation (20).
end

5. Experiments

5.1. Experimental Setup

Datasets. Four real-world high-dimensional image datasets are utilized in the experiments to evaluate the performance of OTMC. Specifically, Handwritten has 2000 objects of 0–9, where the pixel features and profile correlations are used as two different modalities [31]. ORL collects 400 objects in terms of intensity features and Gabor features, which are collected under 10 different class conditions, such as lighting and facial expressions, of 40 subjects [32]. LandUse consists of 2100 satellite images distributed in 21 scene categories, where each satellite image is described by LBP and PHOG features [33]. Scene contains 4485 objects of 15 natural scene categories, in which PHOG and GIST features are used for each object [34]. The detailed information of these datasets is listed in Table 2.
Comparison methods. The comparison methods are grouped into two categories, deep fusion methods and deep clustering methods. Specifically, the deep fusion methods include DCCA [18], DCCAE [19], AE2-Nets [9], and CMIB [10]. These methods conduct k-means on the fusion features extracted by the corresponding deep architectures. The deep clustering methods include DMF [20], MVaDE [25], SiMVC [26], CoMVC [26], and SDMVC [27]. These methods are directly conducted on multimodal data to obtain the clustering results in an end-to-end manner. Moreover, FeatConcate is also used as a comparison method, which performs k-means on concatenated features.
Evaluation metrics. Three well-known and widely used evaluation metrics, i.e., accuracy (ACC) [35], normalized mutual information (NMI) [36], and adjusted rand index (ARI), are used to measure the performance of all the methods. High values of these metrics indicate ideal clustering performance.
ACC measures the ratio of correct cluster elements to all elements as follows:
ACC = i = 0 N 1 δ ( l i , m ( p i ) ) N ,
where l i and p i are the labels and the predictions, respectively; m ( ) and δ ( ) are the Kuhn–Munkres function and the Dirac delta function, respectively; and N is the size of the multimodal dataset.
NMI measures the dependence between the label distribution and prediction distribution as follows:
NMI = 2 I ( L ; P ) H ( L ) + H ( P ) ,
where L and P denote the label distribution and the prediction distribution, respectively; and I ( ) and H ( ) denote the mutual information function and the entropy function, respectively.
ARI is a measure on the basis of the rand index (RI) as follows:
ARI = RI E ( RI ) max ( RI ) E ( RI ) ,
where RI is the rand index, and E denotes the expectation operator.
Implementation details. OTMC is implemented in PyTorch. In the experiments, vector representations of instances are normalized to   0 , 1 , and then are fed into the network for training. During training, the Adam solver is used with a batch size of 100 to optimize the network, and the learning rate is set to 10 3 for all modules of the network. In addition, all the reported final results are the average results of 5 repeated experiments.

5.2. Clustering Performance Evaluation

Table 3 demonstrates the numerical clustering results in terms of ACC, NMI, and ARI. The results show that OTMC obtains the best results on four benchmark datasets. Moreover, there are also two observations about the ranking of comparison methods. First, the clustering results produced by deep fusion and deep clustering methods are better than those produced by FeatConcate, which verifies the capability of the deep methods to model correlations of heterogeneous modalities when fusing complementary knowledge. Second, the deep clustering methods outperform the deep fusion methods in clustering pattern mining, since the deep clustering methods obtain clustering-specific fusion knowledge on multimodal data, rather than the generalized fusion knowledge extracted by the deep fusion methods. In summary, OTMC produces the optimal results compared with ten baseline methods. These observations validate the effectiveness of optimal transport between data and clusters. That is, the modality-specific OT and the modality-common OT grounded in the Wasserstein distance can effectively measure divergences hidden in heterogeneous distributions of modalities. In particular, in comparison with FeatConcate, which relies on the strong-notion Euclidean distance, OTMC achieves more improvements on Handwritten than the other three datasets. The potential reason may be the differences in heterogeneous manifold overlaps, i.e., the heterogeneous manifolds in Handwritten have fewer overlaps than those in the other three datasets, which experience feature engineering to enhance the intersection.
Figure 2 shows the results of the Nemenyi tests on the four datasets. It can be seen that the average rank of OTMC is higher than that of the other methods on all four datasets, which statistically indicates the superiority of OTMC.

5.3. Further Evaluation

Ablation analysis. To investigate the effectiveness of each component of OTMC, an ablation analysis is conducted on Handwritten. As shown in Table 4, modality-specific OT produces the lowest results, which are similar in two modalities, due to the limited information of each modality. Modality-common OT produces higher results than modality-specific OT in two modalities, because it can correct the inconsistencies based on modality-common knowledge when transporting common information of modalities. OTMC produces the optimal results, since OT can fuse the private and shared information of modalities. The above observations validate the effectiveness of the collaboration between modality-specific OT and modality-common OT in the pattern mining of multimodal data.
Convergence analysis. To verify the convergence of OTMC, Figure 3 shows the losses on four datasets in each epoch of training. In the 0–10 epoch, all of the losses on the four datasets decrease rapidly. Then, in the 11–20 epoch, the speed of decrease becomes slow. And in the 21–30 epoch, the change in loss tends to zero. In general, OTMC can converge after 30-epoch training (black dotted line in Figure 3).

6. Conclusions

A multimodal clustering method based on optimal transport (OTMC) is designed in this paper to mine intrinsic patterns from the fusion information of multimodal data. It utilizes the weak-notion distance to measure the differences among heterogeneous manifolds, which overcomes the fragile assumption on strict overlaps between data manifolds in previous MC methods. In particular, multimodal clustering is defined as the transport mapping between multimodal data and clustering prototypes, which effectively fuses the complementary information in multimodal data. Then, the variational solution to OT is derived based on a generative transport plan, which utilizes fusion information to produce accurate clustering patterns. Afterwards, a multimodal OT clustering network is designed to achieve the above definition. Finally, extensive experiments illustrate the superiority of OTMC.
The possible future directions of this research are three-fold: First, in current OTMC, the transport plan is completely implemented by a deep neural network. Hence, the performance of OTMC heavily depends on the degree of fitting between the deep neural network and the real transport plan, which introduces uncertainty and may be a potential limitation. In the future, a variational solution with more compact margins will be explored to address this dependence and achieve a more precise map for pattern mining. Second, the performance of OTMC on different types of multimodal datasets, especially those with varying degrees of distribution overlap or complementarity across modalities via feature engineering, will be checked to further verify the theoretical basis of OTMC. Third, OTMC will be tried in other multimodal data learning cases such as graph data classification and time series prediction.

Author Contributions

Methodology, Y.G.; writing—original draft preparation, Z.Y. and C.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Liaoning Provincial Department of Education Scientific Research Fund Project (Approval Number: LJ22241163204).

Data Availability Statement

The original data presented in the study are openly available in Handwritten at https://archive.ics.uci.edu/ml/datasets/Multiple+Features (accessed on 28 January 2025), ORL at https://cam-orl.co.uk/facedatabase.html (accessed on 28 January 2025), LandUse at http://weegee.vision.ucmerced.edu/datasets/landuse.html (accessed on 28 January 2025), and Scene at https://figshare.com/articles/dataset/15-Scene_Image_Dataset/7007177 (accessed on 28 January 2025).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Kiaei, I.; Lotfifard, S. A two-stage fault location identification method in multiarea power grids using heterogeneous types of data. IEEE Trans. Ind. Inform. 2019, 15, 4010–4020. [Google Scholar] [CrossRef]
  2. Li, Y.; Yang, M.; Zhang, Z. A survey of multi-view representation learning. IEEE Trans. Knowl. Data Eng. 2019, 31, 1863–1883. [Google Scholar] [CrossRef]
  3. Fu, L.; Lin, P.; Vasilakos, A.V.; Wang, S. An overview of recent multi-view clustering. Neurocomputing 2020, 402, 148–161. [Google Scholar] [CrossRef]
  4. Gao, J.; Liu, M.; Li, P.; Laghari, A.A.; Javed, A.R.; Victor, N.; Gadekallu, T.R. Deep incomplete multi-view clustering via in-formation bottleneck for pattern mining of data in extreme-environment IoT. IEEE Internet Things J. 2023, 11, 26700–26712. [Google Scholar] [CrossRef]
  5. Gao, J.; Liu, M.; Li, P.; Zhang, J.; Chen, Z. Deep multiview adaptive clustering with semantic invariance. IEEE Trans. Neural Netw. Learn. Syst. 2024, 35, 12965–12978. [Google Scholar] [CrossRef]
  6. Wang, H.; Yang, Y.; Liu, B. GMC: Graph-based multi-view clustering. IEEE Trans. Knowl. Data Eng. 2020, 32, 1116–1129. [Google Scholar] [CrossRef]
  7. Caron, M.; Bojanowski, P.; Joulin, A.; Douze, M. Deep clustering for unsupervised learning of visual features. In Proceedings of the European Conference on Computer Vision, Munich, Germany, 8–14 September 2018; pp. 139–156. [Google Scholar]
  8. Gao, J.; Li, P.; Laghari, A.A.; Srivastava, G.; Gadekallu, T.R.; Abbas, S.; Zhang, J. Incomplete multiview clustering via semidiscrete optimal transport for multimedia data mining in IoT. ACM Trans. Multim. Comput. Commun. Appl. 2024, 20, 1–20. [Google Scholar] [CrossRef]
  9. Zhang, C.; Liu, Y.; Fu, H. AE2-Nets: Autoencoder in autoencoder networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 2577–2585. [Google Scholar]
  10. Wan, Z.; Zhang, C.; Zhu, P.; Hu, Q. Multi-view information-bottleneck representation learning. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtual, 2–9 February 2021; pp. 10085–10092. [Google Scholar]
  11. Yang, Y.; Wang, H. Multi-view clustering: A survey. Big Data Min. Anal. 2018, 1, 83–107. [Google Scholar] [CrossRef]
  12. Chao, G.; Sun, S.; Bi, J. A survey on multi-view clustering. arXiv preprint 2017, arXiv:1712.06246. [Google Scholar]
  13. Li, P.; Laghari, A.A.; Rashid, M.; Gao, J.; Gadekallu, T.R.; Javed, A.R.; Yin, S. A deep multimodal adversarial cycle-consistent network for smart enterprise system. IEEE Trans. Ind. Inform. 2023, 19, 693–702. [Google Scholar] [CrossRef]
  14. Xiao, Q.; Dai, J.; Luo, J.; Fujita, H. Multi-view manifold regularized learning-based method for prioritizing candidate disease miRNAs. Knowl. Based Syst. 2019, 175, 118–129. [Google Scholar] [CrossRef]
  15. Neema, I.; Ardecani, F.B.; Shoghli, O. Cluster-based deterioration prediction of composite pavements with incorporation of flooding. In Proceedings of the 39th International Symposium on Automation and Robotics in Construction, Bogota, Colombia, 13–15 July 2022; pp. 99–106. [Google Scholar]
  16. Mirghaderi, H.; Hassanizadeh, B. k-most suitable locations problem: Greedy search approach. Int. J. Ind. Syst. Eng. 2022, 42, 80–95. [Google Scholar] [CrossRef]
  17. Rahiminasab, A.; Tirandazi, P.; Ebadi, M.J.; Ahmadian, A.; Salimi, M. An energy-aware method for selecting cluster heads in wireless sensor networks. Appl. Sci. 2020, 10, 7886. [Google Scholar] [CrossRef]
  18. Andrew, G.; Arora, R.; Bilmes, J.A.; Livescu, K. Deep canonical correlation analysis. In Proceedings of the International Conference on Machine Learning, Atlanta, GA, USA, 16–21 June 2013; pp. 1247–1255. [Google Scholar]
  19. Wang, W.; Arora, R.; Livescu, K.; Bilmes, J.A. On deep multi-view representation learning. In Proceedings of the International Conference on Machine Learning, Lille, France, 6–11 July 2015; pp. 1083–1092. [Google Scholar]
  20. Zhao, H.; Ding, Z.; Fu, Y. Multi-view clustering via deep matrix factorization. In Proceedings of the AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017; pp. 2921–2927. [Google Scholar]
  21. Bodaghi, M.; Hosseini, M.; Gottumukkala, R. A multimodal intermediate fusion network with manifold learning for stress detection. arXiv preprint 2024, arXiv:2403.08077. [Google Scholar]
  22. Ma, Z.; Yu, J.; Wang, L.; Chen, H.; Zhao, Y.; He, X.; Wang, Y.; Song, Y. Multi-view clustering based on view-attention driven. Int. J. Mach. Learn. Cybern. 2023, 14, 2621–2631. [Google Scholar] [CrossRef]
  23. Liu, S.; Zhu, C.; Li, Z.; Yang, Z.; Gu, W. View-driven multi-view clustering via contrastive double-learning. Entropy 2024, 26, 470. [Google Scholar] [CrossRef]
  24. Dornaika, F.; Hajjar, S.E.; Charafeddine, J.; Barrena, N. Unified multi-view data clustering: Simultaneous learning of consensus coefficient matrix and similarity graph. Cogn. Comput. 2025, 17, 38. [Google Scholar] [CrossRef]
  25. Yin, M.; Huang, W.; Gao, J. Shared generative latent representation learning for multi-view clustering. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; pp. 6688–6695. [Google Scholar]
  26. Trosten, D.J.; Lokse, S.; Jenssen, R.; Kampffmeyer, M. Reconsidering representation alignment for multi-view clustering. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Virtual, 19–25 June 2021; pp. 1255–1265. [Google Scholar]
  27. Xu, J.; Ren, Y.; Tang, H.; Yang, Z.; Pan, L.; Yang, Y.; Pu, X.; Yu, P.S.; He, L. Self-supervised discriminative feature learning for multi-view clustering. IEEE Trans. Knowl. Data Eng. 2023, 35, 7470–7482. [Google Scholar] [CrossRef]
  28. Gao, Q.; Lian, H.; Wang, Q.; Sun, G. Cross-modal subspace clustering via deep canonical correlation analysis. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; pp. 3938–3945. [Google Scholar]
  29. Mao, Y.; Yan, X.; Guo, Q.; Ye, Y. Deep mutual information maximin for cross-modal clustering. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtual, 2–9 February 2021; pp. 8893–8901. [Google Scholar]
  30. Yang, H.; Deng, Z.; Zhang, W.; Wu, Q.; Choi, K.; Wang, S. End-to-end multiview fuzzy clustering with double representation learning and visible-hidden view cooperation. IEEE Trans. Fuzzy Syst. 2024, 32, 483–497. [Google Scholar] [CrossRef]
  31. Multiple Features – UCI Machine Learning Repository. Available online: https://archive.ics.uci.edu/ml/datasets/Multiple+Features (accessed on 28 January 2025).
  32. The Database of Faces. Available online: https://cam-orl.co.uk/facedatabase.html (accessed on 28 January 2025).
  33. UC Merced Land Use Dataset. Available online: http://weegee.vision.ucmerced.edu/datasets/landuse.html (accessed on 28 January 2025).
  34. 15-Scene Image Dataset. Available online: https://figshare.com/articles/dataset/15-Scene_Image_Dataset/7007177 (accessed on 28 January 2025).
  35. Fränti, P.; Sieranoja, S. Clustering accuracy. Appl. Comput. Intell. 2024, 4, 24–44. [Google Scholar] [CrossRef]
  36. Kvalseth, T.O. Entropy and correlation: Some comments. IEEE Trans. Syst. Man and Cybern. 1987, 17, 517–519. [Google Scholar] [CrossRef]
Figure 1. A scheme of OTMC. Multimodal clustering is defined as OTMC that contains modality-specific optimal transport (MsOT) and modality-common optimal transport (McOT). Naturally, MsOT and McOT have similar structures because both of them model OT from the data to the prototype space. MsOT adaptively learns a soft-assignment matrix P s m between the data X m and the clustering prototypes C s m , and McOT adaptively learns a soft-assignment matrix P c between the multimodal data X and the common clustering prototypes C c . Furthermore, a consistent constraint is applied to the categorical distributions C a t ( ϖ s m ) , C a t ( ϖ s v ) , and C a t ( ϖ c ) , whose details can be found in Section 3.3 and Section 4.2. Note that the number of MsOTs is equal to that of modalities, and the two MsOTs in the figure are only shown for esthetic reasons.
Figure 1. A scheme of OTMC. Multimodal clustering is defined as OTMC that contains modality-specific optimal transport (MsOT) and modality-common optimal transport (McOT). Naturally, MsOT and McOT have similar structures because both of them model OT from the data to the prototype space. MsOT adaptively learns a soft-assignment matrix P s m between the data X m and the clustering prototypes C s m , and McOT adaptively learns a soft-assignment matrix P c between the multimodal data X and the common clustering prototypes C c . Furthermore, a consistent constraint is applied to the categorical distributions C a t ( ϖ s m ) , C a t ( ϖ s v ) , and C a t ( ϖ c ) , whose details can be found in Section 3.3 and Section 4.2. Note that the number of MsOTs is equal to that of modalities, and the two MsOTs in the figure are only shown for esthetic reasons.
Electronics 14 00666 g001
Figure 2. Results of Nemenyi test on four datasets. (a) Handwritten; (b) ORL; (c) LandUse; (d) Scene.
Figure 2. Results of Nemenyi test on four datasets. (a) Handwritten; (b) ORL; (c) LandUse; (d) Scene.
Electronics 14 00666 g002
Figure 3. Convergence analysis on four datasets.
Figure 3. Convergence analysis on four datasets.
Electronics 14 00666 g003
Table 1. Frequently used notations.
Table 1. Frequently used notations.
NotationsDescription
X , X ˜ Multimodal data/reconstructed data space.
X m , X ˜ m m -th modality data/reconstructed data space.
M Number of modalities.
k Number of clusters.
μ Multimodal data measure.
μ s m m -th modality-specific data measure.
C Clustering prototype set.
c j j -th clustering prototype.
C s m m -th modality-specific clustering prototype set.
c j m m -th modality, j -th clustering prototype.
ν Clustering prototype measure over all modalities.
ν s m m -th modality-specific clustering prototype measure.
T Clustering transport map from data to prototypes.
T s m m -th modality-specific transport map.
T c Common transport map over all modalities.
W Multimodal data partition scheme.
W j Set of data split into j -th cluster.
π Coupled measure of measures μ and ν .
G Generative transport plan from prototypes to data.
G s m m -th modality-specific generative transport plan.
C a t Categorical prior distribution.
C a t s m m -th modality-specific categorical prior distribution.
C a t ( ϖ ) Categorical distribution with parameter ϖ .
E s m ( ; θ s m ) Clustering network with parameter θ s m .
P m m -th modality soft-clustering-assignment matrix.
G s m ( ; φ s m ) Generative network with parameter φ s m .
D w ( , ) OT-based topological reconstruction loss.
Table 2. The information of datasets.
Table 2. The information of datasets.
DatasetNumberModalityClass
Handwritten20002 (pixel/profile correlation)10 (handwritten numbers 0–9)
ORL4002 (intensity/Gabor)10 (face image shooting conditions)
LandUse21002 (LBP/PHOG)21 (satellite image scene categories)
Scene44852 (PHOG/GIST)15 (natural scene categories)
Table 3. Average clustering results on Handwritten, ORL, LandUse, and Scene datasets. Best results are marked in bold.
Table 3. Average clustering results on Handwritten, ORL, LandUse, and Scene datasets. Best results are marked in bold.
DatasetHandwrittenORLLandUseScene
MetricACCNMIARIACCNMIARIACCNMIARIACCNMIARI
FeatConcate0.61040.60700.55320.57100.75280.47700.12320.16080.03650.30760.35400.1862
DCCA [18]0.66260.66010.61360.59680.77840.50200.15510.23150.04430.36180.38920.2087
DCCAE [19]0.69170.66960.63270.59400.77520.49930.15620.24410.04420.36440.39780.2147
DMF [20]0.69620.71600.59930.67330.81640.54070.14500.15430.03600.33930.35760.1862
AE2-Nets [9]0.81520.71390.66670.68850.85730.56370.24790.30360.10350.36100.40390.2208
MVaDE [25]0.88750.80760.77650.69500.83560.56430.22480.28480.09360.37820.39920.2178
CoMVC [26]0.82050.81420.75590.70630.86520.53530.26240.30830.10850.38590.41170.2231
SiMVC [26]0.82950.76080.69850.69210.85600.52560.24480.25810.09570.37750.39350.2259
SDMVC [27]0.89900.82140.80070.71040.85570.59420.26810.29860.12010.38570.41320.2259
CMIB [10]0.89720.81780.79880.72070.88230.60040.27160.30160.12420.39540.41770.2383
OTMC0.92150.84960.83350.76500.88370.64960.28140.32540.13120.41810.43730.2621
Table 4. Ablation analysis on Handwritten in terms of ACC, NMI, and ARI. Best results are marked in bold.
Table 4. Ablation analysis on Handwritten in terms of ACC, NMI, and ARI. Best results are marked in bold.
MethodsACCNMIARI
modality-specific OT 10.72150.72760.6212
modality-specific OT 20.72550.70920.6141
modality-common0.90900.83060.8098
OTMC0.92150.84960.8335
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yang, Z.; Shi, C.; Guan, Y. An Optimal-Transport-Based Multimodal Big Data Clustering. Electronics 2025, 14, 666. https://doi.org/10.3390/electronics14040666

AMA Style

Yang Z, Shi C, Guan Y. An Optimal-Transport-Based Multimodal Big Data Clustering. Electronics. 2025; 14(4):666. https://doi.org/10.3390/electronics14040666

Chicago/Turabian Style

Yang, Zheng, Chongyang Shi, and Ying Guan. 2025. "An Optimal-Transport-Based Multimodal Big Data Clustering" Electronics 14, no. 4: 666. https://doi.org/10.3390/electronics14040666

APA Style

Yang, Z., Shi, C., & Guan, Y. (2025). An Optimal-Transport-Based Multimodal Big Data Clustering. Electronics, 14(4), 666. https://doi.org/10.3390/electronics14040666

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop