Next Article in Journal
Feature Residual Analysis Network for Building Extraction from Remote Sensing Images
Next Article in Special Issue
Implementation Aspects of Smart Grids Cyber-Security Cross-Layered Framework for Critical Infrastructure Operation
Previous Article in Journal
Performance of Steel-Bolt-Connected Industrialized Building System Frame Subjected to Hydrodynamic Force
Previous Article in Special Issue
A Systematic Review on Blockchain Adoption
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Fusing Local and Global Information for One-Step Multi-View Subspace Clustering

1
Department of Electrical Engineering, School of Automation, Guangdong University of Technology, Guangzhou 510006, China
2
Brunel Interdisciplinary Power Systems Research Centre, Department of Electronic and Electrical Engineering, Brunel University London, London UB8 3PH, UK
*
Authors to whom correspondence should be addressed.
Submission received: 21 March 2022 / Revised: 4 May 2022 / Accepted: 17 May 2022 / Published: 18 May 2022
(This article belongs to the Special Issue Electrification of Smart Cities)

Abstract

:
Multi-view subspace clustering has drawn significant attention in the pattern recognition and machine learning research community. However, most of the existing multi-view subspace clustering methods are still limited in two aspects. (1) The subspace representation yielded by the self-expression reconstruction model ignores the local structure information of the data. (2) The construction of subspace representation and clustering are used as two individual procedures, which ignores their interactions. To address these problems, we propose a novel multi-view subspace clustering method fusing local and global information for one-step multi-view clustering. Our contribution lies in three aspects. First, we merge the graph learning into the self-expression model to explore the local structure information for constructing the specific subspace representations of different views. Second, we consider the multi-view information fusion by integrating these specific subspace representations into one common subspace representation. Third, we combine the subspace representation learning, multi-view information fusion, and clustering into a joint optimization model to realize the one-step clustering. We also develop an effective optimization algorithm to solve the proposed method. Comprehensive experimental results on nine popular multi-view data sets confirm the effectiveness and superiority of the proposed method by comparing it with many state-of-the-art multi-view clustering methods.

1. Introduction

Clustering is a fundamental unsupervised learning problem that is widely used in the tasks of machine learning [1], computer vision [2], and data mining [3]. It attempts to help to understand the structure of unlabeled data by dividing the entire unlabeled samples into clusters, where the samples in the same cluster are not similar to samples in the other clusters [4,5,6].
With the continuous development of information technology, different features of the object can be easily acquired by different feature extractors, data sources or sensors. For example, an image can be depicted by the color, texture, and edge features. A news report is usually composed of text descriptions and pictures. In the field of autonomous driving, an obstacle can be captured by different types of sensors. These different features can be viewed as multi-view data. Since each view commonly contains view-specific information about the object, using only one view for clustering may yield poor results [7]. Therefore, it is reasonable and appropriate to fuse different views for clustering. It is known that multiple views come from the same object. Hence, multi-view data contain not only the consistency but also the diversity across views. How to reasonably utilize the consistency and diversity to find the underlying clustering structure of multi-view data has become an important research topic .
To deal with the multi-view data, a natural idea is to concatenate these different feature vectors into a new vector and then adopt some existing single-view clustering methods to group the multi-view data. Although this idea is intuitive and simple to deal with multi-view data, it ignores the consistency and complementary information across these views. To address this problem, lots of multi-view clustering methods have been developed to obtain the good clustering performance. For background reading, the reader can refer to the surveys on multi-view clustering [8,9,10]. In this paper, we mainly focus on the multi-view subspace clustering, which has received extensive attention due to its advanced clustering performance and good mathematical interpretability.
Multi-view subspace clustering attempts to construct an ideal subspace representation to describe the multiple linear subspace structure, and the clustering results are then obtained by utilizing the spectral clustering for this obtained subspace representation. The mechanism for computing the subspace representation is based on the self-expressive reconstruction model, where each sample is reconstructed by entire samples. Hence, subspace representation yielded by the self-expressive reconstruction model can exploit the global information but may ignore the local information of multi-view data. Nevertheless, exploring the local structure has been confirmed to improve the learning performance [11]. Moreover, most multi-view subspace clustering methods divide the learning subspace representation and clustering into two individual procedures, which ignores their communications.
To address the above-mentioned issues, in this paper, we propose a novel subspace clustering method fusing local and global information for one-step multi-view subspace clustering (LGOMSC). The proposed method combines the procedures of constructing subspace representation, multi-view information fusion, and clustering into a unified optimization framework. In this framework, as shown in Figure 1, to exploit the local and global information of multi-view data, we integrate graph learning into the self-expressive reconstruction model by adaptively exploring the local structure information for the construction of subspace representation. To capture latent consistency information across views, the proposed method adopts a multi-view information fusion to learn the common subspace representation from these specific subspace representations of different views. Meanwhile, in graph learning, a rank constraint is applied to the Laplacian matrix yielded by the common subspace representation to directly produce the clustering result. Therefore, the proposed method is a one-step multi-view subspace clustering method. The main contributions of the work are summarized as follows:
  • A novel one-step multi-view subspace clustering method is proposed, which fuses the subspace representation (exploring local and global information), multi-view information fusion (constructing a common subspace representation by fusing different view-specific subspace representations), and clustering (imposing rank constraint on the Laplacian matrix from the common subspace representation) as a unified optimization framework to realize the end-to-end clustering.
  • We develop an effective optimization algorithm to solve the proposed method. Comprehensive experiments on nine popular multi-view data sets confirm the effectiveness and superiority of the proposed method by comparing it with some state-of-the-art multi-view clustering methods.
Figure 1. Framework of proposed LGOMSC.
Figure 1. Framework of proposed LGOMSC.
Applsci 12 05094 g001
The rest of this paper is organized as follows. In Section 2, we review the related works. In Section 3, we introduce the formulation of the proposed LGOMSC method. In Section 4, we provide the optimization algorithm to solve the proposed LGOMSC method, including the analysis of the convergence and computation complexity. In Section 5, we conduct the experiments on nine popular multi-view data sets and analyze the experimental results. Finally, we provide the conclusion in Section 6.

2. Related Work

Multi-view clustering is a very powerful data analysis tools for unsupervised learning of data with heterogeneous features. In the past two decades, many multi-view clustering methods have been proposed to achieve robust clustering performance. In the following, we will briefly introduce several multi-view clustering methods from different perspectives.
Subspace-based methods have recently become the mainstay of multi-view clustering research, aiming to discover potential subspace structures across different views. For example, Gao et al. [12] propose a multi-view subspace clustering method that utilizes a common cluster structure to exploit the consistency information across multiple views. Cao et al. [13] propose a diversity-induced multi-view subspace clustering method that adopts the Hilbert–Schmidt independence criterion as the diversity term to explore the complementary information of multi-view data. Luo et al. [14] propose a multi-view subspace clustering method that simultaneously considers the consistency and specificity for learning the subspace representation. Wang et al. [15] propose a multi-view subspace clustering method that considers the complementarity of multi-view data by adopting a position-aware exclusivity term. Guo et al. [16] propose a rank consistency induced multi-view subspace clustering model that learns a consistent subspace structure. Brbić and Kopriva [17] propose a multi-view subspace clustering method that adopts an agreement term to ensure the consistency among these subspace representations. To capture the high-order correlations underlying multi-view data, the tensor technique is adopted to exploit the complementary information among different views. For example, Zhang et al. [18] propose a low-rank tensor constrained multi-view subspace clustering model that adopts a low-rank tensor constraint for the obtained subspace representations. Xie et al. [19] utilize the subspace representations of multiple views as a tensor data and then utilize the tensor-singular value decomposition on the rotated tensor to guarantee the consensus among different views. Zhang et al. [20] propose a tensorized multi-view subspace representation learning that adopts a low-rank constraint model for the subspace representation tensor. Yin et al. [21] propose a multi-view subspace clustering model by organizing the multi-view data as tensorial data, and the tensorial data can be represented by a t-linear combination with sparse and low-rank penalty. Recently, researchers considered partition-level multi-view information fusion and proposed a partition-based clustering model to construct joint optimization of multi-view subspace clustering. For example, Kang et al. [22] propose a unified multi-view subspace clustering model that implements the graph construction, the generation of basic partitions, and the fusion of consensus clustering in an interactive way. Lv et al. [23] propose a partition fusion-based multi-view subspace clustering method that utilizes the different partitions to find a shared partition. Zhang et al. [24] develop a consensus one-step multi-view subspace clustering method that fuses the subspace representation learning, partition learning, and clustering into a whole to iteratively optimize. Kang et al. [25] propose to integrate multi-view information in the partition space and obtain clustering results by assigning each partition with a respective rotation matrix. Furthermore, each view is assigned a weight to consider the differences in the clustering capacity of the views. The anchor-based model is proposed to fit for the large-scale multi-view data. Kang et al. [26] propose a large-scale multi-view subspace clustering method by integrating the anchor graphs from different views for spectral clustering. Wang et al. [27] propose a fast parameter-free multi-view subspace clustering by adaptively learning the anchors and graph structure. Sun et al. [28] propose to combine anchor learning and graph construction into a unified optimization framework, allowing the learned anchors to represent the actual latent data distribution more accurately, leading to a more discriminative clustering structure.
Matrix factorization-based methods refer to obtaining consistent latent representations through matrix factorization. Specifically, a given data matrix can be represented by the product of two or more low-dimensional matrices. Liu et al. [29] extended the traditional single-view non-negative matrix factorization algorithm to multi-view application scenarios and proposed a multi-view clustering algorithm based on non-negative matrix factorization. Guo et al. [30] propose to exploit group sparsity inducing norm in a matrix factorization framework to learn shared sparse subspace representations. Recently, Wang et al. [31] proposed a diversity non-negative matrix factorization multi-view clustering method by introducing a new diversity term to increase the diversity among multi-view representations and linearize the running time. Nie et al. [32] propose a new joint clustering method named Fast Multi-view Matrix Tri-Factorization to reduce the information loss in the matrix factorization process, while reducing the computational complexity and improving the operational efficiency. Liu et al. [33] propose a novel multi-view matrix factorization-based clustering method, which proposes to consider the higher-order relationships among features using an optimal graph regularization strategy and introduces the Hilbert–Schmidt independence criterion (HSIC) to fully explore the complementary information in different views. In addition, researchers have extended matrix factorization from the perspective of intact space learning [34]. For example, Zhang et al. [35] propose a latent multi-view subspace clustering that utilizes the latent representation for subspace clustering. Li et al. [36] propose a flexible multi-view representation learning that utilizes the kernel dependence measure to obtain a latent representation from different views for subspace clustering. Xie et al. [37] propose a multi-view subspace clustering method that fuses graph learning, latent representation, and clustering into a unified optimization framework.
Graph-based methods provides an effective way to solve the nonlinearly separated problems. For example, Tang et al. [38] propose a fusion process using linked matrix factorization to fuse the graph matrices corresponding to all views with multiple sources of information. Nie et al. [39] propose a multi-view graph clustering method based on the idea of manifold learning that can perform local structure learning and multi-view clustering at the same time and can also adaptively learn the weights corresponding to each view. Meanwhile, Nie et al. [40] propose an automatic weighting method to fuse a series of view-specific low-quality graphs into a high-quality unified graph, while extending the Laplacian rank approach to multi-view learning. Similarly, Zhan et al. [41] further design a notable clustering method based on twostep multiple graph fusion strategy. Recently, Zhan et al. [42] proposed a method to learn a consensus graph matrix by all views by minimizing disagreement between different views and constraining the rank of the Laplacian matrix. Wang et al. [43] propose another graph-based multi-view clustering method that automatically fuses multiple graph matrices to generate a unified graph matrix. The learned unified graph matrix can help the graph matrices of all views and gives the clustering indicator matrix. Recently, Zhao et al. [44] proposed to minimize the divergence between graphs using tensor Schatten p-norm regularization and integrate the tensor Schatten p-norm regularization and the manifold learning regularization into a unified framework to learn a shared common graph.
Although most of existing multi-view subspace clustering methods have achieved good clustering performance, they still have some limitations. First, the subspace representation generated by the self-expression reconstruction model usually ignores the local structure of the data set. Second, most multi-view subspace clustering methods usually divide the subspace representation learning process and the subsequent clustering task into two separate processes, ignoring the interactions between them. To address these issues, in this paper we propose an LGOMSC method that considers adding graph learning to explore local information adaptively for obtaining subspace representations. Moreover, LGOMSC performs multi-view information fusion directly on the subspace representation and introduces rank constraints on the Laplacian matrix of the common subspace representation matrix, which helps to naturally partition the data points into the desired number of clusters. Our approach integrates similarity learning, multi-view information fusion and clustering as a unified framework to achieve multi-view clustering in an end-to-end manner.
Duan et al. [45] propose a multi-view subspace clustering (MVSCLG) that also utilizes the local and global information to achieve the end-to-end clustering. The main differences between MVSCLG and LGOMSC include: (1) to explore the consistency between different views, MVSCLG adopts the spectral matrix fusion, but LGOMSC adopts the graph matrix fusion; (2) to achieve end-to-end clustering, MVSCLG adopts a rotation matrix to map the common spectral matrix to the final cluster label matrix, but LGOMSC adopts a rank constraint on the common Laplacian matrix to directly achieve clustering. Moreover, compared with MVSCLG, LGOMSC has some advantages. Firstly, MVSCLG involves many singular value decomposition procedures and contains many variables, which leads to longer running times and more memory usage than LGOMSC. Secondly, LGOMSC contains fewer hyperparameters than MVSCLG, which is more suitable for practical applications.

3. Proposed Method

3.1. Notations

For convenience, we list important mathematic notations that are used throughout the paper in Table 1. Matrices are represented in bold uppercase, while vectors are represented in bold lowercase.

3.2. Formulation

In this section, we provide the detailed modeling process of the proposed LGOMSC method. For a multi-view data set with V views, let X 1 , , X V be the data matrices of the V views and X v = { x 1 v , , x n v } d v × n be the v-th view data, where d v is the dimensionality of the v-th view, and n is the number of data points. Since each view contains view-specific information about the object, we respectively compute the view-specific subspace representation of each view to capture the diversity across views. The objective function of the self-expression model for the multi-view data can be formulated as:
min S v v = 1 V X v X v S v F 2 + λ 1 v = 1 V S v F 2 s . t . 0 s i j v 1 , ( S v ) T 1 = 1 , d i a g ( S v ) = 0
where S v = { s 1 v , s 2 v , , s n v } n × n is the subspace representation matrix of the v-th view, s i j v is the j-th element of s i v , 1 denotes a column vector with all entries of one, d i a g ( ) denotes a vector of the diagonal elements of a matrix.
Since Model (1) adopts the entire data set to linearly reconstruct each data sample, the subspace representation matrix S v captures the global information of the v-th view data. However, this subspace representation obtained by Model (1) ignores the local structure to construct the subspace representations. In other words, two closed data samples should have similar subspace representations. Hence, to exploit the local information of multi-view data, we integrate the graph learning into Model (1) to compute the subspace representation. Hence, the objection function can be formulated as:
min S v v = 1 V X v X v S v F 2 + λ 1 v = 1 V S v F 2 + v = 1 V i = 1 n j = 1 n x i v x j v 2 2 s i j v s . t . 0 s i j v 1 , ( S v ) T 1 = 1 , d i a g ( S v ) = 0
Since multi-view data come from the same object, they should have latent consistency. To characterize this consistency, we adopt a multi-view information fusion term to obtain a common subspace representation matrix U n × n from the subspace representation matrices { S 1 , S 2 S V } . This term can be represented as:
min U v = 1 V S v U F 2 s . t . U 0 , U T 1 = 1
Through minimizing Model (3), this common subspace representation matrix U can make these the subspace representation matrices { S 1 , S 2 S V } to have latent consistency. Hence, we add this multi-view information fusion term into Model (2) as:
min S v , U , F v = 1 V X v X v S v F 2 + λ 1 v = 1 V S v F 2 + v = 1 V i = 1 n j = 1 n x i v x j v 2 2 s i j v + v = 1 V S v U F 2 s . t . 0 s i j v 1 , 1 T s i v = 1 , d i a g ( S v ) = 0 , U 0 , U T 1 = 1
After obtaining the common subspace structure, we can get the affinity matrix W = 1 / 2 ( U + U T ) and perform spectral clustering on such a subspace affinity matrix. However, the constructions of subspace representation and clustering are divided into two individual procedures, which ignore their interactions. To address this problem, we consider introducing a rank constraint [46] on the Laplacian matrix L U = D W , where the degree matrix D is defined as a diagonal matrix whose i-th diagonal element is d i i = j = 1 n w i j . If the common subspace representation matrix U is non-negative, then the Laplacian matrix has the following theorem.
Theorem 1.
The number of connected components in the graph with U is equal to the multiplicity of zero eigenvalue of the Laplacian matrix L U [47].
According to Theorem 1, we consider making the number of zero eigenvalues of the Laplacian matrix L U to be equal to the number of clustering clusters, i.e., r a n k ( L U ) = n c . By adding the rank constraint r a n k ( L U ) = n c into Model (4), the common subspace representation matrix U will have the ideal property. Therefore, we can directly obtain the cluster result from U without discretization.
However, it is difficult to directly solve the rank constraint r a n k ( L U ) = n c . It is known that r a n k ( L U ) = n c is equivalent to i = 1 c σ i ( L U ) = 0 , where σ i ( L U ) denotes the i-th smallest eigenvalues of L U . Since L U is positive semi-definite, σ i ( L U ) 0 . According to Ky Fan’s Theorem [48], i = 1 c σ i ( L U ) = min F T F = I T r ( F T L U F ) . Therefore, to hold i = 1 c σ i ( L U ) = 0 , the objective function of the proposed LGOMSC is formulated as:
min S v , U , F v = 1 V X v X v S v F 2 + λ 1 v = 1 V S v F 2 + v = 1 V i = 1 n j = 1 n x i v x j v 2 2 s i j v + v = 1 V S v U F 2 + 2 λ T r ( F T L U F ) s . t . 0 s i j v 1 , 1 T s i v = 1 , d i a g ( S v ) = 0 , U 0 , U T 1 = 1 , F T F = I c
where λ > 0 is a parameter, F = { f 1 , , f c } n × c is the embedding matrix, and I c c × c denotes the identity matrix.
In Model (5), when λ is large enough, the obtained common subspace representation U makes i = 1 c σ i ( L U ) zero. Hence, r a n k ( L U ) = n c is satisfied. To effectively accelerate the optimization procedure, we determine the value of λ in a heuristic way. Moreover, in Model (5), we integrate subspace representation learning, multi-view information fusion, and clustering into a unified framework. The aim is to exploit the internal relationships of the three procedures to obtain a good clustering performance.

4. Optimization

There are several variables and constraints in the proposed method. To effectively solve these variables from LGOMSC, we developed an alternate optimization algorithm.

4.1. Update S v

When U and F are fixed, the objective function about S v becomes:
min S v X v X v S v F 2 + λ 1 S v F 2 + i = 1 n j = 1 n x i v x j v 2 2 s i j v + S v U F 2 s . t . 0 s i j v 1 , 1 T s i v = 1 , d i a g ( S v ) = 0
In this paper, we adopt a two-step approximation strategy [25] to optimize S v .
Firstly, we ignore the constraints in Model (6) to solve S v as:
min S v X v X v S v F 2 + λ 1 S v F 2 + i = 1 n j = 1 n x i v x j v 2 2 s i j v + S v U F 2
Through making the derivative of Model (7) of S v as zero, we have:
S ^ v = ( ( X v ) T X v + I n + λ 1 I n ) 1 ( ( X v ) T X v + U 1 2 B v )
where b i j v = x i v x j v 2 2 is the ij-th element of B v n × n and I n n × n denotes the identity matrix.
Secondly, through adding the constraints of S v , the solution of S v can be obtained by:
min S i v s i v s ^ i v 2 2 s . t . s i i v = 0 , S i v 0 , 1 T S i v = 1
Model (9) is a constrained quadratic optimization problem, which can be effectively solved by the iterative algorithm in the work [49].

4.2. Update U

When S v and F are fixed, the objective function about U is represented as:
min U v = 1 V i = 1 n j = 1 n ( s i j v u i j ) 2 + 2 λ T r ( F T L U F ) s . t . u i j 0 , 1 T u i = 1
where u i n × 1 is a column vector, u i j is the j-th element of u i .
Noting that T r ( F T L U F ) = 1 / 2 i = 1 n j = 1 n f i , : f j , : 2 2 u i j , we denote h i be a vector with the j-th element h i j = f i , : f j , : 2 2 . Through simple mathematical derivation, problem (10) can be rewritten as follows:
min u i v = 1 V u i s i v + λ 2 V h i 2 2 s . t . u i j 0 , 1 T u i = 1
We define q v = s i v λ 2 V h i , we can obtain:
min u i v = 1 V u i q v 2 2 s . t . u i j 0 , 1 T u i = 1
Model (12) is effectively optimized by an iterative algorithm referring to the work [50].

4.3. Update F

When U and S v are fixed, the objective function about F is represented as:
min F T r ( F T L U F ) s . t . F T F = I c
The optimal solution F yielded by Model (13) is formed by the c eigenvectors of L U corresponding to the c smallest eigenvalues. Finally, the procedure for optimizing Model (5) is described in Algorithm 1.

4.4. Convergence Analysis

In this paper, we adopt an alternate updating algorithm (Algorithm 1) to solve the objective function in Model (5). Since λ is changed during the iteration to accelerate the procedure in the experiment, the objective function of Model (5) is varied during each iteration. Hence, it is difficult to guarantee convergence theoretically. However, in the experiments, the results show that Algorithm 1 for optimizing Model (5) has good convergence.

4.5. Computational Complexity Analysis

According to the optimization process described in Algorithm 1, the computational complexity of LGOMSC consists of updating S v , U , and F . First, the update of S v takes O(n3 + Vn2). Second, the update of U needs O(n2). Third, the update of F costs O(n3) for compute eigenvectors of the Laplacian matrix. Overall, the complexity of Algorithm 1 is O ( ( 2 n 3 + ( V + 1 ) n 2 ) ) T ) , where T is the total number of iterations.
Algorithm 1: Optimization Algorithm for Model (5)
Input: given V view data X 1 , , X V with X v d v × n , the number of clusters c , parameters λ 1 and λ .
Output: U with exact c connected components.
Initialize S v by the optimization problem [51]:
min S v i = 1 n j = 1 n x i v x j v 2 2 s i j v + λ 1 S v F 2 s . t . s i i v = 0 , 0 s i j v 1 , 1 T s i v = 1
Initialize U and F based on s 1 , s 2 s V .
Repeat
Update S v by model (9).
Update U by model (12).
Update F by model (13).
Until  i c σ i ( L U ) < 1.0 e 13 and i c + 1 σ i ( L U ) > 1.0 e 13

5. Experiments

In this section, we use nine popular multi-view data sets to assess the clustering performance of LGOMSC.

5.1. Data Set Descriptions

In the experiments, the nine public multi-view benchmark data sets were 3Sources, 100leaves, BBC, Caltech101, COIL-20, NottingHill, Webkb, Cornell, and Wikipedia Articles. All the data sets are summarized in Table 2.

5.2. Experimental Setting

In this paper, LGOMSC is compared with twelve relevant methods including
  • FeatConcate: Concatenate the features of different views into a vector and utilize k-means to acquire the clustering result. It is regarded as the baseline method.
  • Co-reg_c and Co-reg_p: Centroid-based co-regularization [52] and pairwise co-regularization [52].
  • LMSC: Latent multi-view subspace clustering [35].
  • FMR: Flexible multi-view representation learning for subspace clustering [36].
  • MLRSSC: Multi-view low-rank sparse subspace clustering [17].
  • RMKMC: Robust multi-view k-means clustering [53].
  • mPAC: Multiple partitions aligned clustering [25].
  • LMVSC: Large-scale multi-view subspace clustering in linear time [26].
  • PMSC: Partition level multi-view subspace clustering [22].
  • COMVSC: Consensus one-step multi-view subspace clustering [24].
  • GMC: Graph-based multi-view clustering [43].
  • MVSCLG: Multi-view subspace clustering with local and global information [45].
We conduct these comparison methods from corresponding open-source codes and follow their papers to set the optimal parameters. LGOMSC contains two parameters λ and λ 1 . For λ , it is first set with a proper value in a heuristic way, then in each iteration, λ is divided by two if the number of zero eigenvalues of L U is greater than c and multiplied by two if it is smaller than c. For λ 1 , we adopt the grid search method to empirically choose it in the range of { 1 , 10 , 20 , , 100 } . In this paper, we use L2 norm for data normalization, i.e., ( x i v ) = x i v x i v 2 .

5.3. Experiment Results and Analysis

In the experiments, four popular metrics including accuracy (ACC), normalized mutual information (NMI), F-score and adjusted Rand index (ARI) are utilized to assess the clustering result. These evaluation metrics reflect different natures of the clustering results, thus providing a comprehensive analysis from multiple perspectives. For all four of the evaluation metrics, higher values indicate better results. The comparison results are shown in Table 3,Table 4,Table 5 and Table 6. The best result is highlighted in red font, and the second best result is reported in blue font.
Experiment Analysis. Through observing the clustering results from Table 3,Table 4,Table 5 and Table 6, one can see that our proposed method can obtain the best results on all multi-view data sets except Caltech101. For the Caltech101 data set, LGOMSC is lower than FeatConcate and RMKMC in terms of NMI. However, in terms of ACC, our proposed method exceeds the second best results on the data sets including 3sources, 100leaves, BBC, Caltech101, COIL-20, NottingHill, Webkb, Cornell, and Wikipedia by 6.51%, 7.93%, 2.19%, 2.31%, 4.51%, 8.07%, 2.76%, 8.72% and 1.59%, respectively. For NMI, our proposed method is 8.93% lower than the best method, RMKMC, on the Caltech101 data set. For F-score, our proposed method exceeds the second best method by 1.41%, 14.45%, 4.39%, 3.89%, 6.55%, 11.23%, 3.75%, 7.75% and 0.63% for the corresponding data sets, respectively. In terms of ARI, our proposed method exceeds the second best method by 0.84%, 14.59%, 5.31%, 7.37%, 6.97%, 14.33%, 12.11%, 2.54% and 0.16% for the corresponding data set, respectively. The above results demonstrate the effectiveness and superiority of the proposed method. Hence, our LGOMSC method is a valuable multi-view subspace clustering method.
From the results in these tables, one can see that the baseline method (i.e., FeatConcate) sometimes exhibits comparable performance to the multi-view subspace method and even exceeds some multi-view clustering methods. However, in most cases, this baseline method still has a big gap in comparison with multi-view clustering methods. It confirms that multi-view clustering methods that consider the consistency or complementary information can obtain a good multi-view clustering performance. However, in some data sets, e.g., 3sources and 100leaves, the multi-view k-means method RMKMC, produces even worse results than FeatConcate. This phenomenon has been observed by some previous researchers [18,54].
Multi-view subspace clustering based on intact space learning methods like LMSC and FMR performs clustering on the latent representation space. However, there is a large gap between these methods and our approach, probably because they separate the representation learning and clustering processes, leading to suboptimal clustering results.
Compared with similar multi-view information fusion methods like GMC and LMVSC, our approach achieves a more impressive performance. This is mainly because we fuse graph learning into the self-expression model to jointly explore local and global structural information in the data. More information is used to serve the clustering task and therefore better performance is obtained.
Compared with the partition-based models to construct multi-view subspace clustering for joint optimization methods like COMVSC, mPAC, and PMSC, our approach achieves more impressive performance. This is mainly because we make the clustering structure of the multi-view data revealed while generating the common subspace representation under the rank constraint of Laplacian matrix. Thus, the end-to-end clustering approach facilitates a better clustering performance.
Compared with the MVSCLG method, the overall results from the nine data sets show that our method achieves the best results on all evaluation criteria, except for the Caltech101, Cornell and Wikipedia data sets. It states that using rank constraint can obtain an ideal graph matrix fitting for direct clustering. Our method also has superiorities in terms of running time and memory usage, which will be discussed in the next subsection. Thus, our method is more effective than MVSCLG.
Compared with these state-of-the-art clustering methods, the proposed method can achieve a more impressive clustering performance. The main reason is that it considers the local and global information from the original multi-view data to learn the subspace representation. Moreover, the proposed LGOMSC is an end-to-end model, which fuses the construction of subspace representation, multi-view information fusion, and clustering into a seamless whole. The purpose for this is to dig into their potential correlations.
Statistical Analysis. To demonstrate the statistical properties of our proposed method, we conducted the Friedman test and Nemenyi post-hoc test.
The Friedman test assumes that all the k compared methods hold the same performance on H data sets. Specifically, this model performance evaluation consists of the following two main steps. In the first step, first, sort all methods on each data set from high to low according to the clustering performance index and assign corresponding ordinal values (e.g., 1, 2, …), and then calculate each method on all data sets average rank. In particular, the ordinal values are averaged if the performance of the two methods is the same. Finally, Γ χ 2 and Γ F are calculated, and their mathematical expressions are as follows:
Γ F = ( H 1 ) Γ χ 2 H ( k 1 ) Γ χ 2
where Γ χ 2 = 12 H k ( k + 1 ) ( i = 1 k r i 2 k ( k + 1 ) 2 4 ) , r i represents the average rank of the i-th method over all data sets. Besides, Γ F obeys the F-distribution with the degree of freedom k−1 and (k−1)(H−1). The correctness of the hypothesis is eliminated by comparing the Γ F with its corresponding threshold (the thresholds of the Friedman test can be calculated by q f ( 1 α , k 1 , ( k 1 ) ( H 1 ) ) in R programming language). If the hypothesis is rejected, this indicates a significant difference in the performance of the compared methods. A further Nemenyi post-hoc test is then required to further distinguish between the methods.
In the second step, the Nemenyi post-hoc test calculates the critical distance by Equation (15) to reflect the difference between the average ordinal results of various methods.
C D = q α k ( k + 1 ) 6 H
where q α can be calculate by q t u k e y ( 1 α , k , I n f ) / S q r t ( 2 ) in R programming language.
In our case, the number of compared methods, k, equals 13 and H equals 9. We sort the ACC, NMI, F-score, and ARI of the compared methods from high to low and obtain the average ranking of each method in terms of all data sets.
When α = 0.05, the threshold for the Friedman test was 1.8544. According to Equation (14), the Γ F values can be calculated for different clustering evaluation metrics (ACC, NMI, F-score, and ARI), which are 8.4348, 13.0826, 5.9892 and 11.3655, respectively. These Γ F values are all greater than the threshold of the Friedman test, which rejects the hypothesis that all the methods being compared hold the same performance. Then, we perform the Nemenyi post-hoc test to further distinguish multiple methods. After obtaining the critical distance, CD = 6.2982, according to Equation (15), we can draw the Friedman test chart as Figure 2. For each method, the blue dot marks its average rank. The horizontal lines with the dot at the center indicate the critical distance, CD. If the lines do not have overlapping areas this indicates a significant difference in the comparison methods.
From the figure, we can see that there are significant differences between the method in this paper and Co_reg_c, Co_reg_p, MLRSSC, and PMSC, and that the other methods do not differ from one another significantly. Compared with other methods, our method has the best average ranking regardless of the clustering evaluation index. In summary, our proposed method holds statistical advantages.

5.4. Running Time Analysis

We used MATLAB 2018b to run each clustering method independently and recorded the running time of each clustering method under one hyperparameter combination on each data set in Figure 3. From these results, one can see that FeatConcate is the fastest among most of multi-view data sets. To explore the consistency or complementary information of the multi-view data, these multi-view clustering methods generally need a relatively long running time to produce the final clustering. Our LGOMSC method is faster than Co_reg_c, Co_reg_p, LMSC, FMR, MLRSSC, RMKMC, mPAC, PMSC, and COMVSC on most data sets. On the Caltech101 data set, our LGOMSC method is 4.35 s slower than MLRSSC, and on the Wikipedia data set, our LGOMSC method is 1.89 s slower than MLRSSC and 0.21 s slower than RMKMC. On these data sets, LMVSC and GMC are the fastest, and they are more suitable for solving large-scale clustering problems. They are more concerned with efficiency rather than effectiveness. Therefore, the clustering performance is relatively poor. Although our proposed method is slower than the methods designed specifically for large-scale scenarios, our method is still comparable to LMVSC and GMC for the NottingHill data set. In Table 7, MVSCLG costs more running times than our method. In the NottingHill data set, LGOMSC costs 4744.49 s under one hyperparameter combination. However, since MVSCLG contains three hyperparameters, we need to conduct the grid search strategy to select the optimal one from 420 hyperparameter combinations, which may need 553.52 h. Hence, we ignore it.

5.5. Memory Usage Analysis

We use MATLAB 2018b to run each clustering method independently and recorded the memory usage of each clustering method. Specifically, we recorded the current matlab memory usage once before we ran the program. After the program has finished, we recorded the current matlab memory usage again. Subtracting the first reading from the second reading gives us the memory usage of the method. For example, on the 3sources data set, as can be seen in Figure 4, the memory usage of our method is small compared with the LMVSC, PMSC, COMVSC and MVSCLG methods. The memory usage of our method is slightly larger than the other remaining methods, but there is no significant difference between our method and the other remaining methods in terms of memory usage by an order of magnitude. Therefore, our method has an appropriate space complexity.

5.6. Convergence Study

The objective function of LGOMSC has multiple variables and constraints. We have developed an effective iterative optimization algorithm to solve the proposed method. We conduct the convergence experiment in Figure 5, which provides the convergence curves of LGOMSC on nine multi-view data sets. The x-axis displays the number of iterations, and the y-axis displays the corresponding objective function value. From these results, one can see that the proposed method is well convergent and converges quickly. Within 10 iterations, the objective function value can converge to a stable value.

5.7. Parameter Tuning

We conducted an experiment to analyze the hyperparameter λ 1 in this paper. In the experiment, we tuned λ 1 from a candidate set of { 1 , 10 , 20 , , 100 } . As shown in Figure 6, we give the parameter tuning of LGOMSC on nine multi-view data sets and show the clustering performance under different values. One can see that the proposed method keeps a relatively robust clustering performance under a large range of λ 1 . This helps us to easily select a proper parameter.

6. Conclusions

In this paper, we propose a novel one-step multi-view subspace clustering method, which integrates the self-expression model and graph learning to simultaneously exploit the local and global information of subspace representations from multi-view data. Moreover, to further exploit the hidden relationships between different steps to achieve an end-to-end clustering, our method integrates the subspace representation learning, multi-view information fusion, and clustering tasks into a joint framework. Experimental results on nine popular multi-view data sets confirm the effectiveness of our method by comparing with many baseline methods.
In the future, we have two directions to improve our method. First, since our method is a linear model, we will consider expanding our model to non-linear cases to deal with complex multi-view data. Second, for better fitting for large-scale multi-view subspace clustering, we will adopt anchor-based ideas to improve our method.

Author Contributions

Conceptualization, Y.D. and H.Y.; methodology, H.Y. and C.S.L.; software, Y.D.; experiment, validation and analysis, Y.D. and H.Y.; investigation, Y.D. and H.Y.; resources, H.Y.; data curation, Y.D.; writing—original draft preparation, Y.D. and H.Y.; writing—review and editing, Y.D., H.Y., C.S.L. and L.L.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Natural Science Foundation of China under Grant 61903091; Guangdong Basic and Applied Basic Research Foundation (No. 2020A1515010801).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original contributions presented in the study are included in the article, further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Goldberg, D.; Holland, J. Genetic algorithms and machine learning. Mach. Learn. 1988, 3, 95–99. [Google Scholar] [CrossRef]
  2. Liu, S.; Liang, X.; Liu, L.; Shen, X.; Yang, J.; Xu, C.; Lin, L.; Cao, X.; Yan, S. Matching-cnn meets knn: Quasi-parametric human parsing. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 1419–1427. [Google Scholar]
  3. Witten, I.; Frank, E. Data Mining: Practical Machine Learning Tools and Techniques; Morgan Kaufmann: San Mateo, CA, USA, 2005; pp. 35–40. [Google Scholar]
  4. Berkhin, P. A survey of clustering data mining techniques. In Grouping Multidimensional Data; Kogan, J., Nicholas, C., Teboulle, M., Eds.; Springer: Berlin/Heidelberg, Germany, 2006; pp. 25–71. [Google Scholar]
  5. Astolfi, D.; Pandit, R. Multivariate wind turbine power curve model based on data clustering and polynomial LASSO regression. Appl. Sci. 2021, 12, 72. [Google Scholar] [CrossRef]
  6. Dinh, D.; Fujinami, T.; Huynh, V. Estimating the optimal number of clusters in categorical data clustering by silhouette coef-ficient. In Proceedings of the Twentieth International Symposium on Knowledge and Systems Sciences (ISKSS), Da Nang, China, 29 November–1 December 2019; pp. 1–17. [Google Scholar]
  7. Liu, G.; Lin, Z.; Yan, S.; Sun, J.; Yu, Y.; Ma, Y. Robust recovery of subspace structures by low-rank representation. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 171–184. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  8. Chao, G.; Sun, S.; Bi, J. A survey on multiview clustering. IEEE Trans. Artif. Intell. 2021, 2, 146–168. [Google Scholar] [CrossRef]
  9. Yang, Y.; Wang, H. Multi-view clustering: A survey. Big Data Min. Anal. 2018, 1, 83–107. [Google Scholar]
  10. Fu, L.; Lin, P.; Vasilakos, A.; Wang, S. An overview of recent multi-view clustering. Neurocomputing 2020, 402, 148–161. [Google Scholar] [CrossRef]
  11. Yan, S.; Xu, D.; Zhang, B.; Zhang, H.; Yang, Q.; Lin, S. Graph embedding and extensions: A general framework for dimensionality reduction. IEEE Trans. Pattern Anal. Mach. Intell. 2006, 29, 40–51. [Google Scholar] [CrossRef] [Green Version]
  12. Gao, H.; Nie, F.; Li, X.; Huang, H. Multi-view subspace clustering. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 4238–4246. [Google Scholar]
  13. Cao, X.; Zhang, C.; Fu, H.; Liu, S.; Zhang, H. Diversity-induced multi-view subspace clustering. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 586–594. [Google Scholar]
  14. Luo, S.; Zhang, C.; Zhang, W.; Cao, X. Consistent and specific multi-view subspace clustering. In Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence (AAAI-18), New Orleans, LA, USA, 2–7 February 2018; pp. 3730–3737. [Google Scholar]
  15. Wang, X.; Guo, X.; Lei, Z.; Zhang, C.; Li, S. Exclusivity-consistency regularized multi-view subspace clustering. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 923–931. [Google Scholar]
  16. Guo, J.; Sun, Y.; Gao, J.; Hu, Y.; Yin, B. Rank Consistency induced multiview subspace clustering via low-rank matrix factorization. IEEE Trans. Neural Netw. Learn. Syst. 2021, 1–14. [Google Scholar] [CrossRef]
  17. Brbić, M.; Kopriva, I. Multi-view low-rank sparse subspace clustering. Pattern Recognit. 2018, 73, 247–258. [Google Scholar] [CrossRef] [Green Version]
  18. Zhang, C.; Fu, H.; Liu, S.; Liu, G.; Cao, X. Low-rank tensor constrained multiview subspace clustering. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 1582–1590. [Google Scholar]
  19. Xie, Y.; Tao, D.; Zhang, W.; Liu, Y.; Zhang, L.; Qu, Y. On unifying multi-view self-representations for clustering by tensor multi-rank minimization. Int. J. Comput. Vis. 2018, 126, 1157–1179. [Google Scholar] [CrossRef] [Green Version]
  20. Zhang, C.; Fu, H.; Wang, J.; Li, W.; Cao, X.; Hu, Q. Tensorized multi-view subspace representation learning. Int. J. Comput. Vis. 2020, 128, 2344–2361. [Google Scholar] [CrossRef] [Green Version]
  21. Yin, M.; Gao, J.; Xie, S.; Guo, Y. Multiview subspace clustering via tensorial t-product representation. IEEE Trans. Neural Netw. Learn. Syst. 2018, 30, 851–864. [Google Scholar] [CrossRef]
  22. Kang, Z.; Zhao, X.; Peng, C.; Zhu, H.; Zhou, J.; Peng, X.; Chen, W.; Xu, Z. Partition level multiview subspace clustering. Neural Netw. 2020, 122, 279–288. [Google Scholar] [CrossRef]
  23. Lv, J.; Kang, Z.; Wang, B.; Ji, L.; Xu, Z. Multi-view subspace clustering via partition fusion. Inf. Sci. 2021, 560, 410–423. [Google Scholar] [CrossRef]
  24. Zhang, P.; Liu, X.; Xiong, J.; Zhou, S.; Zhao, W.; Zhu, E.; Cai, Z. Consensus one-step multi-view subspace clustering. IEEE Trans. Knowl. Data Eng. 2020, 1–14. [Google Scholar] [CrossRef]
  25. Kang, Z.; Guo, Z.; Huang, S.; Wang, S.; Chen, W.; Su, Y.; Xu, Z. Multiple partitions aligned clustering. In Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence (IJCAI), Macao, China, 10–16 August 2019; pp. 2701–2707. [Google Scholar]
  26. Kang, Z.; Zhou, W.; Zhao, Z.; Shao, J.; Han, M.; Xu, Z. Large-scale multi-view subspace clustering in linear time. In Proceedings of the Thirty-Fourth AAAI Conference on Artificial Intelligence (AAAI-20), New York, NY, USA, 7–12 February 2020; pp. 4412–4419. [Google Scholar]
  27. Wang, S.; Liu, X.; Zhu, X.; Zhang, P.; Zhang, Y.; Gao, F.; Zhu, E. Fast parameter-free multi-view subspace clustering with consensus anchor guidance. IEEE Trans. Image Process. 2022, 31, 556–568. [Google Scholar] [CrossRef]
  28. Sun, M.; Zhang, P.; Wang, S.; Zhou, S.; Tu, W.; Liu, X.; Zhu, E.; Wang, C. Scalable multi-view subspace clustering with unified anchors. In Proceedings of the 29th ACM International Conference on Multimedia, Chengdu, China, 20–24 October 2021; pp. 3528–3536. [Google Scholar]
  29. Liu, J.; Wang, C.; Gao, J.; Han, J. Multi-view clustering via joint nonnegative matrix factorization. In Proceedings of the 2013 SIAM International Conference on Data Mining (SDM), Austin, TX, USA, 2–4 May 2013; pp. 252–260. [Google Scholar]
  30. Guo, Y. Convex subspace representation learning from multi-view data. In Proceedings of the Twenty-Seventh AAAI Conference on Artificial Intelligence Bellevue, Washington, DC, USA, 14–18 July 2013; pp. 387–393. [Google Scholar]
  31. Wang, J.; Tian, F.; Yu, H.; Liu, C.; Zhan, K.; Wang, X. Diverse non-negative matrix factorization for multiview data representation. IEEE Trans. Cybern. 2018, 48, 2620–2632. [Google Scholar]
  32. Nie, F.; Shi, S.; Li, X. Auto-weighted multi-view co-clustering via fast matrix factorization. Pattern Recognit. 2020, 102, 107207. [Google Scholar] [CrossRef]
  33. Liu, X.; Song, P.; Sheng, C.; Zhang, W. Robust multi-view non-negative matrix factorization for clustering. Digit. Signal Process. 2022, 123, 103447. [Google Scholar] [CrossRef]
  34. Xu, C.; Tao, D.; Xu, C. Multi-view intact space learning. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 2531–2544. [Google Scholar] [CrossRef] [Green Version]
  35. Zhang, C.; Hu, Q.; Fu, H.; Zhu, P.; Cao, X. Latent multi-view subspace clustering. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 4279–4287. [Google Scholar]
  36. Li, R.; Zhang, C.; Hu, Q.; Zhu, P.; Wang, Z. Flexible multi-view representation learning for subspace clustering. In Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence (IJCAI), Macao, China, 10–16 August 2019; pp. 2916–2922. [Google Scholar]
  37. Xie, D.; Zhang, X.; Gao, Q.; Han, J.; Xiao, S.; Gao, X. Multiview clustering by joint latent representation and similarity learning. IEEE Trans. Cybern. 2019, 50, 4848–4854. [Google Scholar]
  38. Tang, W.; Lu, Z.; Dhillon, I.S. Clustering with multiple graphs. In Proceedings of the 2009 Ninth IEEE International Conference on Data Mining, Miami, FL, USA, 6–9 December 2009; pp. 1016–1021. [Google Scholar]
  39. Nie, F.; Cai, G.; Li, X. Multi-view clustering and semi-supervised classification with adaptive neighbours. In Proceedings of the Thirty-First AAAI Conference (AAAI-17), San Francisco, CA, USA, 4–9 February 2017; pp. 2408–2414. [Google Scholar]
  40. Nie, F.; Li, J.; Li, X. Self-weighted multiview clustering with multiple graphs. In Proceedings of the Twenty-Sixth International Joint Conference on Artificial Intelligence, Melbourne, Australia, 19–25 August 2017; pp. 2564–2570. [Google Scholar]
  41. Zhan, K.; Zhang, C.; Guan, J.; Wang, J. Graph learning for multiview clustering. IEEE Trans. Cybern. 2017, 48, 2887–2895. [Google Scholar] [CrossRef]
  42. Zhan, K.; Nie, F.; Wang, J.; Yang, Y. Multiview consensus graph clustering. IEEE Trans. Image Process. 2018, 28, 1261–1270. [Google Scholar] [CrossRef]
  43. Wang, H.; Yang, Y.; Liu, B. GMC: Graph-based multi-view clustering. IEEE Trans. Knowl. Data Eng. 2019, 32, 1116–1129. [Google Scholar] [CrossRef]
  44. Zhao, Y.; Yun, Y.; Zhang, X.; Li, Q.; Gao, Q. Multi-view spectral clustering with adaptive graph learning and tensor schatten p-norm. Neurocomputing 2022, 468, 257–264. [Google Scholar] [CrossRef]
  45. Duan, Y.; Yuan, H.; Lai, L.; He, B. Multi-view subspace clustering with local and global information. In Proceedings of the International Conference on Wavelet Analysis and Pattern Recognition (ICWAPR), Adelaide, Australia, 3–5 December 2021; pp. 1–6. [Google Scholar]
  46. Nie, F.; Wang, X.; Jordan, M.; Huang, H. The constrained Laplacian rank algorithm for graph-based clustering. In Proceedings of the Thirtieth AAAI Conference on Artificial Intelligence (AAAI-16), Phoenix, AZ, USA, 12–17 February 2016; pp. 1969–1976. [Google Scholar]
  47. Mohar, B.; Alavi, Y.; Chartrand, G.; Oellermann, O. The Laplacian spectrum of graphs. Graph Theory Comb. Appl. 1991, 2, 871–898. [Google Scholar]
  48. Fan, K. On a theorem of Weyl concerning eigenvalues of linear transformations I. Proc. Natl. Acad. Sci. USA 1949, 35, 652–655. [Google Scholar] [CrossRef] [Green Version]
  49. Huang, J.; Nie, F.; Huang, H. A new simplex sparse learning model to measure data similarity for clustering. In Proceedings of the the Twenty-fourth International Joint Conference on Artificial Intelligence, Buenos Aires, Argentina, 25–31 July, 2015; pp. 3569–3575. [Google Scholar]
  50. Wang, H.; Yang, Y.; Liu, B.; Fujita, H. A study of graph-based system for multi-view clustering. Knowl.-Based Syst. 2019, 163, 1009–1019. [Google Scholar] [CrossRef]
  51. Nie, F.; Wang, X.; Huang, H. Clustering and projected clustering with adaptive neighbors. In Proceedings of the 20th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, New York, NY, USA, 24–27 August 2014; pp. 977–986. [Google Scholar]
  52. Kumar, A.; Rai, P.; Daumé, H. Co-regularized multi-view spectral clustering. In Proceedings of the 25th Annual Conference on Neural Information Processing Systems, Granada, Spain, 12–15 December 2011; pp. 1413–1421. [Google Scholar]
  53. Cai, X.; Nie, F.; Huang, H. Multi-view k-means clustering on big data. In Proceedings of the Twenty-Third International Joint Conference on Artificial Intelligence, Beijing, China, 3–9 August 2013; pp. 724–730. [Google Scholar]
  54. Yang, Y.; Song, J.; Huang, Z.; Ma, Z.; Sebe, N.; Hauptmann, A.G. Multi-feature fusion via hierarchical regression for multimedia analysis. IEEE Trans. Multimed. 2013, 15, 572–581. [Google Scholar] [CrossRef]
Figure 2. Friedman test charts. (a) Friedman test on ACC, (b) Friedman test on NMI, (c) Friedman test on F-score, (d) Friedman test on ARI.
Figure 2. Friedman test charts. (a) Friedman test on ACC, (b) Friedman test on NMI, (c) Friedman test on F-score, (d) Friedman test on ARI.
Applsci 12 05094 g002
Figure 3. Running time of different methods on nine multi-view data sets.
Figure 3. Running time of different methods on nine multi-view data sets.
Applsci 12 05094 g003
Figure 4. The memory usage representation of compared methods on 3sources.
Figure 4. The memory usage representation of compared methods on 3sources.
Applsci 12 05094 g004
Figure 5. Convergence performance on nine multi-view data sets. (a) 3sources, (b) 100leaves, (c) BBC, (d) Caltech101, (e) COIL-20, (f) NottingHill, (g) Webkb, (h) Cornell, (i) Wikipedia.
Figure 5. Convergence performance on nine multi-view data sets. (a) 3sources, (b) 100leaves, (c) BBC, (d) Caltech101, (e) COIL-20, (f) NottingHill, (g) Webkb, (h) Cornell, (i) Wikipedia.
Applsci 12 05094 g005
Figure 6. The parameter sensitivity of λ 1 on nine multi-view data sets. (a) 3sources, (b) 100leaves, (c) BBC, (d) Caltech101, (e) COIL-20, (f) NottingHill, (g) Webkb, (h) Cornell, (i) Wikipedia.
Figure 6. The parameter sensitivity of λ 1 on nine multi-view data sets. (a) 3sources, (b) 100leaves, (c) BBC, (d) Caltech101, (e) COIL-20, (f) NottingHill, (g) Webkb, (h) Cornell, (i) Wikipedia.
Applsci 12 05094 g006aApplsci 12 05094 g006b
Table 1. Notations and abbreviations.
Table 1. Notations and abbreviations.
NotationDefinition
I n n × n Identity matrix
1 All-ones column vector
nNumber of data sample
cNumber of clusters
VNumber of views
d v Feature dimension of the v-th view
X v d v × n Feature matrix of the v-th view
x i , : , x j , x i j Represented   as   the   i - th   row ,   j - th   column ,   and   ij - th   element   of   matrix   X , respectively
X T The transpose of a matrix
S v n × n Subspace representation matrix of the v-th view
U n × n Common subspace representation matrix
F The Frobenius norm
T r ( ) Trace operator of a matrix
d i a g ( ) Vector of the diagonal elements of a matrix
r a n k ( ) The rank of a matrix
Table 2. Summary of nine multi-view benchmark data sets (dv denotes the dimensionality of the v-th view).
Table 2. Summary of nine multi-view benchmark data sets (dv denotes the dimensionality of the v-th view).
Data SetPointClassViewd1d2d3d4d5d6
3sources16963356036313068
100leaves16001003646464
BBC685544659463346654684
Caltech10114747648402541984512928
COIL-201440203102433046750
NottingHill466053675033042000
Webkb10512218403000
Cornell195521951703
Wikipedia69310212810
Table 3. The clustering performance comparison in terms of ACC on nine multi-view data sets.
Table 3. The clustering performance comparison in terms of ACC on nine multi-view data sets.
ACC (%)3sources100leavesBBCCaltech101COIL-20NottingHillWebkbCornellWikipedia
FeatConcate65.0971.0061.4654.2767.5091.9394.7743.0857.72
Co-reg_c69.1778.5334.7442.0070.3674.7780.4238.4138.59
Co-reg_p66.1875.6035.9942.1472.4272.1483.2436.2620.70
LMSC71.6077.0086.2853.8075.3583.7895.3443.5956.85
FMR70.4169.2585.1147.6972.0182.8593.2443.0856.85
MLRSSC34.881.4433.1454.215.0730.1178.0243.0815.22
RMKMC54.441.0060.4454.1461.6075.4394.0143.5961.04
mPAC76.9247.0658.1059.3673.4090.2878.1245.6456.71
LMVSC63.3171.0684.3856.7274.1789.2595.6255.9059.16
PMSC63.8522.4634.4544.4549.0970.2178.0245.4419.70
COMVSC65.0970.8869.4977.5477.6481.7082.8755.3860.46
GMC65.0986.3869.0565.7487.5731.2477.6438.9731.89
Ours83.4394.3188.4779.8592.08100.0098.3864.6262.63
Table 4. The clustering performance comparison in terms of NMI on nine multi-view data sets.
Table 4. The clustering performance comparison in terms of NMI on nine multi-view data sets.
NMIn (%)3sources100leavesBBCCaltech101COIL-20NottingHillWebkbCornellWikipedia
FeatConcate56.5387.6460.6356.1879.1586.6666.1819.0254.04
Co-reg_c55.0392.0413.3843.6881.6969.798.7912.6026.22
Co-reg_p50.8590.046.5943.5982.1767.7318.9111.577.54
LMSC69.1889.2265.7051.8584.5478.5770.3618.8952.60
FMR57.3485.4466.0446.7778.5366.3859.5921.4451.81
MLRSSC5.7513.331.032.112.660.230.084.862.31
RMKMC40.230.0054.3863.1679.0675.2863.8127.9955.09
mPAC64.1373.9647.4150.5685.8683.1416.7915.4847.53
LMVSC59.3587.4969.4555.7082.2482.3369.4127.8452.81
PMSC48.7564.024.1921.3870.3165.737.9510.426.47
COMVSC50.6987.1957.6752.7087.5075.1623.7324.5153.58
GMC53.7395.3755.6253.7796.319.230.1715.9029.98
Ours77.7497.4375.2054.2397.44100.0084.8439.4056.06
Table 5. The clustering performance comparison in terms of F-score on nine multi-view data sets.
Table 5. The clustering performance comparison in terms of F-score on nine multi-view data sets.
F-score (%)3sources100leavesBBCCaltech101COIL-20NottingHillWebkbCornellWikipedia
FeatConcate66.6163.5460.3256.9161.0588.7792.7833.2750.34
Co-reg_c69.0474.4637.0042.8368.2072.3879.6632.0626.96
Co-reg_p66.0169.6737.7442.3469.6769.0881.7131.1213.00
LMSC65.5868.7276.8153.5871.3383.0493.0242.5948.96
FMR63.8959.0776.5547.2466.5872.1990.0834.5048.22
MLRSSC37.491.9737.8855.939.4236.0379.2742.8819.57
RMKMC47.891.8657.4655.6458.8472.5891.8735.9451.85
mPAC76.9336.6159.1061.1968.7388.1779.1843.3445.81
LMVSC55.8561.6778.4451.9069.2985.3693.8349.5750.13
PMSC57.0518.7738.4843.4649.5666.8479.2743.5519.67
COMVSC56.7462.9562.7271.8371.7980.1181.4746.1851.02
GMC52.8866.2763.0561.5485.3136.9478.6737.0623.00
Ours78.3488.9182.8375.7291.86100.0097.5857.3252.48
Table 6. The clustering performance comparison in terms of ARI on nine multi-view data sets.
Table 6. The clustering performance comparison in terms of ARI on nine multi-view data sets.
ARI (%)3sources100leavesBBCCaltech101COIL-20NottingHillWebkbCornellWikipedia
FeatConcate56.9063.1749.2041.9458.8985.6776.9912.3344.28
Co-reg_c58.3874.206.0527.0366.4664.7016.206.5718.39
Co-reg_p54.8169.361.5226.4668.0360.7127.192.472.80
LMSC56.7268.4169.7037.8069.7478.2080.7910.1842.89
FMR53.7058.6669.1331.8664.8364.6872.8412.0142.12
MLRSSC0.330.12−0.040.930.030.04−0.141.33−0.09
RMKMC34.270.0045.6141.2156.4564.6973.6115.7345.97
mPAC69.7935.8740.2745.1967.0884.7928.178.9738.86
LMVSC43.1661.3671.7534.4867.6781.1580.8132.5444.24
PMSC36.7017.471.1717.5846.3357.260.1810.441.96
COMVSC36.9462.5547.3650.1770.2874.5025.5217.6844.69
GMC32.8765.8647.4639.6484.452.211.023.476.07
Ours70.6388.7977.0657.5491.42100.0092.9235.0846.13
Table 7. Comparing LGOMSC with MVSCLG on nine multi-view data sets.
Table 7. Comparing LGOMSC with MVSCLG on nine multi-view data sets.
Methods 3sources100leavesBBCCaltech101COIL-20NottingHillWebkbCornellWikipedia
MVSCLGACC78.7073.3880.8882.5678.61-91.8270.7763.49
NMI65.1388.5458.1458.0690.28-56.4949.8859.09
F-score76.1366.1069.9579.0772.94-88.0465.4056.16
ARI68.1065.7461.1861.7671.34-67.8951.3549.69
Time (s)59.77947.71209.86764.76443.214744.49163.0540.27107.10
OursACC83.4394.3188.4779.8592.08100.0098.3864.6262.63
NMI77.7497.4375.2054.2397.44100.0084.8439.4056.06
F-score78.3488.9182.8375.7291.86100.0097.5857.3252.48
ARI70.6388.7977.0657.5491.42100.0092.9235.0846.13
Time (s)0.5317.033.9127.877.79102.34.540.564.15
“-” indicates that MVSCLG requires more than 553.52 h on the NottingHill data set.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Duan, Y.; Yuan, H.; Lai, C.S.; Lai, L.L. Fusing Local and Global Information for One-Step Multi-View Subspace Clustering. Appl. Sci. 2022, 12, 5094. https://0-doi-org.brum.beds.ac.uk/10.3390/app12105094

AMA Style

Duan Y, Yuan H, Lai CS, Lai LL. Fusing Local and Global Information for One-Step Multi-View Subspace Clustering. Applied Sciences. 2022; 12(10):5094. https://0-doi-org.brum.beds.ac.uk/10.3390/app12105094

Chicago/Turabian Style

Duan, Yiqiang, Haoliang Yuan, Chun Sing Lai, and Loi Lei Lai. 2022. "Fusing Local and Global Information for One-Step Multi-View Subspace Clustering" Applied Sciences 12, no. 10: 5094. https://0-doi-org.brum.beds.ac.uk/10.3390/app12105094

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop