Next Article in Journal
A Modified TurboEdit Cycle-Slip Detection and Correction Method for Dual-Frequency Smartphone GNSS Observation
Previous Article in Journal
Experimental Characterization of RGB LED Transceiver in Low-Complexity LED-to-LED Link
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Adaptive Weighted Graph Fusion Incomplete Multi-View Subspace Clustering

School of Computer, National University of Defense Technology, Changsha 410073, China
*
Author to whom correspondence should be addressed.
Sensors 2020, 20(20), 5755; https://doi.org/10.3390/s20205755
Submission received: 24 August 2020 / Revised: 29 September 2020 / Accepted: 6 October 2020 / Published: 10 October 2020
(This article belongs to the Section Intelligent Sensors)

Abstract

:
With the enormous amount of multi-source data produced by various sensors and feature extraction approaches, multi-view clustering (MVC) has attracted developing research attention and is widely exploited in data analysis. Most of the existing multi-view clustering methods hold on the assumption that all of the views are complete. However, in many real scenarios, multi-view data are often incomplete for many reasons, e.g., hardware failure or incomplete data collection. In this paper, we propose an adaptive weighted graph fusion incomplete multi-view subspace clustering (AWGF-IMSC) method to solve the incomplete multi-view clustering problem. Firstly, to eliminate the noise existing in the original space, we transform complete original data into latent representations which contribute to better graph construction for each view. Then, we incorporate feature extraction and incomplete graph fusion into a unified framework, whereas two processes can negotiate with each other, serving for graph learning tasks. A sparse regularization is imposed on the complete graph to make it more robust to the view-inconsistency. Besides, the importance of different views is automatically learned, further guiding the construction of the complete graph. An effective iterative algorithm is proposed to solve the resulting optimization problem with convergence. Compared with the existing state-of-the-art methods, the experiment results on several real-world datasets demonstrate the effectiveness and advancement of our proposed method.

Graphical Abstract

1. Introduction

Traditional clustering methods [1,2,3,4] usually use a single view to measure the similarity of samples. With the rapid progress of data collection, individual features are not enough to describe data points. Multiple views usually contain supplementary information, which may be beneficial to explore the basic structure of the data. With the development of information technology, data mining and other technologies, many datasets in the real-world can be presented from different perspectives, called multi-view data. For example, the same text can be expressed in various languages. In biometric recognition scope, faces, fingerprints, palm prints and iris could form the different views of multi-view data. In the field of medical diagnosis, different examinations of patients can be regarded as different views. Multi-view data could provide sufficient information than the traditional single feature representation in revealing the underlying clustering structure. Furthermore, distinct views contain specific information of intra-view and complementary information of inter-view, which are negotiated with each other to boost the performance of clustering [5,6,7,8,9,10,11,12,13,14].
Based on different mechanisms, we can divide the existing multi-view clustering methods into four categories. The first category methods refer to multi-kernel clustering. These methods usually combine multiple pre-defined kernels to reach optimal clustering results [12,15,16,17]. The second kind of approach is co-training and co-regularized [18,19]. They iteratively learn multiple clustering results that can provide predicted clustering indices for the unlabeled data from different views. In this way, the clustering results are forced to be consistent across views. The third strategy collaboratively transforms the multi-view information into a compact common binary code space. Then, the clustering process is measured in the Hamming space and enjoys superior algorithm acceleration [20,21]. The last mechanism is the subspace-based multi-view clustering method. It assumes that high-dimensional data points are drawn from various low-dimensional subspaces, and each cluster can be drawn from one of the subspaces [22,23,24,25]. The essential idea is to find several low-dimensional representations embedded in latent spaces and finally attain a united representation for downstream clustering tasks [26]. Besides, aiming at finding a shared low-dimensional latent representation via matrix decomposition, the non-negative matrix factorization (NMF) [27]-based multi-view clustering methods [28,29,30,31] can also be seen as a branch of the subspace-based multi-view clustering method.
Although the algorithms mentioned above have achieved great success in different scenarios, these traditional multi-view clustering algorithms cannot effectively deal with multi-view data with incomplete features. Therefore, the incomplete multi-view clustering algorithms [32,33,34] have attracted extensive attention. To the best of our knowledge, existing incomplete multi-view clustering algorithms can be classified into two categories: non-negative matrix factorization based methods and graph-based methods. The NMF-based methods aim at directly obtaining a common low-dimensional representation through non-negative matrix decomposition. Most of them take the strategy of combining view-specific and common representations into a unified one [35,36,37]. Another representative approach of the NMF-based method is to fill the missing data with average feature values and then use the weighted non-negative matrix factorization to reduce the impact of the missing samples [38]. These NMF-based methods can directly obtain a consistent representation with incomplete samples. However, it is limited to the following two points: (1) when the number of views is more than two, the common parts of views will be significantly reduced and cannot be learned a shared representation between views; (2) NMF-based methods usually neglect the intrinsic structure of data, resulting in an uncompacted representation.
The graph-based incomplete multi-view clustering algorithms are more effective in exploring the geometric structure of data than NMF-based methods. The construction of the graph is essential for the success of clustering. However, it is impossible to construct a complete graph connecting all samples due to the lack of partial samples in incomplete multi-view clustering. To cover this problem, Gao et al. [39] first fill the missing parts and then learn graphs and representations. Zhao et al. [36] utilize NMF to obtain consistent representations to guide the generation of graphs with local structures. However, when the missing rate is high, the filling strategy will dominate the learning of the representation, resulting in the filled samples being connected with each other. Moreover, information fusion refers to fusing multiple sources to achieve consistency. In this stage, multiple views are treated equally, which is unreasonable in real applications.
To address the above issues, we propose a novel incomplete multi-view clustering method, constructing the graphs between instances in the latent embedding subspace. In this manner, we can deal with multi-view data with any number of views. Furthermore, an adaptive weighted mechanism is induced to fuse the graphs with local-structure into a complete graph.
In this manner, we establish a relation between missing and unmissing samples. An additional sparse regularization term is imposed on the consensus complete graph to eliminate the adverse effects of inconsistency between views and noise or outliers from each view. Specifically, the framework of this paper is illustrated in Figure 1.
Compared with existing methods, the proposed adaptive weighted graph fusion incomplete multi-view subspace clustering (AWGF-IMSC) algorithm has the following contributions:
  • It induces the similarity graph fusion after obtaining latent spaces to extract the local structure of inner views. By virtue of it, noise existing in the original space can be eliminated in latent space and contribute to better graph construction.
  • It incorporates relations between missing samples and complete samples into the complete graph. The sparse constraint imposed on the complete graph improves the view-inconsistency and reduces the disagreements between views, making the proposed method more robust in most cases.
  • The importance of each view is automatically learned and adaptively optimized during the optimization. Consequently, the important view has strong guidance in the learning process. Moreover, there is no limitation to the number of views in our approach. The proposed method is applicable to any multi-view datasets.
The rest of the paper is organized as follows. The next Section 2 denotes the notations and symbols used in this paper. Section 3 introduces methods mostly related to our work. The proposed algorithm and its optimization process are formulated in Section 4. Besides, we also give the analysis of convergence and complexity of the proposed algorithm in this part. Extensive experiment results and analysis are shown in Section 5, before conclusion and prospectives.

2. Notation

For clarity, we give the notation used throughout the paper at the beginning. We use bold letters to represent matrices and vectors. For matrix A , A : , j and A i , j represent its j-th column and ( i , j ) element, respectively. A , Tr ( A ) and A 1 denote the transpose, trace and the inverse operations on matrix A , respectively. · F denotes the Frobenius norm. The 2 , 1 norm is denoted as X 2 , 1 = i = 1 n j = 1 t X i , j 2 = i = 1 n X i , : 2 . Moreover, operator A + turns the negative elements in matrix A to 0 while maintaining the non-negative elements, and vice versa. For multi-view datasets A = { A ( 1 ) , A ( 2 ) , , A ( m ) } , the superscript ( i ) represents the i-th view. In individual A ( i ) R d i × N , each column indicates an instance. N is the number of the samples and d i represents the feature dimension of corresponding i-th view.

3. Related Work

In this section, we will present the work most relevant to the proposed method, i.e., semi-non-negative matrix factorization and subspace learning.

3.1. Semi-Non-Negative Matrix Factorization for Single View

Non-negative matrix factorization (NMF) is a significant branch in the field of matrix factorization. NMF aims at finding two non-negative matrix U R + d × K and V R + K × N to roughly approximate the original data matrix, i.e., X U V . Since many real-world datasets are usually high-dimensional, the NMF methods have been widely applied in image analysis [40], data mining, speech denoising [41] and population genetics, etc. The semi-NMF [42] is an extension of traditional NMF, which only requires the coefficient matrix to be non-negative. Specifically, given the data matrix X R d × N , the semi-NMF utilizes the base matrix U R d × k and the non-negative coefficient matrix V R k × N to approximate the matrix X :
min U , V X U V F 2 s . t . V 0
.
Ding et al. [42] further propose an iterative optimization algorithm to find the local optimal solution. The updating strategy can be concluded as follows:
With V being fixed, U can be updated by U = X V ( V V ) 1 .
With U being fixed, V can be updated by V i , j V i , j ( X U ) i , j + + [ V ( U U ) ] i , j ( X U ) i , j + [ V ( U U ) + ] i , j .
The positive and negative elements of matrix M are denoted as M i , j + and M i , j . And they hold on the property M i , j = M i , j + M i , j .
NMF and semi-NMF methods are also employed universally in multi-view clustering. Many of the multi-view clustering (MVC) methods utilize NMF to reduce dimension on each view or directly reach a consistent latent presentation [28,43]. Especially in the incomplete multi-view scenario, NMF and semi-NMF play significant roles in achieving a consistent representation from different incomplete views. Li et al. [35] learn a shared representation for the paired instances and view-specific representations for unpaired instances via NMF. The complete latent representation can be attained by combining shared and view-specific representations. The method in [44] utilizes weighted semi-NMF to reach a consensus representation. Then, the 2 , 1 norm regularized regression is imposed to align the different basis matrices. Although these NMF-based methods could learn a consensus representation from the incomplete views, the number of views and the absence of local structure limit their performance.

3.2. Subspace Clustering

Subspace clustering is an extension of the traditional clustering method which aims at grouping data in different subspaces [45,46]. The self-representation property [47] of subspace clustering aims to represent data points by the linear combinations of themselves. The formulation can be expressed as:
min Z X X Z F 2 + β R ( Z ) s . t . 0 Z i , j 1 , Z 1 = 1 ,
where X R d × N is the original data, Z is the self-representation coefficient matrix, with each column being a new representation for corresponding data point. β > 0 is a trade-off parameter. Since Z reflects the correlations among samples, it can be regarded as a graph and then we can perform spectral clustering algorithm on it to get the final clustering result.

3.3. Incomplete Multi-View Spectral Clustering with Adaptive Graph Learning (IMSC-AGL)

In paper [48], a novel graph-based multi-view clustering method is proposed to deal with incomplete multi-view scenarios termed incomplete multi-view spectral clustering with adaptive graph learning (IMSC-AGL). IMSC-AGL optimizes the shared graph from the low-dimensional representations individually formed by each view. Moreover, a nuclear-norm constraint is introduced to ensure the low-rank property of the ideal graph. The mathematical formulation can be written as,
min Z ( v ) , E ( v ) , F ( v ) , M v Z ( v ) * λ 1 Tr F ( v ) F ( v ) T M M T + v λ 2 Tr F ( v ) T G ( v ) T L ( v ) G ( v ) F ( v ) + v λ 3 E ( v ) 1 s . t . Y ( v ) = Y ( v ) Z ( v ) + E ( v ) , Z ( v ) 1 = 1 , 0 Z ( v ) 1 , Z i i ( v ) = 0 , F ( v ) T F ( v ) = I , M T M = I ,
where Y v represents the complete samples in v-th view. Z ( v ) denotes the respective v -th view’s graph. F ( v ) represents the clustering indicator matrix with proper size. Moreover, M refers to the final shared clustering indicator matrix. Although IMSC-AGL achieves considerable performance in various applications, it can still be improved from the number of hyper-parameters and considering to fuse multiple information in a weighted manner.

4. Method

4.1. Adaptive Weighted Graph Fusion Incomplete Multi-View Subspace Clustering

In this section, we present our adaptive graph fusion incomplete multi-view subspace clustering method (AWGF-IMSC) in detail and give a unified objective function.
For incomplete multi-view data, we remove the incomplete instances and reform as X ( i ) R d i × n i , where d i and n i represent the feature dimension and the numbers of visible samples of i-th view, respectively. We assume that semi-NMF factorizes the input data X ( i ) R d i × n i into base matrix U ( i ) R d i × k and coefficient matrix V ( i ) R k × n i . k is the dimension of target space and is commonly set to the number of the clusters of X ( i ) . Considering that the missing samples differ in each view, we learn latent representations of the corresponding visible samples in each view. Therefore, the semi-NMF for individual view can be formulated as:
min U ( i ) , V ( i ) i = 1 m X ( i ) U ( i ) V ( i ) F 2 s . t . V ( i ) 0 .
To further exploit the intra-view similarity structure and the underlying subspace structure, we utilize the self-representation property [47] on the k × n i dimensional latent representation V ( i ) to construct the graph. Thus, we can obtain the different graphs Z ( i ) of individual views by solving the following problem:
min V ( i ) , Z ( i ) i = 1 m V ( i ) V ( i ) Z ( i ) F 2 s . t . V ( i ) 0 , 0 Z j , k ( i ) 1 , Z ( i ) 1 = 1 ,
where the constraint 0 Z j , k ( i ) 1 and Z ( i ) 1 = 1 guarantee a good probabilistic explanation for Z ( i ) . After obtaining the graphs on each view, the natural idea is to integrate the multiple incomplete information into a complete one. In order to establish the correspondence between the incomplete and complete graphs, we denote the index matrix O ( i ) . The index matrix O ( i ) R n i × N can extract the visible instances of view i from the complete graph. To be specific, the matrix O ( i ) is defined as:
O j , k ( i ) = 1 , if x j ( i ) is the k - th sample in complete dataset . 0 , otherwise .
Through the index matrix, we can achieve the transformation between complete and incomplete graphs: Z ( i ) = O ( i ) Z * O ( i ) or O ( i ) Z ( i ) O ( i ) = Z ^ * . In the second condition, O ( i ) expands the graph Z ( i ) into Z ^ * , where Z ^ * has the same size with Z * , but the irrelevant items to view i are zero.
Owing to the size of the graph and the similarity magnitude differing among views, it is unreasonable to directly add up the multiple graphs. Consequently, we aim to integrate the multiple information into a completed graph with adaptive learning weights { α i } i = 1 m . With the help of the index matrix, relevant elements can be extracted from Z * . Then, we can adaptively fuse { Z ( i ) } i = 1 m into a complete graph with auto-learning weights, as illustrated in Equation (6).
min α i i = 1 m α i Z ( i ) O ( i ) Z * O ( i ) F 2 s . t . α i 0 , i = 1 m α i = 1 ,
where α i is the weight for i-th view. It is automatically learned and optimized to illustrate the importance of i-th view. In this manner, the complete graph is learned by a weighted combination of incomplete graphs. Besides, with the fusion of beneficial information, the inconsistencies between different views, noise and outliers in individual view are also integrated into the complete graph. Considering that, an additional sparse constraint is added on Z * . Therefore, integrating the above parts into an unified objective function, we have our optimization goal as:
min U ( i ) , V ( i ) , α i , Z ( i ) Z * i = 1 m X ( i ) U ( i ) V ( i ) F 2 + V ( i ) V ( i ) Z ( i ) F 2 + α i Z ( i ) O ( i ) Z * O ( i ) F 2 + λ 2 Z * 1 s . t . V ( i ) 0 , 0 Z j , k ( i ) 1 , Z ( i ) 1 = 1 , i = 1 m α i = 1 , α i 0 ,
λ 1 and λ 2 are non-negative trade-off parameters. In the proposed framework, we have four terms: using semi-NMF to obtain latent representation, conducting graph construction with self-representation, adaptive graph fusion and sparse regularizer. Finally, we get a full-size graph Z * incorporating all the sample information in the latent subspace.

4.2. Optimization Algorithm for AWGF-IMSC

The constraint problem in Equation (7) is not jointly convex with regard to all the variables. In this section, we propose an alternating iterative algorithm to solve this optimization problem.

4.2.1. Update U ( i )

With V ( i ) , Z ( i ) , α i and Z * fixed, for each U ( i ) , we need to solve the following problem,
min U X ( i ) U ( i ) V ( i ) F 2
Each of U ( i ) can be solved separately since views are independent from each other. Therefore, the optimization problem that we minimize can be rewritten as:
L ( U ( i ) ) = Tr ( X ( i ) X ( i ) 2 V ( i ) U ( i ) X ( i ) + V ( i ) U ( i ) U ( i ) V ( i ) )
The solution for U ( i ) can be easily obtained by setting the derivation w.r.t. U ( i ) to zero.
L ( U ( i ) ) U ( i ) = 2 X ( i ) V ( i ) + 2 U ( i ) V ( i ) V ( i )
Then, we can get the optimal closed-form solution:
U ( i ) = X ( i ) V ( i ) ( V ( i ) V ( i ) ) 1

4.2.2. Update V ( i )

Fixing U ( i ) , Z ( i ) , α i and Z * , the minimum problem for optimizing V ( i ) can be simplified as:
L ( V ( i ) ) = i = 1 m X ( i ) U ( i ) V ( i ) F 2 + V ( i ) V ( i ) Z ( i ) F 2 s . t . V ( i ) 0 ,
We can update V ( i ) in Equation (12) referring to the update strategy in semi-NMF, since the semi-NMF and subspace learning processes are isolated from each other. The partial derivation of L ( V ( i ) ) with respect to V ( i ) can be obtained as:
L ( V ( i ) ) V ( i ) = V ( i ) ( I Z ( i ) Z ( i ) + Z ( i ) Z ( i ) ) + U ( i ) U ( i ) V ( i ) U ( i ) X ( i )
According to the optimization of semi-NMF and the KKT condition, we can get
( V ( i ) P ( i ) + U ( i ) U ( i ) V ( i ) U ( i ) X ( i ) ) j , k V j , k ( i ) = 0 ,
where P ( i ) = I Z ( i ) Z ( i ) + Z ( i ) Z ( i ) . Note that U ( i ) U ( i ) = ( U ( i ) U ( i ) ) + ( U ( i ) U ( i ) ) . Based on this, we can achieve the updating rule for V ( i ) :
V j , k ( i ) V j , k ( i ) ( V ( i ) ( P ( i ) ) ) j , k + [ ( U ( i ) U ( i ) ) V ( i ) ] j , k + ( U ( i ) X ( i ) ) j , k + ( V ( i ) ( P ( i ) ) + ) j , k + [ ( U ( i ) U ( i ) ) + V ( i ) ] j , k + ( U ( i ) X ( i ) ) j , k .

4.2.3. Update Z ( i )

When U ( i ) , V ( i ) , α i and Z * fixed, the optimization for Z ( i ) can be simplified as:
min Z ( i ) i = 1 m V ( i ) V ( i ) Z ( i ) F 2 + λ 1 α i Z ( i ) O ( i ) Z * O ( i ) F 2 s . t . 0 Z j , k ( i ) 1 , Z ( i ) 1 = 1 .
Denoting H ( i ) = O ( i ) Z * O ( i ) , we can obtain the following equivalent question
min Z ( i ) i = 1 m Tr ( Z ( i ) ( V ( i ) V ( i ) + λ 1 α i 2 I ) Z ( i ) 2 Z ( i ) ( V ( i ) V ( i ) + λ 1 α i H ( i ) ) + V ( i ) V ( i ) + λ 1 H ( i ) H ( i ) ) .
Setting the derivative with respect to Z : , j ( i ) to zero, we can get
L ( Z : , j ( i ) ) Z : , j ( i ) = i = 1 m j = 1 n i Q ( i ) Z : , j ( i ) + Q ( i ) Z : , j ( i ) p ( i ) = 0 ,
where Q ( i ) = V ( i ) V ( i ) + λ 1 α i 2 I and p ( i ) = 2 ( V ( i ) V ( i ) ) : , j + 2 λ 1 α i H : , j ( i ) . For each view, we can obtain the following closed-form solution
Z : , j ( i ) = ( Q ( i ) + Q ( i ) ) 1 p ( i )

4.2.4. Update α i

By fixing U ( i ) , V ( i ) , Z ( i ) and Z * and removing other terms, the optimization for α i can be transformed into solving Equation (18).
min α i i = 1 m α i Z ( i ) O ( i ) Z * O ( i ) F 2 s . t . i = 1 m α i = 1 , α i 0
Note that H ( i ) α i Z ( i ) O ( i ) Z * O ( i ) . For each view, we can obtain the following Lagrange function:
L ( α i ) = α i 2 Tr ( Z ( i ) Z ( i ) ) 2 α i Tr ( Z ( i ) H ( i ) ) + ( Tr ( H ( i ) H ( i ) ) + γ ( i = 1 m α i 1 ) ,
where γ is the Lagrange multipliers. Setting the derivative of L ( α i ) w.r.t. α i to zero, we can obtain:
α i = 2 Tr ( Z ( i ) H ( i ) ) γ 2 Tr ( Z ( i ) Z ( i ) )
According to the constraint i = 1 m α i = 1 , we can compute γ and further get each α i .

4.2.5. Update Z *

With U ( i ) , V ( i ) , Z ( i ) and α i fixed, we need to minimize the following objective for Z ( i ) .
min Z * i = 1 m λ 1 α i Z ( i ) O ( i ) Z * O ( i ) F 2 + λ 2 Z * 1
We can get the equivalent element-wise equation:
min Z * λ 1 i = 1 r α i Z : , j ( i ) r Z j , q * 2 + λ 2 | Z k , q * |
Note that j , p differ in different views. r is the count of views in which samples j and p exist simultaneously. If O j , k ( i ) = 1 and O p , q ( i ) = 1 , the elements Z k , q * in Z * are composed of the weighted sum of the corresponding instance Z j , p ( i ) in view i. Specifically, Equation (22) obtains a unique solution:
Z k , q * = 2 λ 1 r i = 1 r α i Z j , p ( i ) λ 2 2 λ 1 r 2 , if i = 1 r α i Z j , p ( i ) > λ 2 2 λ 1 r 2 λ 1 r i = 1 r α i Z j , p ( i ) + λ 2 2 λ 1 r 2 , if i = 1 r α i Z j , p ( i ) < λ 2 2 λ 1 r 0 , otherwise
As can be seen in Equation (23), the solution of optimal Z * is a weighted combination of self-representation graph over the view which corresponding samples are visible. Moreover, the noise and outliers will be given a very small value to make the Z * sparse. Therefore, we can get a robust and complete graph revealing all of the relationships of the samples.

4.3. Convergence and Computational Complexity

We end up in this section by analyzing the convergence analysis and computational complexity of our proposed method.
Convergence analysis: We first analyze the convergence of the proposed method. Algorithm 1 is a convex problem during the updating of each variable. Each sub-problem obtains a global optimum solution and the value of the objective function is non-increasing until converges. Experiment results in the next section demonstrate this in practice.
Computational complexity analysis: With the optimization process outlined in Algorithm 1, the total time complexity consists of five parts referring to the alternate steps. For incomplete multi-view setting, dimensionality and the number of complete samples varies across different views. With notations in the following, the first stage for computing U ( i ) needs O ( d i n i k + n i k 2 + k 3 + d i k 2 ) for the i-th view. The time cost of d × n dimensional matrix X multiplying n × k dimensional matrix V is O ( d i n i k ) . Similarly, the time costs of V V and ( V V ) 1 are O ( n i k 2 ) and O ( k 3 ) , respectively. At last, the result of X V times ( V V ) 1 costs O ( d i k 2 ) . Therefore the total time cost of updating U ( i ) is O ( d i n i k + n i k 2 + k 3 + d i k 2 ) for each view. Similarly to updating U ( i ) , the time cost of updating V ( i ) is O ( q ( k n i 2 + d i k 2 + k 2 n i + k d i n i ) ) , where q is the number of iterations. The time cost of updating Z ( i ) is O ( n i ( n i k 2 + n i 3 ) ) . The time cost of updating α i is O ( m n i 3 ) . At last, solving Z * acquires O ( m n i 2 ) . After all, the time complexity of our algorithm is O ( m n i ( n i k 2 + n i 3 + n i + n i k ) ) .
Algorithm 1: AWGF-IMSC
Sensors 20 05755 i001

5. Experiment

5.1. Datasets

To demonstrate the effectiveness of our proposed algorithm AWGF_IMSC, we do comparisons with six baseline methods on four benchmark datasets. The statistical information of the datasets is displayed in Table 1. Detailed introductions are as follows:
  • BUAA-visnir face database (BUAA) [49]. The dataset BUAA used in this paper contains 1350 instances of 150 categories. Each instance has visible images (VIS) and near infrared images (NIR), which naturally form a two-view dataset. Both VIS and NIR images are 640×480 pixels. Then, they are resized into 10 × 10 matrix and vectorized into 100-dimensional features.
  • Caltech7 [50]. The Caltech7 dataset is a subset of the Caltech101 dataset, containing seven categories (Face, Motorbikes, Dolla-Bill, Garfield, Snoopy, Stop-Sign and Windsorchair) and 1474 instances. The original images of dataset Caltech7 differ in size. We follow the work in [48], selecting two of five given features as the multi-view dataset. The selected two views refer to 512 dimensional GIST features [51] and 928 dimensional local binary patterns(LBP) features [51].
  • One-hundred plant species leaves dataset (100Leaves) [52]. The 100Leaves dataset contains 1600 instances from 100 categories. The original images of 100Leaves differ in size, too. Shape descriptor, fine scale margin and texture histogram features constitute three-views to depict samples from different perspectives.
  • Mfeat handwritten digit dataset (Mfeat) [53]. This dataset contains 2000 samples. The size of the original images of dataset Mfeat is 891 × 702 pixels. The public multi-view dataset of it has six views. In our experiments, we select 76-dimensional features of Fourier coefficients of the character shapes and 240-dimensional features of pixel averages.
As shown in Figure 2, we randomly select six pictures of different categories from four original datasets for display.

5.2. Baselines

We conduct extensive experiments, comparing with several state-of-the-art incomplete multi-view clustering methods. Brief introductions are given below.
  • Best single view (BSV). BSV first fills the missing samples with the average feature values of its view. The affinity matrices can be constructed by Gaussian kernel. Then, we perform spectral clustering algorithm on the similarity matrix of each view and report the best clustering performance.
  • Partial multi-view clustering (PVC) [35]. This method supposes that the instances available in both views should have a common representation. The view-specific instances which are missing in another view should maintain the specific information. Based on the NMF, this method integrates the common and view-specific representations in the latent space to form a unified representation.
  • Multiple incomplete view clustering via weighted non-negative matrix factorization with 2 , 1 regularization (MIC) [38]. This paper first fills the missing instances with an average value of features and then learns a 2 , 1 regularized latent subspace by weighted NMF.
  • Incomplete multi-modal visual data grouping (IMG) [36]. IMG proposes to use the latent representation to generate a complete graph, which establishes a connection between missing data from different views.
  • Doubly aligned incomplete multi-view clustering (DAIMC) [37]. The proposed method first aligns the samples into a common representation by semi-NMF and then aligns the base matrices with the help of 2 , 1 regularized regression modal.
  • Incomplete multi-view spectral clustering with adaptive graph learning (INMF-AGL) [48] induces a co-regularization term to learn the common representation, which integrates the graph learning and spectral clustering.
For the compared methods, we run their demo with the suggest or default parameters and repeat five times to obtain average results. Note that the PVC and IMG methods can only deal with the two-view scenarios. In our experiments, we combine two different views and report the best result.
Following most existing works, we utilize accuracy (ACC) and normalized mutual information (NMI) to measure the clustering results; higher values representing better clustering performance. Then we give the definition of the two metrics. Denoting true positive (TP), false positive (FP), false negative (FN) and true negative (TN), we can obtain the ACC and NMI definitions as follows:
A C C = T P + T N T P + T N + F P + F N ,
indicating the percentage of correct predicted results in the total samples.
NMI quantifies the amount of information contained in a random variable about another random variable. For clarity, the expression can be formulated as:
N M I ( X , Y ) = 2 I ( X , Y ) H ( X ) + H ( Y ) ,
where mutual information I ( X , Y ) is x y p ( x , y ) log p ( x , y ) p ( x ) p ( y ) , p ( x , y ) is the joint probability distribution of X and Y, p ( x ) is the marginal probability distribution of X. H ( X ) = i p ( x i ) log p ( x i ) is the information entropy, regarded as the uncertainty of random variables.

5.3. Experiment Setting

In our experiments, we generate incomplete data from complete multi-view datasets in the way of One-complete, which means that we randomly select one of the views to be complete. The rest of the views suffer different incomplete ratio (IR) from 10%, 20%, 30%, 40%, 50%, 60%, 70%. For the two-view datasets BUAA, Caltech7 and Mfeat, we randomly select one view as the complete view. The incomplete case occurs in the rest view with randomly removing 10–70% samples. For more than two-view occasions, one view is chosen randomly and the rest views suffer 10–70% missing. The datasets used in this paper can be found in our Github (https://github.com/Jeaninezpp/Incomplete-multi-view-datasets). In our experiments, we perform our proposed method five times as the compared method for fairness. Our code is available at https://github.com/Jeaninezpp/AWGF-code.

5.4. Experiment Results and Analysis

Experiment results on different datasets of various compared method are enumerated in Table 2, Table 3, Table 4 and Table 5. These four tables present the ACC results of the above algorithms on four benchmark datasets. Each row shows the accuracy of compared methods under a certain incomplete ratio. We highlight the best results in bold. Each column represents the evolution of the accuracy of the corresponding method as the incomplete ratio increase. Under each incomplete ratio, we can get the sequence number by sorting the accuracy from high to low. The sequence number is regarded as the rank of the algorithms under a certain incomplete ratio. Then, we can obtain the average rank by taking the average of the ranks over each algorithm. The average rank illustrates the robustness of the method in terms of incomplete ratio.
Furthermore, we depict the NMI metric in Figure 3 with line charts. Based on these results, we have the following observations:
  • Compared with the proposed method, the BSV method yields worse clustering performance. This is mainly because directly filling the missing instances with the average features will lead them to be clustered into the same group. The weighted NMF methods DAIMC and MIC perform better than BSV at a low missing rate since the NMF-based methods learn a shared representation to exploit the complementary information across views. Besides, the weighted manner reduces the negative impact of the missing instances. However, with the increasing incomplete ratio, these two methods suffer a sharp decline, especially apparent in Mfeat dataset. Methods like IMG and INMF_AGL involving the graph construction perform better than them. Our proposed method integrates the advantages of NMF-based and graph-based methods, adaptively fusing the graph learned from each embedding space. Therefore, our AWGF_IMSC method reaches the best clustering performance in most cases.
  • Comparing with the INMF_AGL [48], our proposed method consistently further improves the clustering performance and achieves better results among the benchmark datasets. In addition to the cases on Mfeat, the INMF_AGL method performs higher accuracy than AWGF_IMSC under the 20–50% missing rate. However, our performance exceeds it under the incomplete ratio in 60% and 70%. Although both of INMF_AGL and our AWGF_IMSC adopt subspace clustering to build graph structure in each view, the clustering results demonstrate the effectiveness of graph fusion instead of indicator fusion in INMF_AGL.
  • AWGF-IMSC shows clear advantages over other compared baselines under various incomplete ratios, with three best and one second-best results out of the total four datasets. For example, on the BUAA dataset (Table 3), our method transcends the second best method by 7.94%, 7.58%, 7.18%, 7.53%, 14.93%, 9.24% and 7%, respectively. More significant improvements can be seen on dataset Caltech7. In Table 2, the ACCs of the proposed method are 23.42%, 13.10%, 15.02%, 9.74%, 10.48%, 10.8% and 11.43% higher than the second best INMF_AGL method. These significant results verify the effectiveness of the proposed adaptive weighted graph-based fusion learning for incomplete multi-view clustering. Our method achieves the best average rank in Caltech7, BUAA and 100Leaves. In Mfeat, the average rank of the proposed method is second best, which is only 0.43 more than the best, but 0.85 less than the third.
  • As shown in Figure 3, we can also observe that the proposed algorithm outperforms other methods on all of the datasets under various incomplete ratio. Besides, our method appears a relatively stable trend as the missing rate increases. Moreover, the abnormal phenomenon of BSV in dataset Caltech7 (Figure 3a) maybe because the preserved complete view has an excellent structure when generating large missing datasets. Therefore, the results of the BSV will be outstanding. Other methods are affected by the negative impact of missing samples and thus produce a lower effect than BSV, while our method is still superior to all compared methods by a more significant proportion, further illustrating the effectiveness and superiority of the proposed method.

5.5. Analysis of the Parameter Sensitivity

In this section, we analyze the impact of the hyper-meters λ 1 and λ 2 in AWGF-IMSC on clustering performance. The parameters are chosen ranging from 10 3 , 10 2 , , 10 3 by grid search. Figure 4 and Figure 5 plot the NMI results by varying λ 1 and λ 2 in a large range on BUAA, Caltech7, 100Leaves and Mfeat.
We have the following observations from Figure 4 and Figure 5: (i) All the two hyper-parameters are effective in improving the clustering performance; (ii) AWGF-IMSC is practically stable against these parameters that it achieves competitive performance in a wide range of parameter settings in a low missing rate; (iii) With the incomplete ratio increasing, the combinations of a relatively smaller λ 1 and a more prominent λ 2 tend to achieve better performance. The reason is that the λ 2 controls the impact of the sparse regularizer of the complete graph. Imposing sparseness requirements on the graph within a specific range will filter out the noise in the graph and the inconsistency between the views. (iv) For different datasets, we can conclude that most of the datasets are relatively stable when λ 1 { 1 e 3 , 1 e 2 , 1 e 1 } , λ 2 { 10 , 100 , 1000 } .

5.6. Convergence Analysis

Our algorithm is theoretically guaranteed to converge to a local minimum, as illustrated in the optimization part. For the convergence analysis, we conduct experiments on the four datasets with all incomplete ratios and all suggest parameter scope. We randomly select from each dataset and draw the evolution of the objective value, as shown in Figure 6. In the above experiments, we observe that our algorithm’s objective values monotonically decrease at each iteration and usually converge in less than 20 iterations. These results verify our proposed algorithm’s convergence.

6. Conclusions

This article proposes a novel incomplete multi-view clustering method to fuse the local-structure contained graph with adaptive view-importance learning. We incorporate representation learning and incomplete graph fusion into a unified framework, whereas two processes can negotiate with each other, serving for graph learning tasks. A sparse regularization is imposed on the complete graph to do it more robustly to the view-inconsistency. Moreover, the importance of different views is automatically learned, further guiding the construction of the complete graph. We conduct experiments to illustrate the effectiveness and superiority of the proposed method. Some recent works utilize deep neural networks, such as GAN (generative adversarial network), to generate missing features to solve incomplete multi-view clustering problems. The utilization of neural networks to handle multi-view clustering and incomplete multi-view clustering will be advanced considerations in the future.

Author Contributions

All authors contributed to the manuscript preparation of the paper. Conceptualization, P.Z. and S.W.; methodology, P.Z.; validation, P.Z. and J.H.; formal analysis, P.Z. and Z.C. (Zhen Cheng); investigation, P.Z. and X.G.; data curation, P.Z.; writing–original draft preparation, P.Z.; writing–review and editing, S.W., J.H., Z.C. (Zhen Cheng) and Z.C. (Zhiping Cai); visualization, P.Z. and X.G.; supervision, E.Z., Z.C. (Zhiping Cai) and S.W.; funding acquisition, E.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Key R&D Program of China 2018YFB1003203 and the National Natural Science Foundation of China (Grant NO.61672528, NO. 61773392, NO. 62072465).

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AWGF-IMSCAdaptive weighted graph fusion incomplete multi-view subspace clustering
MVCMulti-view clustering
NMFNon-negative matrix factorization
BUAABUAA-visnir face database
100LeavesOne-hundred plant species leaves dataset
MfeatMfeat handwritten digit dataset
VISVisual image
NIRNear infrared image
LBPLocal binary patterns
IRIncomplete ratio
ACCAccuracy
NMINormalized mutual information
TP, FP, FN, TNTrue positive, false positive, false negative, true negative
BSVBest single view
PVCPartial multi-view clustering
MICMultiple incomplete view clustering via weighted non-negative matrix factorization with 2 , 1 regularization
IMGIncomplete multi-modal visual data grouping
DAIMCDoubly aligned incomplete multi-view clustering
INMF-AGLIncomplete multiview spectral clustering with adaptive graph learning

References

  1. Zhao, Q.; Zhang, Y.; Qin, Q.; Luo, B. Quantized Residual Preference Based Linkage Clustering for Model Selection and Inlier Segmentation in Geometric Multi-Model Fitting. Sensors 2020, 20, 3806. [Google Scholar] [CrossRef]
  2. Biabani, M.; Fotouhi, H.; Yazdani, N. An Energy-Efficient Evolutionary Clustering Technique for Disaster Management in IoT Networks. Sensors 2020, 20, 2647. [Google Scholar] [CrossRef]
  3. Deng, T.; Ye, D.; Ma, R.; Fujita, H.; Xiong, L. Low-rank local tangent space embedding for subspace clustering. Inf. Sci. 2020, 508, 1–21. [Google Scholar] [CrossRef]
  4. Peng, X.; Feng, J.; Zhou, J.T.; Lei, Y.; Yan, S. Deep subspace clustering. IEEE Trans. Neural Netw. Learn. Syst. 2020. [Google Scholar] [CrossRef]
  5. Chen, J.; Zhao, Z.; Ye, J.; Liu, H. Nonlinear Adaptive Distance Metric Learning for Clustering. In Proceedings of the 13th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD’07), San Jose, CA, USA, 12–15 August 2007; Association for Computing Machinery: New York, NY, USA, 2007; pp. 123–132. [Google Scholar] [CrossRef]
  6. Gönen, M.; Alpaydin, E. Localized Multiple Kernel Learning. In Proceedings of the 25th International Conference on Machine Learning (ICML’08), Helsinki, Finland, 5–9 July 2008; Association for Computing Machinery: New York, NY, USA, 2008; pp. 352–359. [Google Scholar] [CrossRef] [Green Version]
  7. Chaudhuri, K.; Kakade, S.M.; Livescu, K.; Sridharan, K. Multi-View Clustering via Canonical Correlation Analysis. In Proceedings of the 26th Annual International Conference on Machine Learning (ICML’09), Montreal, QC, Canada, 14–18 June 2009; Association for Computing Machinery: New York, NY, USA, 2009; pp. 129–136. [Google Scholar] [CrossRef] [Green Version]
  8. Yu, S.; Tranchevent, L.; Liu, X.; Glanzel, W.; Suykens, J.A.; De Moor, B.; Moreau, Y. Optimized data fusion for kernel k-means clustering. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 34, 1031–1039. [Google Scholar] [CrossRef]
  9. Huang, H.C.; Chuang, Y.Y.; Chen, C.S. Multiple kernel fuzzy clustering. IEEE Trans. Fuzzy Syst. 2012, 20, 120–134. [Google Scholar] [CrossRef] [Green Version]
  10. Liu, X.; Wang, L.; Yin, J.; Zhu, E.; Zhang, J. An efficient approach to integrating radius information into multiple kernel learning. IEEE Trans. Cybern. 2013, 43, 557–569. [Google Scholar]
  11. Gönen, M.; Margolin, A.A. Localized Data Fusion for Kernel k-Means Clustering with Application to Cancer Biology. In Advances in Neural Information Processing Systems 27; Curran Associates, Inc.: New York, NY, USA, 2014; pp. 1305–1313. [Google Scholar]
  12. Liu, X.; Dou, Y.; Yin, J.; Wang, L.; Zhu, E. Multiple Kernel k-Means Clustering with Matrix-Induced Regularization. In Proceedings of the Thirtieth AAAI Conference on Artificial Intelligence, Phoenix, AZ, USA, 12–17 February 2016; pp. 1888–1894. [Google Scholar]
  13. Wang, S.; Zhu, E.; Hu, J.; Li, M.; Zhao, K.; Hu, N.; Liu, X. Efficient multiple kernel k-means clustering with late fusion. IEEE Access 2019, 7, 61109–61120. [Google Scholar] [CrossRef]
  14. Chowdhary, C.L.; Mittal, M.; Pattanaik, P.; Marszalek, Z. An Efficient Segmentation and Classification System in Medical Images Using Intuitionist Possibilistic Fuzzy C-Mean Clustering and Fuzzy SVM Algorithm. Sensors 2020, 20, 3903. [Google Scholar] [CrossRef]
  15. Du, L.; Zhou, P.; Shi, L.; Wang, H.; Fan, M.; Wang, W.; Shen, Y.D. Robust Multiple Kernel K-Means Using 2,1 Norm. In Proceedings of the 24th International Conference on Artificial Intelligence, Buenos Aires, Argentina, 25–31 July 2015; pp. 3476–3482. [Google Scholar]
  16. Zhou, S.; Liu, X.; Li, M.; Zhu, E.; Liu, L.; Zhang, C.; Yin, J. Multiple kernel clustering with neighbor-kernel subspace segmentation. IEEE Trans. Neural Netw. Learn. Syst. 2019, 31, 1351–1362. [Google Scholar] [CrossRef] [Green Version]
  17. Wang, S.; Liu, X.; Zhu, E.; Tang, C.; Liu, J.; Hu, J.; Xia, J.; Yin, J. Multi-view Clustering via Late Fusion Alignment Maximization. In Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence, Macao, China, 10–16 August 2019; pp. 3778–3784. [Google Scholar] [CrossRef] [Green Version]
  18. Kumar, A.; Daumé, H., III. A Co-Training Approach for Multi-View Spectral Clustering. In Proceedings of the 28th International Conference on Machine Learning (ICML’11), Bellevue, WA, USA, 28 June–2 July 2011; Omnipress: Madison, WI, USA, 2011; pp. 393–400. [Google Scholar]
  19. Kumar, A.; Rai, P.; Daumé, H. Co-Regularized Multi-View Spectral Clustering. In Proceedings of the 24th International Conference on Neural Information Processing Systems (NIPS’11), Guangzhou, China, 14–18 November 2017; Curran Associates Inc.: Red Hook, NY, USA, 2011; pp. 1413–1421. [Google Scholar]
  20. Zhang, Z.; Liu, L.; Shen, F.; Shen, H.T.; Shao, L. Binary multi-view clustering. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 41, 1774–1782. [Google Scholar] [CrossRef]
  21. Jiang, G.; Wang, H.; Peng, J.; Chen, D.; Fu, X. Graph-based Multi-view Binary Learning for Image Clustering. arXiv 2019, arXiv:1912.05159. [Google Scholar]
  22. Zhang, C.; Fu, H.; Hu, Q.; Cao, X.; Xie, Y.; Tao, D.; Xu, D. Generalized latent multi-view subspace clustering. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 42, 86–99. [Google Scholar] [CrossRef]
  23. Zhang, C.; Hu, Q.; Fu, H.; Zhu, P.; Cao, X. Latent Multi-view Subspace Clustering. In 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR); IEEE Computer Society: Los Alamitos, CA, USA, 2017; pp. 4333–4341. [Google Scholar] [CrossRef]
  24. Brbić, M.; Kopriva, I. Multi-view low-rank sparse subspace clustering. Pattern Recognit. 2018, 73, 247–258. [Google Scholar] [CrossRef] [Green Version]
  25. Kang, Z.; Zhao, X.; Peng, C.; Zhu, H.; Zhou, J.T.; Peng, X.; Chen, W.; Xu, Z. Partition level multiview subspace clustering. Neural Netw. 2020, 122, 279–288. [Google Scholar] [CrossRef]
  26. Liu, G.; Lin, Z.; Yan, S.; Sun, J.; Yu, Y.; Ma, Y. Robust Recovery of Subspace Structures by Low-Rank Representation. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 171–184. [Google Scholar] [CrossRef] [Green Version]
  27. Lee, D.D.; Seung, H.S. Learning the parts of objects by non-negative matrix factorization. Nature 1999, 401, 788–791. [Google Scholar] [CrossRef]
  28. Liu, J.; Wang, C.; Gao, J.; Han, J. Multi-View Clustering via Joint Nonnegative Matrix Factorization. In Proceedings of the 2013 SIAM International Conference on Data Mining, Austin, TX, USA, 2–4 May 2013; pp. 252–260. [Google Scholar] [CrossRef] [Green Version]
  29. Zong, L.; Zhang, X.; Zhao, L.; Yu, H.; Zhao, Q. Multi-view clustering via multi-manifold regularized non-negative matrix factorization. Neural Netw. 2017, 88, 74–89. [Google Scholar] [CrossRef] [Green Version]
  30. Yang, Z.; Liang, N.; Yan, W.; Li, Z.; Xie, S. Uniform Distribution Non-Negative Matrix Factorization for Multiview Clustering. IEEE Trans. Cybern. 2020. [Google Scholar] [CrossRef]
  31. Yin, M.; Huang, W.; Gao, J. Shared Generative Latent Representation Learning for Multi-View Clustering. In Proceedings of the Thirty-Fourth AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; pp. 6688–6695. [Google Scholar]
  32. Zhao, L.; Chen, Z.; Yang, Y.; Wang, Z.J.; Leung, V.C. Incomplete multi-view clustering via deep semantic mapping. Neurocomputing 2018, 275, 1053–1062. [Google Scholar] [CrossRef]
  33. Yu, X.; Li, H.; Zhang, Z.; Gan, C. The Optimally Designed Variational Autoencoder Networks for Clustering and Recovery of Incomplete Multimedia Data. Sensors 2019, 19, 809. [Google Scholar] [CrossRef] [Green Version]
  34. Liu, X.; Li, M.; Tang, C.; Xia, J.; Xiong, J.; Liu, L.; Kloft, M.; Zhu, E. Efficient and effective regularized incomplete multi-view clustering. IEEE Trans. Pattern Anal. Mach. Intell. 2020. [Google Scholar] [CrossRef]
  35. Li, S.Y.; Jiang, Y.; Zhou, Z.H. Partial Multi-View Clustering. In Proceedings of the Twenty-Eighth AAAI Conference on Artificial Intelligence, Québec City, QC, Canada, 27–31 July 2014; pp. 1968–1974. [Google Scholar]
  36. Zhao, H.; Liu, H.; Fu, Y. Incomplete Multi-Modal Visual Data Grouping. In Proceedings of the Twenty-Fifth International Joint Conference on Artificial Intelligence, New York, NY, USA, 9–15 July 2016; pp. 2392–2398. [Google Scholar]
  37. Hu, M.; Chen, S. Doubly Aligned Incomplete Multi-View Clustering. In Proceedings of the 27th International Joint Conference on Artificial Intelligence, Stockholm, Sweden, 13–19 July 2018; pp. 2262–2268. [Google Scholar]
  38. Shao, W.; He, L.; Yu, P.S. Multiple Incomplete Views Clustering via Weighted Nonnegative Matrix Factorization with 2,1 Regularization. In Proceedings of the 2015th European Conference on Machine Learning and Knowledge Discovery in Databases, Porto, Portugal, 7–11 September 2015; Springer: Berlin/Heidelberg, Germany, 2015; pp. 318–334. [Google Scholar] [CrossRef]
  39. Gao, H.; Peng, Y.; Jian, S. Incomplete multi-view clustering. In International Conference on Intelligent Information Processing; Springer: Berlin/Heidelberg, Germany, 2016; pp. 245–255. [Google Scholar] [CrossRef] [Green Version]
  40. Wang, J.; Lai, S.; Li, M. Improved image fusion method based on NSCT and accelerated NMF. Sensors 2012, 12, 5872–5887. [Google Scholar] [CrossRef]
  41. Zhou, Q.; Feng, Z.; Benetos, E. Adaptive noise reduction for sound event detection using subband-weighted NMF. Sensors 2019, 19, 3206. [Google Scholar] [CrossRef] [Green Version]
  42. Ding, C.H.Q.; Li, T.; Jordan, M.I. Convex and Semi-Nonnegative Matrix Factorizations. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 32, 45–55. [Google Scholar] [CrossRef] [Green Version]
  43. He, M.; Yang, Y.; Wang, H. Learning latent features for multi-view clustering based on NMF. In International Joint Conference on Rough Sets; Springer: Berlin/Heidelberg, Germany, 2016; pp. 459–469. [Google Scholar]
  44. Hu, M.; Chen, S. One-pass incomplete multi-view clustering. In Proceedings of the AAAI Conference on Artificial Intelligence 2019, Honolulu, HI, USA, 27–28 January 2019; Volume 33, pp. 3838–3845. [Google Scholar]
  45. Parsons, L.; Haque, E.; Liu, H. Subspace clustering for high dimensional data: A review. Acm Sigkdd Explor. Newsl. 2004, 6, 90–105. [Google Scholar] [CrossRef]
  46. Hoppenstedt, B.; Reichert, M.; Kammerer, K.; Probst, T.; Schlee, W.; Spiliopoulou, M.; Pryss, R. Dimensionality Reduction and Subspace Clustering in Mixed Reality for Condition Monitoring of High-Dimensional Production Data. Sensors 2019, 19, 3903. [Google Scholar] [CrossRef] [Green Version]
  47. Elhamifar, E.; Vidal, R. Sparse subspace clustering: Algorithm, theory, and applications. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 2765–2781. [Google Scholar] [CrossRef] [Green Version]
  48. Wen, J.; Xu, Y.; Liu, H. Incomplete multiview spectral clustering with adaptive graph learning. IEEE Trans. Cybern. 2018, 1418–1429. [Google Scholar] [CrossRef]
  49. Huang, D.; Sun, J.; Wang, Y. The Buaa-Visnir Face Database Instructions; Technical Report IRIP-TR-12-FR-001; School Computer Science and Engineering, Beihang University: Beijing, China, 2012. [Google Scholar]
  50. Fei-Fei, L.; Fergus, R.; Perona, P. Learning generative visual models from few training examples: An incremental Bayesian approach tested on 101 object categories. In Proceedings of the 2004 Conference on Computer Vision and Pattern Recognition Workshop, Washington, DC, USA, 27 June–2 July 2004. [Google Scholar]
  51. Oliva, A.; Torralba, A. Modeling the shape of the scene: A holistic representation of the spatial envelope. Int. J. Comput. Vis. 2001, 42, 145–175. [Google Scholar] [CrossRef]
  52. Mallah, C.; Cope, J.; Orwell, J. Plant leaf classification using probabilistic integration of shape, texture and margin features. Signal Process. Pattern Recognit. Appl. 2013, 5, 45–54. [Google Scholar]
  53. van Breukelen, M.; Duin, R.P.; Tax, D.M.; Den Hartog, J. Handwritten digit recognition by combined classifiers. Kybernetika 1998, 34, 381–386. [Google Scholar]
Figure 1. Framework of the proposed adaptive weighted graph fusion incomplete multi-view subspace clustering (AWGF-IMSC). It is a novel incomplete multi-view clustering method to fuse the local-structure contained graph with adaptive view-importance learning. Incomplete graphs of different scales are fused into a complete graph with automatically learning weights. In addition, the constructed complete graph will further guide the learning process of incomplete graphs and latent representations.
Figure 1. Framework of the proposed adaptive weighted graph fusion incomplete multi-view subspace clustering (AWGF-IMSC). It is a novel incomplete multi-view clustering method to fuse the local-structure contained graph with adaptive view-importance learning. Incomplete graphs of different scales are fused into a complete graph with automatically learning weights. In addition, the constructed complete graph will further guide the learning process of incomplete graphs and latent representations.
Sensors 20 05755 g001
Figure 2. Sample images of the four datasets. Row from top to bottom represents images from BUAA, Caltech7, 100Leaves and Mfeat, respectively.
Figure 2. Sample images of the four datasets. Row from top to bottom represents images from BUAA, Caltech7, 100Leaves and Mfeat, respectively.
Sensors 20 05755 g002
Figure 3. NMI results with different incomplete ratios on four incomplete datasets.
Figure 3. NMI results with different incomplete ratios on four incomplete datasets.
Sensors 20 05755 g003
Figure 4. Parameter study on four incomplete datasets. Pictures depict the diversification of NMI under different parameter combinations in the condition of missing 10% instances.
Figure 4. Parameter study on four incomplete datasets. Pictures depict the diversification of NMI under different parameter combinations in the condition of missing 10% instances.
Sensors 20 05755 g004
Figure 5. The influence of parameters λ 1 and λ 2 on NMI when 30% instances are missing on four datasets.
Figure 5. The influence of parameters λ 1 and λ 2 on NMI when 30% instances are missing on four datasets.
Sensors 20 05755 g005
Figure 6. The convergence curves of the objective values on four datasets.
Figure 6. The convergence curves of the objective values on four datasets.
Sensors 20 05755 g006
Table 1. Statistics of the datasets.
Table 1. Statistics of the datasets.
SamplesViewsClustersFeature
BUAA13502150100100
Caltech7147427512928
100Leaves16003100646464
Mfeat200021076240
Table 2. ACCs and average ranks on Caltech7 under different incomplete ratios. Bold numbers denote the best result.
Table 2. ACCs and average ranks on Caltech7 under different incomplete ratios. Bold numbers denote the best result.
IR\MethodBSVMICIMGDAIMCINMF_AGLOurs
10%0.33280.40070.51890.41050.52630.7605
20%0.31450.38860.50270.39690.57940.7104
30%0.34360.32050.48370.41220.57910.7293
40%0.41390.33660.50800.34230.60070.6981
50%0.48610.34930.42900.35500.57020.6750
60%0.52450.34460.49430.41550.58740.6954
70%0.43240.35390.48370.41090.57990.6942
Average Rank4.435.713.294.572.001.00
Table 3. ACCs and average ranks on BUAA under different incomplete ratios. Bold numbers denote the best result.
Table 3. ACCs and average ranks on BUAA under different incomplete ratios. Bold numbers denote the best result.
IR\MethodBSVMICIMGDAIMCINMF_AGLOurs
10%0.29640.01930.34240.32030.56520.6446
20%0.29970.01930.34240.27750.56370.6395
30%0.30060.01930.34240.23410.55660.6284
40%0.29970.01930.34240.23360.55050.6258
50%0.29650.01930.34240.23730.47300.6223
60%0.29750.01930.34240.24130.52930.6217
70%0.29790.01930.34240.25960.54100.6110
Average Rank4.146.003.004.862.001.00
Table 4. ACCs and average ranks on 100Leaves under different incomplete ratios. Bold numbers denote the best result.
Table 4. ACCs and average ranks on 100Leaves under different incomplete ratios. Bold numbers denote the best result.
IR\MethodBSVMICIMGDAIMCINMF_AGLOurs
10%0.10690.62080.56610.66280.82230.8433
20%0.10640.57680.51430.57500.78350.8104
30%0.10600.52340.46070.47400.75290.7776
40%0.10560.47440.43020.43700.70260.7291
50%0.10750.46650.40010.30680.65510.7028
60%0.10650.44490.36680.34050.63160.6478
70%0.10570.42610.37400.38750.58930.6194
Average Rank6.003.144.714.142.001.00
Table 5. ACCs and average ranks on Mfeat under different incomplete ratios. Bold numbers denote the best result.
Table 5. ACCs and average ranks on Mfeat under different incomplete ratios. Bold numbers denote the best result.
IR\MethodBSVMICIMGDAIMCINMF_AGLOurs
10%0.15200.64770.54420.86700.86500.8100
20%0.14840.57210.50040.71510.84150.7866
30%0.14710.52200.51330.57370.81770.7995
40%0.14970.45390.45540.50420.81480.7864
50%0.14580.36400.40370.54230.79890.7680
60%0.15240.35830.34240.58190.72140.7515
70%0.14760.34470.34240.64460.70270.7256
Average Rank6.004.294.712.711.431.86

Share and Cite

MDPI and ACS Style

Zhang, P.; Wang, S.; Hu, J.; Cheng, Z.; Guo, X.; Zhu, E.; Cai, Z. Adaptive Weighted Graph Fusion Incomplete Multi-View Subspace Clustering. Sensors 2020, 20, 5755. https://doi.org/10.3390/s20205755

AMA Style

Zhang P, Wang S, Hu J, Cheng Z, Guo X, Zhu E, Cai Z. Adaptive Weighted Graph Fusion Incomplete Multi-View Subspace Clustering. Sensors. 2020; 20(20):5755. https://doi.org/10.3390/s20205755

Chicago/Turabian Style

Zhang, Pei, Siwei Wang, Jingtao Hu, Zhen Cheng, Xifeng Guo, En Zhu, and Zhiping Cai. 2020. "Adaptive Weighted Graph Fusion Incomplete Multi-View Subspace Clustering" Sensors 20, no. 20: 5755. https://doi.org/10.3390/s20205755

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop