Abstract
This paper deals with clustering for multiview data. Multiview clustering has been a research hot spot in many domains or applications, such as information retrieval, biology, chemistry, and marketing. Exploring information from multiple views, one can hope to find a clustering that is more accurate than the ones obtained using the individual views. The aim is to search for clustering patterns that perform a consensus between the patterns from different views. Inspired by variable weighting and co-regularized strategy, this paper studies co-regularized weighting multiview clustering algorithms. Two co-regularized weighting multiview clustering algorithms are proposed from two aspects: pairwise co-regularization and centroid-based co-regularization. Experimental results obtained both on synthetic and real datasets show that the proposed algorithms outperform the main existing multiview clustering algorithms.
Introduction
The ever growing complexity of data constitutes a central challenge for the data mining community. This complexity may concern various aspects, such as the size of the dataset, the complexity of the features, the temporality or scalability, and more generally the multiplicity of data.
Multiview data are instances that have multiple views from different feature spaces. Data with multiple representations are quite common in scientific, economic, and social domains such as biology, medicine, marketing, and social networks. For example, the variables of the nucleated blood cell data 1 were divided into views of density, geometry, “color,” and texture, each representing a view of particular measurements on the nucleated blood cells. In the past decade, multiview data have raised interests in the so-called multiview clustering.2–7 Multiview clustering needs to exploit the information from multiple views and takes the differences among different views into consideration in order to produce a more accurate and robust partitioning of the data.
A fundamental issue of multiview clustering is to combine multiple clustering from each individual views. Usually, there exist two main approaches in multiview learning: centralized and distributed. Distributed strategy is also known as clustering ensembles problematic or a posteriori fusion. Centralized algorithms make use of multiple representations simultaneously to discover hidden patterns from the data. Most of the existing work in multiview clustering follows the centralized approach with extensions to existing clustering algorithms.3–5
In semisupervised learning, co-regularization makes the hypotheses that learned from different views of the data agree with each other on unlabeled data. 8 The framework employs two main assumptions: (1) the true objective functions in each view should agree on the labels for the unlabeled data (compatibility), and (2) the views are independent given the class label (conditional independence). The compatibility assumption allows us to shrink the space of possible target hypotheses by searching only over the compatible functions. The independence assumption makes it unlikely for compatible classifiers to agree on wrong labels. In the case of clustering, this would mean that a data point in both views would be assigned to the correct cluster with high probability.
In addition, currently, most of multiview clustering algorithms consider the every view equally in terms of importance, but if there is a poor quality view, it will seriously affect the clustering results. Such as the data distribution is too concentrated to separate completely or there exists noise or outliers in a view, these will make clustering result in this view worse than other views. It is essential to assign weights to different views. Variable weighting clustering was introduced in clustering analysis.8–10 Recently, Tzortzis and Likas 5 proposed a weighted combination of exemplar-based mixture models that assigns different weights to the views and learns those weights automatically. Chen et al. 7 proposed TW-k-means algorithm, considering both view weights and variable weights.
In this paper, we propose two fuzzy clustering algorithms for multiview datasets by co-regularizing the clustering hypotheses across views. Co-regularization is a well-known technique in semisupervised literature; however, it is not known on using it for unsupervised learning problems.9,10 Inspired by the co-regularization and variable weighting strategy, we propose two novel fuzzy multiview clustering objective functions that implicitly combine multiple views of the data to achieve a better clustering performance. The weights of views are introduced to distinguish the impacts of different views in clustering. A co-regularized term is also added to its objective function to mine the hidden patterns from different views.
Related works on multiview clustering
In this section, we introduce some necessary notations and briefly introduce some multiview clustering approaches presented in the past. In the following, we consider a dataset of R views to be clustered where
Collaborative fuzzy clustering
Pedrycz
11
uses the FCM model and derives the collaborative variant CoFC for the multiview context. The collaboration between the views only concerns the membership degree
Centralized method for multiple-view clustering (CoFKM)
Cleuziou and Exbrayat
9
presented a CoFKM. This approach is an extension of the FCM method based on a centralized strategy. This method introduces a penalty term which aims at reducing the disagreement between different views. The objective function of CoFKM is defined as follows
We can see that CoFKM overcomes the shortcomings of CoFC, but it considers the every view equally in terms of importance. However, if there is a poor quality view which is very common in real-world datasets, it may affect the clustering performance.
Co-regularized weighting multiview clustering (CoWMVC)
We assume that we are given data having multiple representations (i.e. views). Given a dataset of R views, let
With
Here, we propose two co-regularization-based approaches to make the clustering hypotheses on different views agree with each other. We construct an objective function by introducing a penalty term which aims at reducing the disagreement between different views.
Our first co-regularization scheme enforces that a view pair (v, w) should have high pairwise similarity (pair-wise co-regularization criteria). Our second co-regularization scheme enforces different views to look similar by regularizing them toward a common consensus (centroid-based co-regularization). The idea is different from previously proposed consensus clustering approaches that commit to individual clusterings in the first step and then combine them to a consensus in the second step.
Pairwise CoWMVC
We propose a co-regularized approach based on FCM, which aims at minimizing in each view and penalizing the disagreement between all pairs of views. Thus, the criterion to be minimized can be written as
This co-regularized term is made to penalize our criterion. It can be considered as a divergence between views, the lower Δ is, the lower the disagreement. The advantage of equation (5) is that the pairwise co-regularized term has the same order of magnitude with local inertia, then the sum of these expressions can be considered as a coherent global criterion
Given that V and W are fixed,
Given that U and W are fixed,
Given that U and V are fixed,
Since the proposed method produces a fuzzy partition for each view, in order to get a global clustering, we have to merge these results. We propose to build a global partition of data using an assignment rule
And assign the data xi to the cluster k according the maximum of
Therefore, the pairwise co-regularized weighting multiview clustering algorithm CoWMVC(P) is summarized as follows:
Set the cluster number C; initialize the membership matrices, cluster centers, and weights for all views; and set the termination criterion. For each view, update the cluster center using equation (9). For each view, update the membership using equation (8). Compute the weight for rth view using equation (10) Repeat step (2) to step (5), until the termination criterion is satisfied. Merge these results to get a global clustering using equation (11)
Centroid-based CoWMVC
In this section, we present an alternative regularization scheme that regularizes each view-specific set of membership
We can rewrite the objective function as follows
Minimizing the objective function, we can obtain the updating equations of the cluster center, the membership, and the weights for views, respectively.
Given that V and W are fixed,
The updating equation of consensus membership is obtained as follows
Given that U and W are fixed,
Given that U and V are fixed,
In centroid-based CoWMVC algorithm, as we have got the consensus membership
Experiments
Datasets
We conduct experiments and validate our approach on two synthetic datasets and two real-world datasets from UCI Machine Learning repository. We give a brief description of each dataset here.
Synthetic data 1: Our first synthetic dataset consists of two views and is generated in a manner akin to Yi et al.
12
which first chooses the cluster ci each sample belongs to and then generates each of the views Synthetic data 2: Our second synthetic dataset consists of three views. Moreover, the features are correlated. Each view still has two clusters. The cluster means and the covariance for three views are given below
Multiple feature dataset: The multiple features dataset consists in a set of 2000 handwritten numbers (digitalized pictures) described by six different views (Fourier coefficients, profile correlations, Karhunen–Loeve coefficients, pixel averages, Zernike moments, morphological features). Ten homogeneous classes (200 objects per class) have to be recovered. Image segmentation dataset: The image segmentation dataset consists in a set of 2310 samples described by two different views. It has seven classes and each sample has nine features. The descriptions of the two datasets are summarized in Table 1.
Evaluation
The evaluation of a clustering result is still an open problem, since one does not always know the true label of objects. Here, we choose three well-known different evaluations to measure the results quality of the compared approaches. The two evaluation criterion are described in Table 2.
Descriptions of MF, IS dataset, and the composition of each view.
IS: image segmentation dataset; MF: multiple feature dataset. RGB is one view of the dataset.
Evaluation criterion.
NMI: normalized mutual information; RI: rand index.
Experiments on two synthetic datasets.
CoFCM: collaborative fuzzy clustering from multiple weighted views; CoFKM: centralized method for multiple-view clustering; COWMVC(C): Centroid-Based Co-regularized weighting multi-view clustering; CoWMVC(P): pairwise co-regularized weighting multiview clustering algorithm; NMI: normalized mutual information; RI: rand index.
Experiments on two real-world datasets.
CoFCM: collaborative fuzzy clustering from multiple weighted views; CoFKM: centralized method for multiple-view clustering; COWMVC(C): Centroid-Based Co-regularized weighting multi-view clustering; CoWMVC(P): pairwise co-regularized weighting multiview clustering algorithm; IS: image segmentation dataset; MF: multiple feature dataset; NMI: normalized mutual information; RI: rand index.
From Tables 3 and 4, we can see that in most cases, the proposed CoWMVC(P) and CoWMVC(C) have a better evaluation index value under the same condition. That means, the clustering result of CoWMVC(P) and CoWMVC(C) is much closer to the actual class label than that of other three methods. For two real-world UCI datasets, it also can be noted that feature concatenation actually performs worse than single view on these two datasets.
For the four datasets we observe that the three multiview clustering algorithms have a very strong improvement over the single view approach, which is the main contribution to multiview approach.
Finally, it can be clearly seen from the results that the proposed algorithms can improve the performance of clustering.
We also test our model to see how the co-regularization parameter η affects the quality of the result. Our reported results are for the pairwise CoWMVC algorithm (Figure 1). Similar trends were observed for the centroid CoWMVC algorithm and therefore we do not report them here. For the range of η shown in the plot, these results indicate that although the performance of our algorithms depends on the co-regularized parameter η, it is still reasonably stable across a wide range of η.
NMI scores of co-regularized weighting multiview clustering as a function for (a) MF dataset and (b) IS dataset. IS: image Segmentation dataset; MF: multiple feature dataset; NMI: normalized mutual information.
Conclusion and future works
In this work, we studied the clustering problem on complex multiview data. We proposed a CoWMVC approach. The approach uses the philosophy of co-regularization to make the clusterings in different views agree with each other. Experimental results show that the proposed algorithms exhibit better multiview clustering performance.
Many perspectives and generalizations will be studied further in future. For example the possibility to deal with different number of clusters in each dataset, extending the proposed framework to the case where some of the views having missing data.
Footnotes
Declaration of conflicting interests
The author(s) declared no potential conflicts of interest with respect to the research, authorship, and/or publication of this article.
Funding
The author(s) disclosed receipt of the following financial support for the research, authorship, and/or publication of this article: This work is supported by the Specialized Research Fund for the Doctoral Program of Higher Education of China (Grant No. 20130093110009) and the National Natural Science Foundation of China (Grant No. 61373055).
