Skip to main content
Log in

Adaptively local consistent concept factorization for multi-view clustering

  • Methodologies and Application
  • Published:
Soft Computing Aims and scope Submit manuscript

Abstract

Many real-world datasets consist of multiple views of data items. The rough method of combining multiple views directly through feature concatenation cannot uncover the optimal latent structure shared by multiple views, which would benefit many data analysis applications. Recently, multi-view clustering methods have emerged and been applied to solving many machine learning problems. However, most multi-view clustering methods ignore the joint information of multi-view data or neglect the quality difference between different views of data, resulting in decreased learning performance. In this paper, we discuss a multi-view clustering algorithm based on concept factorization that effectively fuses useful information to derive a better representation for more effective clustering. We incorporate two regularizers into the concept factorization framework. Specifically, one regularizer is adopted to force the coefficient matrix to move smoothly on the underlying manifold. The other regularizer is used to learn the latent clustering structure from different views. Both of these regularizers are incorporated into the concept factorization framework to learn the latent representation matrix. Optimization problems are solved efficiently via an iterative algorithm. The experimental results on seven real-world datasets demonstrate that our approach outperforms the state-of-the-art multi-view clustering algorithms.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10
Fig. 11
Fig. 12
Fig. 13
Fig. 14
Fig. 15

Similar content being viewed by others

Notes

  1. http://mlg.ucd.ie/datasets.

  2. http://lig-membres.imag.fr/grimal/data.html.

  3. http://mlg.ucd.ie/datasets/segment.html.

  4. http://archive.ics.uci.edu/ml/datasets/Multiple+Features.

  5. http://archive.ics.uci.edu/ml/datasets.html.

  6. http://lig-membres.imag.fr/grimal/data.html.

References

  • Brunet JP, Tamayo P, Golub TR, Mesirov JP (2004) Metagenes and molecular pattern discovery using matrix factorization. Proc Natl Acad Sci 101(12):4164–4169

    Article  Google Scholar 

  • Cai D, He X, Han J (2010) Locally consistent concept factorization for document clustering. IEEE Trans Knowl Data Eng 23(6):902–913

    Article  Google Scholar 

  • Cai D, He X, Han J, Huang TS (2010) Graph regularized nonnegative matrix factorization for data representation. IEEE Trans Patt Anal Mach Intell 33(8):1548–1560

    Google Scholar 

  • Chao G, Sun S, Bi J (2017) A survey on multi-view clustering. arXiv preprint arXiv:171206246

  • Chaudhuri K, Kakade SM, Livescu K, Sridharan K (2009) Multi-view clustering via canonical correlation analysis. In: Proceedings of the 26th annual international conference on machine learning, pp 129–136

  • Dalal N, Triggs B (2005) Histograms of Oriented Gradients for Human Detection. In: Schmid C, Soatto S, Tomasi C (eds) International conference on computer vision and pattern recognition (CVPR ’05), IEEE computer society, San Diego, United States, vol 1, pp 886–893, 10.1109/CVPR.2005.177, https://hal.inria.fr/inria-00548512

  • Dempster AP, Laird NM, Rubin DB (1977) Maximum likelihood from incomplete data via the em algorithm. J Roy Stat Soc: Ser B (Methodol) 39(1):1–22

    MathSciNet  MATH  Google Scholar 

  • Gao S, Yu Z, Jin T, Yin M (2019) Multi-view low-rank matrix factorization using multiple manifold regularization. Neurocomputing 335:143–152

    Article  Google Scholar 

  • Greene D, Cunningham P (2009) A matrix factorization approach for integrating multiple data views. In: Joint European conference on machine learning and knowledge discovery in databases, Springer, pp 423–438

  • Guan Z, Zhang L, Peng J, Fan J (2015) Multi-view concept learning for data representation. IEEE Trans Knowl Data Eng 27(11):3016–3028

    Article  Google Scholar 

  • He X, Kan MY, Xie P, Chen X (2014) Comment-based multi-view clustering of web 2.0 items. In: Proceedings of the 23rd international conference on World wide web, ACM, pp 771–782

  • Jiang Y, Liu J, Li Z, Lu H (2014) Semi-supervised unified latent factor learning with multi-view data. Mach Vis Appl 25(7):1635–1645

    Article  Google Scholar 

  • Lee DD, Seung HS (2001) Algorithms for non-negative matrix factorization. In: Advances in neural information processing systems, pp 556–562

  • Liu J, Wang C, Gao J, Han J (2013) Multi-view clustering via joint nonnegative matrix factorization. In: Proceedings of the 2013 SIAM international conference on data mining, SIAM, pp 252–260

  • Lowe D (2004) Distinctive image features from scale-invariant keypoints. Int J Comput Vis 60(2):91–110

    Article  Google Scholar 

  • Lu M, Zhang L, Zhao XJ, Li FZ (2016) Constrained neighborhood preserving concept factorization for data representation. Knowl Based Syst 102:127–139

    Article  Google Scholar 

  • Richfield O, Alam MA, Calhoun V, Wang YP (2016) Learning schizophrenia imaging genetics data via multiple kernel canonical correlation analysis. In: 2016 IEEE international conference on bioinformatics and biomedicine (BIBM), IEEE, pp 507–511

  • Sun S (2013) A survey of multi-view machine learning. Neural Comput Appl 23(7–8):2031–2038

    Article  Google Scholar 

  • Wang H, Yang Y, Li T (2016) Multi-view clustering via concept factorization with local manifold regularization. In: 2016 IEEE 16th International Conference on Data Mining (ICDM), IEEE, pp 1245–1250

  • Wang H, Yang Y, Liu B, Fujita H (2019) A study of graph-based system for multi-view clustering. Knowl Based Syst 163:1009–1019

    Article  Google Scholar 

  • Wang M, Huang T, Deng L (2017) A group sparse joint nonnegative matrix factorization model for multiple data integration. In: 2017 14th International Computer Conference on Wavelet Active Media Technology and Information Processing (ICCWAMTIP), IEEE, pp 20–23

  • Xu W, Gong Y (2004) Document clustering by concept factorization. In: Proceedings of the 27th annual international ACM SIGIR conference on Research and development in information retrieval, ACM, pp 202–209

  • Xu W, Liu X, Gong Y (2003) Document clustering based on non-negative matrix factorization. In: Proceedings of the 26th annual international ACM SIGIR conference on Research and development in informaion retrieval, ACM, pp 267–273

  • Yang Y, Wang H (2018) Multi-view clustering: a survey. Big Data Min Anal 1(2):83–107

    Article  Google Scholar 

  • Zhang X, Gao H, Li G, Zhao J, Huo J, Yin J, Liu Y, Zheng L (2018) Multi-view clustering based on graph-regularized nonnegative matrix factorization for object recognition. Inf Sci 432:463–478

    Article  MathSciNet  Google Scholar 

Download references

Acknowledgements

This work was supported in part by Scientific Research Foundation for Advanced Talents (No. jit-b-202045), by the Natural Science Foundation of the Jiangsu Higher Education Institutions of China under Grant No. 19KJA550002, by the Six Talent Peak Project of Jiangsu Province of China under Grant No. XYDXX-054, and by the National Key Research and Development Program (Nos. 2018YFA0701700 and 2018YFA0701701).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Mei Lu.

Ethics declarations

Conflict of interest

The authors declare that they have no conflict of interest.

Ethical approval

This article does not contain any studies with human participants performed by any of the authors.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

(Proofs of theorem1)

(Proofs of theorem1)

The objective function O of ALCCF in Eq. (10) is bounded from below by zero obviously. In order to prove Theorem 1, we need show that the function O is nonincreasing under the update formulae Eq. (13), (14). Since the last two terms of O are only related to \(\mathbf{{V}}\), the update formula for \(\mathbf{{U}}\) in ALCCF is same as CF. Therefore, the convergence proof of CF can be used to show that O is nonincreasing under the update formula in Eq. (13). Please see (Lee and Seung 2001; Brunet et al. 2004) for details.

Next, we will prove that O is nonincreasing under the update formula Eq. 14. To prove it, we will use auxiliary function similar to that used in the expectation–maximization algorithm (Dempster et al. 1977). The auxiliary function is defined as follows; \(G(v,v')\) is an auxiliary function for F(v) if the conditions

$$\begin{aligned} G(v,v') \ge F(v),\quad G(v,v) = F(v) \end{aligned}$$

are satisfied.

Now, we show a very useful lemma as follows.

Lemma 1

If G is an auxiliary function of F, then F is nonincreasing under the update

$$\begin{aligned} {v^{(K + 1)}} = \mathop {\arg \min }\limits _v G(v,{v^{(K)}}) \end{aligned}$$
(20)

Proof

$$\begin{aligned} F({v^{(K + 1)}}) \le G({v^{(K + 1)}},{v^{(K)}}) \le G({v^{(K)}},{v^{(K)}}) = F({v^{(K)}}) \end{aligned}$$

\(\square \)

Next we will show that with a proper auxiliary function the update step for \(\mathbf{{V}}\) in Eq.(14) is exactly the update in Eq. (20).

We use \({F_{ab}}\) to denote the part of O which is only relevant to \({v_{ab}}\). And we can easily check that

$$\begin{aligned} F_{ab}'= & {} {\left( {\frac{{\partial O}}{{\partial {\mathbf{{V}}^{(v)}}}}} \right) _{ab}} = \left( - 2{\mathbf{{K}}^{(v)}}{\mathbf{{U}}^{(v)}} + 2{\mathbf{{V}}^{(v)}}{({\mathbf{{U}}^{(v)}})^T}{\mathbf{{K}}^{(v)}}{\mathbf{{U}}^{(v)}} \right. \nonumber \\&\left. + \sum \limits _{s = 1}^{{n_v}} {{\theta _{vs}}(\mathrm{{2}}{\mathbf{{V}}^{(s)}} - 2{({\mathbf{{V}}^{(v)}})^T})} \mathrm{{ + }}{2\alpha \mathbf{{LV}}} \right) _{ab} \nonumber \\\end{aligned}$$
(21)
$$\begin{aligned} F_{ab}^{''}= & {} {\left( {\frac{{\partial O}}{{\partial {\mathbf{{V}}^{(v)}}}}} \right) _{ab}} = 2({({\mathbf{{U}}^{(v)}})^T}{\mathbf{{K}}^{(v)}}{\mathbf{{U}}^{(v)}})_{bb}\nonumber \\&+ 2\sum \limits _{s = 1}^{{n_v}} {{\theta _{vs}}} + 2\alpha \mathbf{{L}}_{aa}^{(v)} \end{aligned}$$
(22)

Since the update is essentially element-wise, it is sufficient to show that each \({F_{ab}}\) is nonincreasing under the update step of Eq. (14).

Lemma 2

Function

$$\begin{aligned}&G(v,v_{ab}^{(K)}) = {F_{ab}}(v_{ab}^{(K)}) + F_{ab}'(v_{ab}^{(K)})(v -v_{ab}^{(K)})\nonumber \\&\quad + \frac{{{{\left( {{\mathbf{{V}}^{(v)}}{{({\mathbf{{U}}^{(v)}})}^T}{\mathbf{{K}}^{(v)}}{\mathbf{{U}}^{(v)}} \quad \quad + \sum \limits _{s = 1}^{{n_v}} {{\theta _{vs}}{\mathbf{{V}}^{(v)}}} + \alpha {\mathbf{{D}}^{(v)}}{\mathbf{{V}}^{(v)}}} \right) }_{ab}}}}{{v_{ab}^{(K)}}}{(v - v_{ab}^{(K)})^2} \quad \end{aligned}$$
(23)

is an auxiliary function for \({F_{ab}}\), the part of O which is only relevant to \({v_{ab}}\).

Proof

Firstly, it is obvious that \(G(v,v) = {F_{ab}}(v)\).

Secondly, we compare the Taylor series expansion of \({F_{ab}}(v)\)

$$\begin{aligned}&{F_{ab}}(v) = {F_{ab}}(v_{ab}^{(K)}) + F_{ab}'(v_{ab}^{(K)})(v - v_{ab}^{(K)})\nonumber \\&\quad + {\left( ({{{({\mathbf{{U}}^{(v)}})^T}{\mathbf{{K}}^{(v)}}{\mathbf{{U}}^{(v)}})_{bb} + \sum \nolimits _{s = 1}^{{n_v}} {{\theta _{vs}}{\mathbf{{V}}^{(s)}}} + \alpha \mathbf{{L}}_{aa}^{(v)})}} \right) }{(v - v_{ab}^{(K)})^2} \nonumber \\ \end{aligned}$$
(24)

with Eq.(23) to find that \(G(v,v_{ab}^{(K)}) \ge {F_{ab}}(v)\) is equivalent to

$$\begin{aligned} \begin{aligned}&\frac{{{\mathbf{{V}}^{(v)}}{({\mathbf{{U}}^{(v)}})^T}{\mathbf{{K}}^{(v)}}{\mathbf{{U}}^{(v)}} + \sum \nolimits _{s = 1}^{{n_v}} {{\theta _{vs}}{\mathbf{{V}}^{(v)}}} + \alpha {\mathbf{{D}}^{(v)}}{\mathbf{{V}}^{(v)}}}}{{v_{ab}^{(K)}}}\\&\ge {\mathbf{{K}}^{(v)}}{\mathbf{{U}}^{(v)}} + \sum \limits _{s = 1}^{{n_v}} {{\theta _{vs}}{\mathbf{{V}}^{(s)}}} + \alpha \mathbf{{L}}_{aa}^{(v)}. \end{aligned} \end{aligned}$$
(25)

We have

$$\begin{aligned} \begin{aligned}&{\left( {{\mathbf{{V}}^{(v)}}{({\mathbf{{U}}^{(v)}})^T}{\mathbf{{K}}^{(v)}}{\mathbf{{U}}^{(v)}}} \right) _{ab}} \\&\quad = \sum \limits _{l = 1}^k {v_{al}^{(K)}{{({({\mathbf{{U}}^{(v)}})^T}{\mathbf{{K}}^{(v)}}{\mathbf{{U}}^{(v)}})}_{lb}}} \\&\quad \ge v_{ab}^{(K)}{({({\mathbf{{U}}^{(v)}})^T}{\mathbf{{K}}^{(v)}}{\mathbf{{U}}^{(v)}})_{bb}} \end{aligned} \end{aligned}$$
(26)

and

$$\begin{aligned} \begin{aligned} \alpha {({\mathbf{{D}}^{(v)}}{\mathbf{{V}}^{(v)}})_{ab}}&= \alpha \sum \nolimits _{j = 1}^M {\mathbf{{D}}_{aj}^{(v)}v_{jb}^{(K)} \ge } \alpha (\mathbf{{D}}_{aa}^{(v)}v_{ab}^{(K)})\\&\ge \alpha {({\mathbf{{D}}^{(v)}} - {\mathbf{{S}}^{(v)}})_{aa}}v_{ab}^{(K)} = \alpha \mathbf{{L}}_{aa}^{(v)}v_{ab}^{(K)}. \end{aligned} \end{aligned}$$
(27)

Therefore, Eq. (25) holds and we get \(G(v,v_{ab}^{(K)}) \ge {F_{ab}}(v)\). \(\square \)

Now, we can show the convergence of Theorem 1:

Proof

of Theorem 1 Replacing \(G(v,v_{ab}^{(K)})\) in Eq. (20) by Eq.(23) results in the update rule:

$$\begin{aligned}&v_{ab}^{(K + 1)} = v_{ab}^{(K)} \nonumber \\&\quad - v_{ab}^{(K)}\frac{{F_{ab}'(v_{ab}^{(K)})}}{{2{{\left( {{\mathbf{{V}}^{(v)}}{{({\mathbf{{U}}^{(v)}})}^T}{\mathbf{{K}}^{(v)}}{\mathbf{{U}}^{(v)}}} \right) }_{ab}} + 2{{\left( {\sum \nolimits _{s = 1}^{{n_v}} {{\theta _{vs}}{\mathbf{{V}}^{(v)}}} } \right) }_{ab}} + 2\alpha {{\left( {{\mathbf{{D}}^{(v)}}{\mathbf{{V}}^{(v)}}} \right) }_{ab}}}}\nonumber \\&\quad = v_{ab}^{(K)}\frac{{{{\left( {{\mathbf{{K}}^{(v)}}{\mathbf{{U}}^{(v)}} + \sum \nolimits _{s = 1}^{{n_v}} {{\theta _{vs}}{\mathbf{{V}}^{(s)}}} + \alpha {\mathbf{{S}}^{(v)}}{\mathbf{{V}}^{(v)}}} \right) }_{ab}}}}{{{{\left( {{\mathbf{{V}}^{(v)}}{{({\mathbf{{U}}^{(v)}})}^T}{\mathbf{{K}}^{(v)}}{\mathbf{{U}}^{(v)}} + \sum \nolimits _{s = 1}^{{n_v}} {{\theta _{vs}}{\mathbf{{V}}^{(v)}}} + \alpha {\mathbf{{D}}^{(v)}}{\mathbf{{V}}^{(v)}}} \right) }_{ab}}}} \end{aligned}$$
(28)

Since Eq.(20) is an auxiliary function, \({F_{ab}}\) is nonincreasing under this updating rule. \(\square \)

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Lu, M., Zhang, L. & Li, F. Adaptively local consistent concept factorization for multi-view clustering. Soft Comput 26, 1043–1055 (2022). https://doi.org/10.1007/s00500-021-06526-2

Download citation

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s00500-021-06526-2

Keywords

Navigation