Skip to main content
Log in

Fast density-weighted low-rank approximation spectral clustering

  • Published:
Data Mining and Knowledge Discovery Aims and scope Submit manuscript

Abstract

While spectral clustering can produce high-quality clusterings on small data sets, computational cost makes it infeasible for large data sets. Affinity Propagation (AP) has a limitation that it is hard to determine the value of parameter ‘preference’ which can lead to an optimal clustering solution. These problems limit the scope of application of the two methods. In this paper, we develop a novel fast two-stage spectral clustering framework with local and global consistency. Under this framework, we propose a Fast density-Weighted low-rank Approximation Spectral Clustering (FWASC) algorithm to address the above issues. The proposed algorithm is a high-quality graph partitioning method, and simultaneously considers both the local and global structure information contained in the data sets. Specifically, we first present a new Fast Two-Stage AP (FTSAP) algorithm to coarsen the input sparse graph and produce a small number of final representative exemplars, which is a simple and efficient sampling scheme. Then we present a density-weighted low-rank approximation spectral clustering algorithm to operate those representative exemplars on the global underlying structure of data manifold. Experimental results show that our algorithm outperforms the state-of-the-art spectral clustering and original AP algorithms in terms of speed, memory usage, and quality.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Similar content being viewed by others

References

  • Baker C (1977) The numerical treatment of integral equations. Clarendon, Oxford

    MATH  Google Scholar 

  • Balasubramanian M, Schwartz EL, Tenenbaum JB, de Silva V, Langford JC (2002) The Isomap algorithm and topological stability. Science 295(5552): 7

    Article  Google Scholar 

  • Belabbas M-A, Wolfe PJ (2009) Spectral methods in machine learning and new strategies for very large datasets. Proc Natl Acad Sci USA 106: 369–374

    Article  Google Scholar 

  • Berkhin P (2002) Survey of clustering data mining techniques. Technical Report. Accrue Software. http://www.ee.ucr.edu/~barth/EE242/clustering_survey.pdf

  • Ding CHQ, He X, Zha H, Gu M, Simon HD (2001) A min–max cut algorithm for graph partitioning and data clustering. In: Proceedings of IEEE international conference on data mining (ICDM), pp 107–114

  • Donath WE, Hofmann AJ (1973) Lower bounds for the partitioning of graphs. IBM J Res Dev 17: 420–425

    Article  MATH  Google Scholar 

  • Drineas P, Mahoney MW (2005) On the Nyström method for approximating a Gram matrix for improved kernel-based learning. J Mach Learn Res 6: 2153–2175

    MathSciNet  Google Scholar 

  • Duda RO, Hart PE, Stork DG (2001) Pattern classification. Wiley, New York

    MATH  Google Scholar 

  • Fei-Fei L, Fergus R, Perona P (2004) Learning generative visual models from few training examples: an incremental Bayesian approach tested on 101 object categories. In: IEEE conference on computer vision and pattern recognition. Workshop on generative model based vision, pp 178–178

  • Fowlkes C, Belongie S, Chung F, Malik J (2004) Spectral grouping using the Nyström method. IEEE Trans Pattern Anal Mach Intell 26(2): 214–225

    Article  Google Scholar 

  • Freitas ND, Wang Y, Mahdaviani M, Lang D (2006) Fast Krylov methods for N-body learning. Adv Neural Inf Process Syst 18: 251–258

    Google Scholar 

  • Frey BJ, Dueck D (2007) Clustering by passing messages between data points. Science 305(5814): 972–976

    Article  MathSciNet  Google Scholar 

  • Georghiades AS, Belhumeur PN, Kriegman DJ (2001) From few to many: illumination cone models for face recognition under variable lighting and pose. IEEE Trans Pattern Anal Mach Intell 23(6): 643–660

    Article  Google Scholar 

  • Givoni I, Frey BJ (2009) Semi-supervised affinity propagation with instance-level constraints. In: Proceedings of the 14th international workshop on Artificial Intelligence and Statistics (AISTATS), vol 5. JMLR W&CP, Clearwater Beach, FL, pp 161–168

  • Hagen L, Kahng A (1992) New spectral methods for ratio cut partitioning and clustering. IEEE Trans Computer Aided Des Integr Circuits Syst 11(9): 1074–1085

    Article  Google Scholar 

  • Han J, Kamber M (2001) Data mining: concepts and techniques. Morgan Kaufmann Publishers, San Francisco

    Google Scholar 

  • Jain A, Murty M, Flynn P (1999) Data clustering: a review. ACM Comput Surv 31(3): 264–323

    Article  Google Scholar 

  • Jia Y, Wang J, Zhang C, Hua X (2008) Finding image exemplars using fast sparse affinity propagation. In: Proceedings of ACM multimedia, pp 639–642

  • Johnson W, Lindenstrauss J (1984) Extensions of Lipschitz mappings into a Hilbert space. Contemp Math 26: 189–206

    MathSciNet  MATH  Google Scholar 

  • Kschischang F, Frey BJ, Loeliger H-A (2001) Factor graphs and the sum-product algorithm. IEEE Trans Inform Theory 47(2): 498–519

    Article  MathSciNet  MATH  Google Scholar 

  • Kumar S, Mohri M, Talwalkar A (2009) Sampling techniques for the Nyström method. In: Proceedings of the 14th international workshop on Artificial Intelligence and Statistics (AISTATS), vol 5. JMLR W&CP, Clearwater Beach, FL, pp 304–311

  • Lazebnik S, Schmid C, Ponce J (2006) Beyond bags of features: spatial pyramid matching for recognizing natural scene categories. In: Proceedings of IEEE conference on computer vision and pattern recognition (CVPR), pp 2169–2178

  • Lee JA, Verleysen M (2005) Nonlinear dimensionality reduction of data manifolds with essential loops. Neurocomputing 67(1): 29–53

    Google Scholar 

  • Leone M, Sumedha , Weigt M (2007) Clustering by soft-constraint affinity propagation: applications to gene-expression data. Bioinformatics 23(20): 2708–2715

    Article  Google Scholar 

  • Liu T, Moore AW, Gray AG, Yang K (2005) An investigation of practical approximate nearest neighbor algorithms. Adv Neural Inf Process Syst 17: 825–832

    Google Scholar 

  • Liu W, He J, Chang S (2010) Large graph construction for scalable semi-supervises learning. In: Proceedings of the 27th international conference on machine learning (ICML), pp 679-686

  • Madigan D, Raghavan N, Dumouchel W, Nason M, Posse C, Ridgeway G (2002) Likelihood-based data squashing: a modeling approach to instance construction. Data Min Knowl Discov 6: 173–190

    Article  MathSciNet  MATH  Google Scholar 

  • Mahadevan S (2008) Fast spectral learning using Lanczos eigenspace projections. In: Proceedings of the 23rd national conference on artificial intelligence (AAAI), pp 1472–1475

  • Mitra P, Murthy CA, Pal SK (2002) Density-based multiscale data condensation. IEEE Trans Pattern Anal Mach Intell 24(6): 734–747

    Article  Google Scholar 

  • Ng A, Jordan M, Weiss Y (2002) On spectral clustering: analysis and an algorithm. Adv Neural Inf Process Syst 14: 849–856

    Google Scholar 

  • Ouimet M, Bengio Y (2005) Greedy spectral embedding. In: Proceedings of the 10th international workshop on artificial intelligence and statistics (AISTATS), pp 253-260

  • Papadimitriou CH, Steiglitz K (1998) Combinatorial optimization: algorithms and complexity. Dover, New York

    MATH  Google Scholar 

  • Shi J, Malik J (2000) Normalized cuts and image segmentation. IEEE Trans Pattern Anal Mach Intell 22(8): 888–905

    Article  Google Scholar 

  • Song Y, Chen W-Y, Bai H, Lin C-J, Chang EY (2008) Parallel spectral clustering. In: Proceedings of learning and principles and practice of knowledge discovery in databases (ECML/PKDD), pp 374–389

  • Strehl A, Ghosh J (2002) Cluster ensembles—a knowledge reuse framework for combining multiple partitions. J Mach Learn Res 3: 583–617

    Article  MathSciNet  Google Scholar 

  • Talwalkar A, Kumar S, Rowley H (2008) Large-scale manifold learning. In: Proceedings of IEEE conference on computer vision and pattern recognition (CVPR), pp 1–8

  • Tenenbaum JB, de Silva V, Langford JC (2000) A global geometric framework for nonlinear dimensionality reduction. Science 290(5500): 2319–2323

    Article  Google Scholar 

  • Vidal R, Ma Y, Piazzi J (2004) A new GPCA algorithm for clustering subspaces by fitting, differentiating and dividing polynomials. In: Proceedings of IEEE conference on computer vision and pattern recognition (CVPR), pp 510–517

  • von Luxburg U (2007) A tutorial on spectral clustering. Stat Comput 17(4): 395–416

    Article  MathSciNet  Google Scholar 

  • Wang F, Zhang C (2008) Label propagation through linear neighborhoods. IEEE Trans Knowl Data Eng 20(1): 55–67

    Article  Google Scholar 

  • Williams C, Seeger M (2001) Using the Nyström method to speed up kernel machines. Adv Neural Inf Process Syst 13: 682–688

    Google Scholar 

  • Wittkop T, Baumbach J, Lobo F, Rahmann S (2007) Large scale clustering of protein sequences with FORCE–A layout based heuristic for weighted cluster editing. BMC Bioinformatics 8(1): 396

    Article  Google Scholar 

  • Wu M, Schölkopf B (2007) A local learning approach for clustering. Adv Neural Inf Process Syst 19: 1529–1536

    Google Scholar 

  • Xiao J, Wang J, Tan P, Quan L (2007) Joint affinity propagation for multiple view segmentation. In: Proceedings of IEEE 11th international conference on computer vision (ICCV), pp 1–7

  • Xu R, Wunsch D (2005) Survey of clustering algorithms. IEEE Trans Neural Netw 16(3): 645–678

    Article  Google Scholar 

  • Yan D, Huang L, Jordan M (2009) Fast approximate spectral clustering. In: Proceedings of the 15th ACM SIGKDD international conference on knowledge discovery and data mining, pp 907–916

  • Yann L, Corinna C (2009) The MNIST database of handwritten digits. http://yann.lecun.com/exdb/mnist/

  • Zelnik-Manor L, Perona P (2005) Self-tuning spectral clustering. Adv Neural Inf Process Syst 17: 1601–1608

    Google Scholar 

  • Zhang K, Kwok J (2008) Improved Nyström low rank approximation and error analysis. In: Proceedings of the 25th international conference on machine learning (ICML), pp 273–297

  • Zhang K, Kwok J (2009) Density-weighted Nyström method for computing large kernel Eigen-Systems. Neural Comput 21: 121–146

    Article  MathSciNet  MATH  Google Scholar 

  • Zhou D, Bousquet O, Lal TN, Weston J, Schölkopf B (2004) Learning with local and global consistency. Adv Neural Inf Process Syst 16: 321–328

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Fanhua Shang.

Additional information

Responsible editor: Charu Aggarwal.

Rights and permissions

Reprints and permissions

About this article

Cite this article

Shang, F., Jiao, L.C., Shi, J. et al. Fast density-weighted low-rank approximation spectral clustering. Data Min Knowl Disc 23, 345–378 (2011). https://doi.org/10.1007/s10618-010-0207-5

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s10618-010-0207-5

Keywords

Navigation