Next Article in Journal
Recurrence-Based Synchronization Analysis of Weakly Coupled Bursting Neurons under External ELF Fields
Next Article in Special Issue
An Image Encryption Scheme Based on Logistic Quantum Chaos
Previous Article in Journal
Finite Element Iterative Methods for the Stationary Double-Diffusive Natural Convection Model
Previous Article in Special Issue
Free-Space QKD with Modulating Retroreflectors Based on the B92 Protocol
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Quantum Density Peak Clustering Algorithm

1
College of Cyber Security, Jinan University, Guangzhou 510632, China
2
College of Information Science and Technology, Jinan University, Guangzhou 510632, China
3
Guangxi Key Laboratory of Cryptography and Information Security, Guilin 541004, China
*
Author to whom correspondence should be addressed.
Entropy 2022, 24(2), 237; https://doi.org/10.3390/e24020237
Submission received: 17 December 2021 / Revised: 28 January 2022 / Accepted: 1 February 2022 / Published: 3 February 2022
(This article belongs to the Special Issue Practical Quantum Communication)

Abstract

:
A widely used clustering algorithm, density peak clustering (DPC), assigns different attribute values to data points through the distance between data points, and then determines the number and range of clustering by attribute values. However, DPC is inefficient when dealing with scenes with a large amount of data, and the range of parameters is not easy to determine. To fix these problems, we propose a quantum DPC (QDPC) algorithm based on a quantum D i s t C a l c circuit and a Grover circuit. The time complexity is reduced to O ( log ( N 2 ) + 6 N + N ) , whereas that of the traditional algorithm is O ( N 2 ) . The space complexity is also decreased from O ( N · log N ) to O ( log N ) .

1. Introduction

Cluster analysis originated from taxonomy, as an ancient skill mastered by human beings. In the past, people used to classify goods based on their experience and professional knowledge. With the development of modern society, people have higher and higher requirements for classification [1,2]. Classification based only on experience and professional knowledge has been gradually eliminated, and computer technology is now used for cluster analysis, using algorithms to address huge and complex cluster tasks [3,4]. Therefore, clustering algorithms have been proposed for applications in various settings [5,6]. Moreover, the world of massive data that we live in also makes the clustering process indispensable. Many research fields are faced with the problem of a large amount of data [7,8]. If there is no preprocessing such as clustering or data dimension reduction, it is difficult to carry out subsequent analysis [9,10,11]. For example, in the area of machine learning, the original entry of almost all important algorithms is a large amount of large-scale data. It is difficult to use these data without clustering or a dimensionality reduction [12,13,14]. In the field of quantum communication, quantum communication equipment is only supplied to few large companies. Many parties in quantum communication may be classical. Clustering algorithms can help communication parties to deal with the transmitted information more conveniently [15,16,17]. In the area of data dimension reduction, we are familiar with the principal component analysis algorithm (PCA) [18], multidimensional scaling (MDS), linear discrimination (LDA), locally linear embedding (LLE), and so on [19,20,21,22]. However, the dimension reduction algorithm will inevitably reduce the attribute value of the data. If the operation is improper, the data will lose accuracy and the results will have deviated. A clustering algorithm can be used to avoid such problems. Nowadays, clustering algorithms can be divided in the following way.
Partition-based clustering algorithms include K-means [23], K-medians [24], and kernel K-means algorithms [25]. Hierarchy-based clustering algorithms include BIRCH, CURE, and the CHAMELEON algorithm [26]. Density-based clustering algorithms include DBSCAN, mean-shift (MS) [27], and the density peak clustering algorithm (DPC) [28]. Each clustering algorithm has its own advantages and disadvantages, and each algorithm has its own suitable scenarios [29]. The advantage of the DPC algorithm is that there is no need to define the number of clusters, as in the K-means algorithm. Secondly, it can detect non-spherical data, which has high application value in computer image processing. In addition, it can automatically identify abnormal points, which is also a prominent advantage of many clustering algorithms.
In 2014, Rodriguez and Laio proposed a DPC algorithm, which can automatically find the cluster center and achieve efficient clustering of arbitrarily shaped data sets [30]. DPC is a clustering algorithm based on density, and its input parameters are less than those of the K-means algorithm [31,32] and the K-medians algorithm [33,34]. The process of DPC clustering does not need to map data to vector space, which reduces the computational complexity of the algorithm.
However, the DPC algorithm still has its drawbacks. When it deals with large amounts of data, the speed of the algorithm is significantly reduced. The algorithm computes the distance between the current data point and each data point in the set, so the complexity of this operation is O ( N 2 ) , whereas N represents the number of data points [35]. At the same time, the process of the algorithm stores the distance between each data point and its remaining data points, which requires a large amount of storage space.
Some years ago, quantum technology was introduced to speed up the classical algorithms with large data volume, such as the Internet of Things industry and computer vision [36,37,38]. Typical quantum algorithms include the quantum K-means algorithm [32] and quantum principal component analysis [18]. They are not simple quantum version of classical algorithms. The running speeds of these quantum algorithms are greatly reduced. In this paper, we propose a QDPC algorithm, which applies a quantum D i s t C a l c circuit to speed up the DPC algorithm.
In Section 2, the principle and flow of the classical DPC algorithm are introduced in detail. In Section 3, we propose the QDPC algorithm and its corresponding quantum circuit. In Section 4, the simulation experiments are discussed. An analysis of complexity and our conclusions are presented in Section 5.

2. Preliminary

2.1. Notation and Definitions

DPC is an algorithm that does not require iteration and can find the clustering center in one run. Distance information is the most important form of information that one must collect in the DPC algorithm. Based on the distance, one can compute the local density value.
The main ideas of DPC are based on the following assumptions:
*
The clustering center has a relatively high local density value and is surrounded by data points with a low local density value.
*
The clustering center is far away from any point with a higher local density value.
For each data point x i , the algorithm computes two attribute values of the data point: its local density ρ i and its distance δ i from the nearest higher density point. Both attribute values depend only on the distance d i j between the current data point x i and the rest of the data points x j .
The local density ρ i of data point x i is defined as
ρ i = j χ ( d i j d c ) ,
where d i j is the distance between data point x i and x j , d c is the cutoff distance. The function χ is defined as
χ ( x ) = 1 x < 0 0 o t h e r w i s e ,
which indicates the number of data points with distance from the data point x i less than the cutoff distance.
The distance from the higher density point of data point x i is defined as
δ i = min j , ρ i < ρ j ( d i j ) ,
where δ i records the nearest distance from data point x i to all data points with higher local density. If δ i is very small, there is a data point x j with a higher local density around this data point x i . As for the data point with the highest local density, it cannot find a data point with a higher local density, and its distance is δ i = max j ( d i j ) conventionally. It can be found that when the distance δ i is relatively large, this data point is the clustering center.
Therefore, after the two attribute values ρ i and δ i of each data point are obtained, these data points are divided according to the rules.
  • If the value sof ρ i and δ i are both anomalously large, it is the clustering center;
  • If the value of ρ i is relatively large and δ i is relatively small, it is the point in a cluster;
  • If the value of ρ i is relatively small and δ i is relatively large, it is an outlier.
According to the above rules, the algorithm can accurately find every clustering center of the cluster and cluster each data point.

2.2. The Workflow of the Classical DPC Algorithm

The main processes of the algorithm consist of calculating two attribute ρ i and δ i values of each data point. Suppose we have a data set with large amounts of data points D = { x 1 , x 2 , x 3 , , x N } , and the dimension of each data point is d. The steps of the DPC algorithm are as follows:
a.
Calculate the local density ρ i of each data point x i .
b.
For each data point x i , the nearest distance of x i is found in all data points with higher local density than x i , and record this distance as δ i .
c.
According to ρ i and δ i of each data point to determine the clustering center. If ρ i and δ i of a data point are relatively large, it is the clustering center.
d.
Assign each data point to the nearest clustering center.
It should be noted that if δ i of a data point is large and ρ i is small, then the point is an exception. It does not need to be assigned to any cluster.

3. QDPC Algorithm

The classical algorithm takes the largest proportion of time to calculate the distance in the whole algorithm, so quantum circuits are used to optimize this part [39,40]. In quantum technology, fidelity is an important concept, which is similar to cosine similarity [41,42] in the classical framework. Fidelity can measure the similarity between two quantum states. If the value of fidelity is 1, the two quantum states are the same; if the value of fidelity is 0, the two quantum states are orthogonal. Therefore, the distance between data points can be calculated via fidelity only if the classical data are encoded into a quantum state.
The most commonly used quantum circuit to achieve fidelity is the S w a p T e s t . This quantum circuit was proposed by Aïmeur et al in [43]. By taking the inner product of two quantum states | ϕ and | ψ , the S w a p T e s t circuit is used to calculate the fidelity of quantum states, as shown in Figure 1.
Based on the S w a p T e s t circuit, the quantum D i s t C a l c circuit [44] in Figure 2 can calculate the distance between data points x i and x j . The distance is stored in the third register.

Procedure of the QDPC Algorithm

Consider a set with N data points D = { x 1 , x 2 , x 3 , , x N } . The dimension of each data point is d. Regardless of the number of clusters, QDPC will calculate two attribute values ρ i and δ i for each data point. Then the clustering center is determined using these two attribute values.
An overview of the circuit for solving the QDPC is shown in Figure 3.
The procedure used to cluster x i includes the following seven steps:
(i).
Prepare six registers in | 0 log N | 0 log N | 0 n + log ( n + d ) | 0 | 0 | 0 log N , and apply an H gate on each qubit in the first and second registers. The third register records the quantum state of the distance from two data points x i and x j . The fourth register stores the intermediate conversion value a i j , which will be explained in more detail later. The fifth register is an ancillary qubit. The last register, the sixth register, records the attribute value ρ i . By means of quantum D i s t C a l c , the system state is
| Ψ 0 = 1 N i = 1 N j = 1 N | i | j | d ( x i , x j ) | 0 | 0 | 0 log N + | G ,
where | G is a garbage state.
(ii).
Set a desired threshold as d m a x and set a i j { 0 , 1 } to indicate whether two data points x i and x j are close together. The value of a i j is 1 if the distance of two data points d ( x i , x j ) d m a x , otherwise 0. Then we can easily store this value a i j in the fourth register. The system state is
| Ψ 1 = 1 N i = 1 N j = 1 N | i | j | d ( x i , x j ) | a i j | 0 | 0 log N + | G .
(iii).
Take a control-sum operation on the first, fourth, and sixth register. The first register | i is the control qubit, and the sixth register stores the sum of the values of a i j , whereas the index i is fixed. Since the local density property of the data point x i is ρ i = j = 1 N a i j , the value stored in the sixth register is | ρ i . Now the system is
| Ψ 2 = 1 N i = 1 N | i j = 1 N | j | d ( x i , x j ) | a i j | 0 | ρ i + | G .
(iv).
Perform a control conditional rotation [45], where the first and the second registers are control qubits, and the fifth ancillary register is the target. Set the fifth register to | 1 , when ρ j > ρ i , otherwise set to | 0 . The whole system is divided into two parts, as shown below
| Ψ 3 = 1 N i = 1 N | i j , ρ j > ρ i | j | d ( x i , x j ) | a i j | 1 + j , ρ j ρ i | j | d ( x i , x j ) | a i j | 0 | ρ i + | G .
(v).
Apply a projection operation { | 0 0 | , | 1 1 | } on the fifth register, and keep the state when the measurement result is | 1 1 | . The system is
| Ψ 4 = Γ | Ψ 3 = α i = 1 N j , ρ i < ρ j N | i | j | d ( x i , x j ) | a i j | 1 | ρ i + | G ,
where α is the normalized parameter, and | α | 2 = 1 . The third register and the last register store the attribute values δ i and ρ i of each data point x i , respectively.
(vi).
Perform a bit flip operation [46] on the third register, and the value is changed from d ( x i , x j ) to d ( x i , x j ) ¯ . By doing this, the minimum value in the third register becomes the maximum value. In order to make the following Grover algorithm run under more convenient conditions, we change the target of the search to the maximum value of the two attributes. Data points that meet these two requirements are the center of clustering. Now the system is
| Ψ 5 = α i = 1 N j , ρ i < ρ j N | i | j | d ( x i , x j ) ¯ | a i j | 1 | ρ i + | G .
(vii).
Apply the Grover algorithm [47] to find the index i of data point x i with maximum ρ i and the index j of the found data point x i with maximum δ i with a full successful probability. The index i that meets the requirements is the center of a cluster.

4. Simulation Results

We clustered three differently distributed (horizontally, circularly, and discrete) data sets using our QDPC algorithm, implemented on Baidu’s quantum platform Paddle Quantum. Limited to the lack of QRAM devices, thread concurrency was used to read out all the data in a data set at one time. Data were generated by a random function with s e e d = 21 , and the number of data N was fixed as 20 , 40 , 80 , 250 , 500 , and 1000. Table 1 gives the common evaluation indicators purity, F-score and adjusted Rand index (ARI) of the two algorithms on the circularly distributed data. In the table, all the values lie between 0.95 and 1. When N = 20 , 40 , 80 , the clustering results of the DPC algorithm are the same as those of the QDPC algorithm. For N = 20 , 40 , 80 , the values of QDPC are greater than those of DPC, so the QDPC performs better than the DPC.
We also depict the clustering performance of two algorithms when N is fixed at 250 in Figure 4. All the points are accompanied with their indexes. The points colored yellow are the centers of the clusters. Other points colored the same are clustering together, so both DPC and QPDC cluster the data into two groups. But DPC performs slightly worse than QDPC, since the points with indexes 14, 58 and 33 are colored with green, which should be purple.
The experiment was repeated 10 times and the average running times are recorded in Table 2. It can be seen that with increasing N, the running time of QDPC increases linearly, and that of the DPC increases exponentially. When N is fixed at 250 , 500 , or 1000, QDPC is faster than DPC, but when N is less than 80, DPC is faster than QDPC. The reason for this may be the fact that we simulated these results on a classical computer. If we ran the QDPC algorithm on a real quantum computer, the results may show an improvement.

5. Discussion and Conclusions

We now analyze the complexity of the QDPC algorithm step by step. In (i), a quantum D i s t C a l c circuit is applied to obtain the distance between two data points. The time complexity of this step depending on the distance definition is log ( N · N ) [44]. In (ii) and (iii), we convert the value d ( x i , x j ) into a i j and add up a i j . The time complexity of these two steps can be measured by the number of register accesses and the quantum addition circuit. Therefore, in general, the time complexity of this part is O ( N + 5 N ) [48]. In (iv), (v), and (vi), we perform the conditional rotation operation, the projection operation, and the bit flip operation. The time complexity is relatively negligible compared with other steps. Finally, step (vii) requires the application of the Grover algorithm, which introduces a time complexity of O ( N ) [49,50]. Thus, the time complexity of the whole algorithm is O ( log ( N 2 ) + 6 N + N ) . The space complexity is the space size of the quantum registers, i.e., log N + log N + n + log ( n + d ) + 1 + 1 + log N = 3 log N + n + log ( n + d ) + 2 .
For the DPC algorithm, the most time-consuming step is to calculate the distance between data points. It can be seen that the total distances of 1 2 N ( N 1 ) times need to be calculated [51]. So the complexity of the classical DPC algorithm is O ( N 2 ) . The space complexity of DPC depends on the space stored, ρ i and δ i , for each point. The space required is N · log N + N · n + log ( n + d ) bits.
A corresponding comparison between classical and quantum algorithms is shown in Table 3. Based on Table 3, the QDPC algorithm costs less than the DPC algorithm in terms of both time and space complexities.
In this paper, we have proposed a QDPC algorithm that is more efficient in both time and space than the classical algorithm. We applied it to two key circuits, a quantum D i s t C a l c circuit and a Grover circuit. The quantum D i s t C a l c circuit calculates the distance between data points in the data set, from which two important attribute values, ρ i and δ i , required by the QDPC algorithm are obtained. Then, the Grover algorithm is used to search the index of clustering center points that meet the conditions from the data set. In the future, we will investigate some possible application scenarios of the QDPC algorithm and compare the efficiency of algorithms on different data set structures.

Author Contributions

Z.W.: Conceptualization, methodology, software, data curation, writing—original draft. T.S.: supervision, project administration, writing—review and editing. Y.Z.: methodology, investigation. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Guangxi Key Laboratory of Cryptography and Information Security (No. GCIS201922), and Fundamental Research Funds for the Central Universities (No. 21620433).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The authors would like to thank the reviewers for their valuable comments and suggestions.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Aïmeur, E.; Brassard, G.; Gambs, S. Quantum clustering algorithms. In Proceedings of the 24th International Conference on Machine Learning, Corvalis, OR, USA, 20–24 June 2007; pp. 1–8. [Google Scholar]
  2. Nielsen, M.A.; Chuang, I. Quantum computation and quantum information. Am. J. Phys. 2002, 70, 558. [Google Scholar] [CrossRef] [Green Version]
  3. Harrow, A.W.; Hassidim, A.; Lloyd, S. Quantum Algorithm for Linear Systems of Equations. Phys. Rev. Lett. 2009, 103, 150502. [Google Scholar] [CrossRef] [PubMed]
  4. Yu, C.H.; Gao, F.; Wang, Q.L.; Wen, Q.Y. Quantum algorithm for association rules mining. Phys. Rev. A 2016, 94, 042311. [Google Scholar] [CrossRef] [Green Version]
  5. Kerenidis, I.; Prakash, A. Quantum recommendation systems. arXiv 2016, arXiv:1603.08675. [Google Scholar]
  6. Greche, L.; Jazouli, M.; Es-Sbai, N.; Majda, A.; Zarghili, A. Comparison between Euclidean and Manhattan distance measure for facial expressions classification. In Proceedings of the 2017 International Conference on Wireless Technologies, Embedded and Intelligent Systems (WITS), Fez, Morocco, 19–20 April 2017; pp. 1–4. [Google Scholar]
  7. Wiebe, N.; Kapoor, A.; Svore, K.M. Quantum nearest-neighbor algorithms for machine learning. Quantum Inf. Comput. 2018, 15, 318–358. [Google Scholar]
  8. Zidan, M.; Abdel-Aty, A.H.; El-shafei, M.; Feraig, M.; Al-Sbou, Y.; Eleuch, H.; Abdel-Aty, M. Quantum classification algorithm based on competitive learning neural network and entanglement measure. Appl. Sci. 2019, 9, 1277. [Google Scholar] [CrossRef] [Green Version]
  9. Lloyd, S.; Mohseni, M.; Rebentrost, P. Quantum algorithms for supervised and unsupervised machine learning. arXiv 2013, arXiv:1307.0411. [Google Scholar]
  10. Kerenidis, I.; Landman, J.; Luongo, A.; Prakash, A. q-means: A quantum algorithm for unsupervised machine learning. arXiv 2018, arXiv:1812.03584. [Google Scholar]
  11. Shaikh, T.A.; Ali, R. Quantum computing in big data analytics: A survey. In Proceedings of the 2016 IEEE International Conference on Computer and Information Technology (CIT), Nadi, Fiji, 8–10 December 2016; pp. 112–115. [Google Scholar]
  12. Yang, N. KNN Algorithm Simulation Based on Quantum Information. In Proceedings of the Student-Faculty Research Day Conference, CSIS, New York, NY, USA, 3 May 2019. [Google Scholar]
  13. Rebentrost, P.; Mohseni, M.; Lloyd, S. Quantum support vector machine for big data classification. Phys. Rev. Lett. 2014, 113, 130503. [Google Scholar] [CrossRef]
  14. Nakahara, M. Quantum Computing: From Linear Algebra to Physical Realizations; CRC Press: Boca Raton, FL, USA, 2008. [Google Scholar]
  15. Song, T.; Li, P.; Weng, J. Concise security bounds for sending-or-not-sending twin-field quantum key distribution with finite pulses. Phys. Rev. A 2021, 103, 042408. [Google Scholar] [CrossRef]
  16. Song, T.; Tan, X.; Weng, J. Statistical fluctuation analysis of measurement-device-independent quantum random-number generation. Phys. Rev. A 2019, 99, 022333. [Google Scholar] [CrossRef]
  17. Jiang, C.; Yu, Z.W.; Hu, X.L.; Wang, X.B. Unconditional security of sending or not sending twin-field quantum key distribution with finite pulses. Phys. Rev. Appl. 2019, 12, 024061. [Google Scholar] [CrossRef] [Green Version]
  18. Lloyd, S.; Mohseni, M.; Rebentrost, P. Quantum principal component analysis. Nat. Phys. 2014, 10, 631–633. [Google Scholar] [CrossRef] [Green Version]
  19. Kopczyk, D. Quantum machine learning for data scientists. arXiv 2018, arXiv:1804.10068. [Google Scholar]
  20. Hussain Shah, S.; Javed Iqbal, M.; Bakhsh, M.; Iqbal, A. Analysis of Different Clustering Algorithms for Accurate Knowledge Extraction from Popular DataSets. Inf. Sci. Lett. 2020, 9, 4. [Google Scholar]
  21. Lakshmi, S.A.; Mary, S.A.S.A. Group Mosquito Host Seeking Algorithm Based Self Organizing Technique for Genetic Algorithm. Appl. Math. Inf. Sci. 2019, 13, 231–238. [Google Scholar] [CrossRef]
  22. Mustafa, W. Shrink: An Efficient Construction Algorithm for Minimum Vertex Cover Problem. Inf. Sci. Lett. 2021, 10, 9. [Google Scholar]
  23. Arora, P.; Varshney, S. Analysis of k-means and k-medoids algorithm for big data. Procedia Comput. Sci. 2016, 78, 507–512. [Google Scholar] [CrossRef] [Green Version]
  24. Whelan, C.; Harrell, G.; Wang, J. Understanding the k-medians problem. In Proceedings of the International Conference on Scientific Computing (CSC); The Steering Committee of the World Congress in Computer Science, Computer Engineering and Applied Computing (WorldComp): Athens, Greece, 2015; p. 219. [Google Scholar]
  25. Likas, A.; Vlassis, N.; Verbeek, J.J. The global k-means clustering algorithm. Pattern Recognit. 2003, 36, 451–461. [Google Scholar] [CrossRef] [Green Version]
  26. Montanaro, A. Quantum algorithms: An overview. NPJ Quantum Inf. 2016, 2, 1–8. [Google Scholar] [CrossRef]
  27. Comaniciu, D.; Meer, P. Mean shift analysis and applications. In Proceedings of the Seventh IEEE International Conference on Computer Vision, Corfu, Greece, 20–25 September 1999; Volume 2, pp. 1197–1203. [Google Scholar]
  28. Ester, M.; Kriegel, H.P.; Sander, J.; Xu, X. A density-based algorithm for discovering clusters in large spatial databases with noise. Kdd 1996, 96, 226–231. [Google Scholar]
  29. He, X.; Cai, D.; Shao, Y.; Bao, H.; Han, J. Laplacian regularized gaussian mixture model for data clustering. IEEE Trans. Knowl. Data Eng. 2010, 23, 1406–1418. [Google Scholar] [CrossRef] [Green Version]
  30. Rodriguez, A.; Laio, A. Clustering by fast search and find of density peaks. Science 2014, 344, 1492–1496. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  31. Yu, S.S.; Chu, S.W.; Wang, C.M.; Chan, Y.K.; Chang, T.C. Two improved k-means algorithms. Appl. Soft Comput. 2018, 68, 747–755. [Google Scholar] [CrossRef]
  32. LIU, X.j.; YUAN, J.b.; XU, J.; DUAN, B.j. Quantum k-means algorithm. J. Jilin Univ. 2018, 2, 539–544. [Google Scholar]
  33. Zhang, G.; Zhang, C.; Zhang, H. Improved K-means algorithm based on density Canopy. Knowl.-Based Syst. 2018, 145, 289–297. [Google Scholar] [CrossRef]
  34. Pandit, S.; Gupta, S. A comparative study on distance measuring approaches for clustering. Int. J. Res. Comput. Sci. 2011, 2, 29–31. [Google Scholar] [CrossRef] [Green Version]
  35. Gultom, S.; Sriadhi, S.; Martiano, M.; Simarmata, J. Comparison analysis of K-means and K-medoid with Ecluidience distance algorithm, Chanberra distance, and Chebyshev distance for big data clustering. IOP Conf. Ser. Mater. Sci. Eng. 2018, 420, 012092. [Google Scholar] [CrossRef] [Green Version]
  36. Brassard, G.; Hoyer, P.; Mosca, M.; Tapp, A. Quantum amplitude amplification and estimation. Contemp. Math. 2002, 305, 53–74. [Google Scholar]
  37. Giovannetti, V.; Lloyd, S.; Maccone, L. Architectures for a quantum random access memory. Phys. Rev. A 2008, 78, 52310. [Google Scholar] [CrossRef] [Green Version]
  38. Zidan, M.; Aldulaimi, S.; Eleuch, H. Analysis of the Quantum Algorithm based on Entanglement Measure for Classifying Boolean Multivariate Function into Novel Hidden Classes: Revisited. Appl. Math 2021, 15, 643–647. [Google Scholar]
  39. Kerenidis, I.; Landman, J. Quantum spectral clustering. arXiv 2020, arXiv:2007.00280. [Google Scholar] [CrossRef]
  40. Kapil, S.; Chawla, M. Performance evaluation of K-means clustering algorithm with various distance metrics. In Proceedings of the 2016 IEEE 1st International Conference on Power Electronics, Intelligent Control and Energy Systems (ICPEICES), Delhi, India, 4–6 July 2016; pp. 1–4. [Google Scholar]
  41. Sahu, L.; Mohan, B.R. An improved K-means algorithm using modified cosine distance measure for document clustering using Mahout with Hadoop. In Proceedings of the 2014 9th International Conference on Industrial and Information Systems (ICIIS), Gwalior, India, 15–17 December 2014; pp. 1–5. [Google Scholar]
  42. Lee Rodgers, J.; Nicewander, W.A. Thirteen ways to look at the correlation coefficient. Am. Stat. 1988, 42, 59–66. [Google Scholar] [CrossRef]
  43. Aïmeur, E.; Brassard, G.; Gambs, S. Machine learning in a quantum world. In Conference of the Canadian Society for Computational Studies of Intelligence; Springer: Berlin/Heidelberg, Germany, 2006; pp. 431–442. [Google Scholar]
  44. Aïmeur, E.; Brassard, G.; Gambs, S. Quantum speed-up for unsupervised learning. Mach. Learn. 2013, 90, 261–287. [Google Scholar] [CrossRef] [Green Version]
  45. Kaye, P. Reversible addition circuit using one ancillary bit with application to quantum computing. arXiv 2004, arXiv:quant-ph/0408173. [Google Scholar]
  46. Buhrman, H.; Cleve, R.; Watrous, J.; De Wolf, R. Quantum fingerprinting. Phys. Rev. Lett. 2001, 87, 167902. [Google Scholar] [CrossRef] [Green Version]
  47. Durr, C.; Hoyer, P. A quantum algorithm for finding the minimum. arXiv 1996, arXiv:quant-ph/9607014. [Google Scholar]
  48. Ruan, Y.; Xue, X.; Liu, H.; Tan, J.; Li, X. Quantum algorithm for k-nearest neighbors classification based on the metric of hamming distance. Int. J. Theor. Phys. 2017, 56, 3496–3507. [Google Scholar] [CrossRef]
  49. Grover, L.K. A fast quantum mechanical algorithm for database search. In Proceedings of the Twenty-Eighth Annual ACM Symposium on Theory of Computing, Philadelphia, PA, USA, 22–24 May 1996. [Google Scholar]
  50. Boyer, M.; Brassard, G.; Høyer, P.; Tapp, A. Tight bounds on quantum searching. Fortschritte Phys. Prog. Phys. 1998, 46, 493–505. [Google Scholar] [CrossRef] [Green Version]
  51. Singh, A.; Yadav, A.; Rana, A. K-means with Three different Distance Metrics. Int. J. Comput. Appl. 2013, 67, 13–17. [Google Scholar] [CrossRef]
Figure 1. (a) Quantum S w a p T e s t circuit to obtain the similarity between two quantum states | ϕ and | ψ . (b) Details of quantum S w a p T e s t circuit.
Figure 1. (a) Quantum S w a p T e s t circuit to obtain the similarity between two quantum states | ϕ and | ψ . (b) Details of quantum S w a p T e s t circuit.
Entropy 24 00237 g001
Figure 2. Quantum D i s t C a l c circuit for calculating the distance between x i and x j .
Figure 2. Quantum D i s t C a l c circuit for calculating the distance between x i and x j .
Entropy 24 00237 g002
Figure 3. Overview of the quantum circuit for the QDPC algorithm, where | d i j represents | d ( x i , x j ) .
Figure 3. Overview of the quantum circuit for the QDPC algorithm, where | d i j represents | d ( x i , x j ) .
Entropy 24 00237 g003
Figure 4. Experimental results of two algorithms when the data are circularly distributed and the number N is fixed as 250. (a) Clustering performance of DPC algorithm; (b) clustering performance of QDPC algorithm.
Figure 4. Experimental results of two algorithms when the data are circularly distributed and the number N is fixed as 250. (a) Clustering performance of DPC algorithm; (b) clustering performance of QDPC algorithm.
Entropy 24 00237 g004
Table 1. Evaluation indicators of circularly distributed data.
Table 1. Evaluation indicators of circularly distributed data.
Seed = 21DPCQDPC
PurityF-ScoreARIPurityF-ScoreARI
N = 201.01.01.01.01.01.0
N = 401.01.01.01.01.01.0
N = 801.01.01.01.01.01.0
N = 2500.9880000.9761040.9523851.01.01.0
N = 5000.9920000.9840660.9681920.9960000.9920000.984032
N = 10000.9900000.9801640.9603600.9980000.9960000.992008
Table 2. Comparison of the complexity of the two algorithms in simulation experiments.
Table 2. Comparison of the complexity of the two algorithms in simulation experiments.
Algorithms N = 20 N = 40 N = 80 N = 250 N = 500 N = 1000
DPC(/s) 0.00500488 0.01701570 0.06606030 0.65659642 2.54531145 10.08315921
QDPC(/s) 0.02302074 0.03603339 0.06806207 0.19617867 0.39135551 0.77470422
Difference(/s) 0.01801586 0.0190176 0.00200177 0.46041775 2.15395594 9.30845499
Table 3. Theoretical comparison of the complexity of the two algorithms.
Table 3. Theoretical comparison of the complexity of the two algorithms.
ComplexityDPCQDPC
time O ( N 2 ) O ( log ( N 2 ) + 6 N + N )
space N · log N + N · n + log ( n + d ) 3 log N + n + log ( n + d ) + 2
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Wu, Z.; Song, T.; Zhang, Y. Quantum Density Peak Clustering Algorithm. Entropy 2022, 24, 237. https://doi.org/10.3390/e24020237

AMA Style

Wu Z, Song T, Zhang Y. Quantum Density Peak Clustering Algorithm. Entropy. 2022; 24(2):237. https://doi.org/10.3390/e24020237

Chicago/Turabian Style

Wu, Zhihao, Tingting Song, and Yanbing Zhang. 2022. "Quantum Density Peak Clustering Algorithm" Entropy 24, no. 2: 237. https://doi.org/10.3390/e24020237

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop