Abstract
Quantum machine learning has experienced significant progress in both software and hardware development in the recent years and has emerged as an applicable area of near-term quantum computers. In this work, we investigate the feasibility of utilizing quantum machine learning (QML) on real clinical datasets. We propose two QML algorithms for data classification on IBM quantum hardware: a quantum distance classifier (qDS) and a simplified quantum-kernel support vector machine (sqKSVM). We utilize these different methods using the linear time quantum data encoding technique (\({\mathrm{log}}_{2}N\)) for embedding classical data into quantum states and estimating the inner product on the 15-qubit IBMQ Melbourne quantum computer. We match the predictive performance of our QML approaches with prior QML methods and with their classical counterpart algorithms for three open-access clinical datasets. Our results imply that the qDS in small sample and feature count datasets outperforms kernel-based methods. In contrast, quantum kernel approaches outperform qDS in high sample and feature count datasets. We demonstrate that the \({\mathrm{log}}_{2}N\) encoding increases predictive performance with up toβ+β2% area under the receiver operator characteristics curve across all quantum machine learning approaches, thus, making it ideal for machine learning tasks executed in Noisy Intermediate Scale Quantum computers.
Similar content being viewed by others
Introduction
Quantum technologies promise to revolutionize the future of information and computation using quantum devices to process massive amounts of data. To date, considerable progress has been made from both software and hardware points of view. Many researches are underway to simplify quantum algorithms1,2,3,4,5,6,7,8 in order to implement them on existing, so-called Noisy Intermediate Scale Quantum (NISQ) computers9. As a result, small quantum devices based on photons, superconductors, or trapped ions are capable of efficiently running scalable quantum algorithms6,7,10. Quantum Machine Learning (QML) is a particularly interesting approach, as it is suited for existing NISQ architectures11,12,13,14,15. While conventional machine learning is generally applied to process large amounts of data, many research fields cannot provide such large datasets. One example is medical research, where collecting cohorts that represent certain characteristics of diseases routinely results in small datasets16. NISQ devices can efficiently execute algorithms with shallow depth and a low number of qubits9. Therefore, it appears logical to exploit the potential of QML executed on NISQ devices incorporating clinical datasets.
However, the execution of QML algorithms in the form of practical quantum gate operations is non-trivial. First, the classical data needs to be encoded into quantum states. For this purpose, prior QML algorithms assume that a quantum random access memory (QRAM) device for storing the data is present17. Nevertheless, to date, such practical devices are not available. Second, since the output of quantum algorithms are obviously quantum states, the efficient classical bits of information must be extracted through quantum measurements. To date, various classical data encoding approaches have been proposed6,7,18,19,20,21. In particular, encoding classical numerical features into quantum states has the advantage to utilize \({\mathrm{log}}_{2}N\) number of qubits (a.k.a. linear time encoding) in relation to \(N\) number of input features18,19,20,21. This approach allows to utilize NISQ devices with a small number of qubits and to minimize quantum noise, while at the same time maintaining quantum speedup14. In contrast, to date, this approach in combination with quantum machine learning appears to be underrepresented.
In light of the above proceedings, we hypothesize that clinically-relevant quantum prediction models can be built on NISQ devices employing the \({\mathrm{log}}_{2}N\) encoding, having prediction performances comparable to classic ML approaches.
In our work, we propose two quantum machine learning approaches that rely on the \({\mathrm{log}}_{2}N\) encoding approach, thus, not requiring the presence of a fault-tolerant quantum circuit for implementation of quantum RAM17. Previously proposed techniques for estimation of the inner product with Hadamard Test and Swap Test assume that there is a quantum RAM or a quantum circuit that store both index of data and their values22,23. To construct a quantum database (QDB) from classical data, \(\left(n+m+1\right)\)-qubits are required for \(M\) sample and \(N\)-feature counts, where \(n={\mathrm{log}}_{2}N\), \(m={\mathrm{log}}_{2}M\), and 1 is considered as qubit register24. In contrast, the \({\mathrm{log}}_{2}N\) encoding technique utilizes only \(n\) qubit and \(O\left(Mn\right)\) steps to classically access to data without allocating extra qubits to the index of entries of dataset. First, we demonstrate a simple and efficient quantum distance classifier (qDC) executable on existing NISQ devices. Second, we present a simplified quantum-kernel SVM (sqKSVM) approach using quantum kernels which can be executed once without optimization instead of twice with optimization as in case of the quantum-kernel SVM (qKSVM) approach6,7.
In order to test our hypothesis, we demonstrate the performance of the qDC and the sqKSVM approaches using real clinical data and compare their performances to qKSVM, as well as to classic computing counterparts such as k-nearest neighbors25 and classic support vector machines26.
Results
Dataset
This study incorporated three open-access clinical datasets that have been presented and evaluated in various contexts27,28,29. Each dataset underwent redundancy reduction by correlation matrix analysis30 followed by a tenfold cross-validation split with a training-validation ratio of 80β20%16. Training sets of the folds were subjects of feature ranking analysis31 and the highest-ranking eight as well as 16 (if available) features were selected for further analysis. The resulted dataset configurations were analyzed by class imbalance ratios and the quantum advantage score (a.k.a. difference geometry)20 for quantum kernel methods. Table 1 demonstrates the characteristics of the data configurations as well as the results of the imbalance ratio and the quantum advantage scores (for estimation of the quantum advantage scores \({(g}_{CQ})\), see Appendix E of the supplementary material).
Encoding strategies
This study relies on the data encoding strategy which uses sequences of Pauli-Y gate rotations (\({R}_{y}\)) and \(CNOT\) gates (see Appendix A of the supplementary material) to result in a number of \({\mathrm{log}}_{2}N\) encoding qubits18,19,21. \({R}_{y}\) puts each qubit \(q\) in a superposition state \({R}_{y}\left(2\theta \right)\left|q\rangle \right.=\mathrm{cos}\left(\theta \right)\left|0\rangle \pm \mathrm{sin}\left(\theta \right)\right.\left|1\rangle \right.\) and \(CNOT\) s entangle qubits. The data encoding feature map with the application of \({R}_{y}\) and \(CNOT\) is given by32
where \(\left|\varphi \left(\overrightarrow{\theta }\right)\rangle \right.={U}_{\varphi \left(\overrightarrow{\theta }\right)}{\left|0\rangle \right.}^{\otimes {\mathrm{log}}_{2}N}\). In Eq.Β (1), \(\varphi \left(\overrightarrow{\theta }\right)\) is the encoding map from the Euclidean space to the Hilbert space and \({U}_{\varphi \left(\overrightarrow{\theta }\right)}\) is the model circuit for data encoding, which maps \({\left|0\rangle \right.}^{\otimes {\mathrm{log}}_{2}N}\) to another ket vector of input data \(\left|\varphi \left(\overrightarrow{\theta }\right)\rangle \right.\). To find a relationship between the input data and \(\theta\), see Appendix A of the supplementary material.
In contrast, previously proposed quantum ML-specific encoding utilizes a block of the Hadamard gates followed by a block of Pauli-Z gate rotations (\({R}_{z}\)) are applied to each qubit7. To entangle the qubits, nearest neighbor \(CNOT\) s are also applied. The features of data samples are considered as angles of \({R}_{z}\) rotations and the required number of qubits for data encoding are equal to the number of features. The data encoding feature map with the application of the Hadamard, \({R}_{z}\) and \(CNOT\) is given by7
where \(\left|\varphi \left(\overrightarrow{x}\right)\rangle \right.={U}_{\varphi \left(\overrightarrow{x}\right)}{H}^{\otimes N}{\left|0\rangle \right.}^{\otimes N}\). \({U}_{\varphi \left(\overrightarrow{x}\right)}\) is the model circuit for N features data encoding.
In order to compare the predictive performance of the above two data encoding strategies, the qDC, the sqKSVM and the qKSVM (see Appendix C of the supplementary material) approaches were compared utilizing a number of \(N=8\) features. This analysis was executed using the Pennylane simulator environment33, while the sqKSVM was also evaluated on the IBMQ Melbourne machine (see βMethodsβ). Table 2 demonstrates the cross-validation area under the receiver operator characteristics curve (AUC) performance values of the quantum ML algorithms in relation to the \({\mathrm{log}}_{2}N\) and \(N\) encoding qubit strategies.
Quantum and classic machine learning predictive performance evaluation
The quantum distance classifier (qDC) first calculates the distance between the state vector of a test sample and each state vector of the training sample in set \(P\) and set \(Q\) and, then, assigns a label of the test sample to the label of the closest set. In the qDC, we divide the training set, with \(M\) number of samples, based on their labels \(\left\{a, b\in R\right\}\) into two subset \(\left\{P\right\}\) and \(\left\{Q\right\}\), where \(\left\{P\right\}\) contains only label \(a\) with the number of samples \(MP\) and \(\left\{Q\right\}\) contains only label \(b\) with the number of samples \(MQ\) with \(MP+MQ=M\). The task is to determine the label of the given test sample \(\left\{{y}_{k}\right\}\), if \({y}_{k}=a\) or \({y}_{k}=b\). Mathematically, if \(\left|v\rangle \right.\) is the state vector of the test sample as well as \(\left|u\rangle \right.\in P\) and \(\left|w\rangle \right.\in Q\), then the label of \(\left|v\rangle \right.\) is determined by \({y}_{k}=a\), if \(min\left(\left|\left|u\rangle \right.-\left|v\rangle \right.\right|\right) \le min\left(\left|\left|w\rangle \right.-\left|v\rangle \right.\right|\right)\), otherwise \({y}_{k}=b\). The distance between the vectors is given by8 i.e.
where \(\left|\bullet \right|\) is the norm \({l}_{2}\) of a vector. Therefore, the task is to calculate the inner product \(\langle u|v\rangle\) with a quantum computer.
The two different approaches to estimate \(\langle u|v\rangle\) with quantum computers are Hadamard Test22 and the Swap Test23. For the simplified quantum kernel SVM (sqKSVM), we first need to note that the standard form of the quantum kernelized binary classifiers is
where \(\tilde{y }\) is the unknown label, \({y}_{i}\) is the label of the \(i\) th training sample, \({{\alpha }_{i}}^{*}\) is the \(i\) th component of the support vector \({\overrightarrow{\alpha }}^{*}=\left({{\alpha }_{1}}^{*}, {{\alpha }_{2}}^{*},\dots , {{\alpha }_{M}}^{*}\right)\), \(M\) is the number of training data, and \(K\left({\overrightarrow{x}}_{i},\overrightarrow{\tilde{x }}\right)\) is the kernel matrix of all the training-test pairs.
For a given dataset \(D= {\left\{\left({\overrightarrow{x}}_{i}, {y}_{i}\right):{\overrightarrow{x}}_{i}\in {R}^{M}, {y}_{i}\in \left\{-1, 1\right\} \right\}}_{i=1,\dots ,M}\), one option to bypass the drawbacks of the qKSVM algorithm (see Appendix C of the supplementary material) as presented in6,7 is to set uniform weight \({{\alpha }_{i}}^{*}=1\), in case of \(IR=0.5\) (balanced dataset). Otherwise, \({{\alpha }_{i}}^{*}=IR\) for the majority class and \({{\alpha }_{j}}^{*}=1-IR\) for the minority class. Thresholding the value \(\sum_{i=1}^{M}{{{y}_{i}\alpha }_{i}}^{*}K\left({\overrightarrow{x}}_{i},\overrightarrow{\tilde{x }}\right)\) yields the binary output as following
In Eq.Β (5), \(K\left({\overrightarrow{x}}_{i},\overrightarrow{\tilde{x }}\right)\) is defined as (see Appendix F of the supplementary material)
The dataset configurations were utilized to estimate the performance of quantum and classic machine learning algorithms incorporated in this study. Performance estimation was done by confusion matrix analytics34. Prediction models were built based on the given training subset, followed by evaluating the respective validation subsets of each fold. Average area under the receiver operator characteristics curve (AUC) was calculated across validation cases for each predictive model. To build predictive models, quantum ML approaches included the qDC, the sqKSVM and the qKSVM (see Appendix C of supplementary material) were utilized. Classic machine learning approaches were k-nearest neighbors (ckNN)25 and support vector machines (cSVM)26. See Table 3 for the comparison of cross-validation AUC performances of quantum and classic computing algorithm within the dataset configurations.
Estimation of the probability of errors rate
Our experimental demonstrations are performed on the 15-qubit IBMQ Melbourne processor based on superconducting transmon qubits. The experiment has been conducted on the Wisconsin Breast Cancer dataset with 8 and 16 features, given, that this dataset provided the highest predictive cross-validation performance. On the NISQ device and simulator, each circuit is run with a fixed number of measurement shots (=β8192). We plot scatter diagrams for the inner product values from the simulator and the inner product from the NISQ device in Fig.Β 1. To show the correlation between the experimental and the simulator values of the inner products, we also fit optimal lines using least square regression in Fig.Β 1. To measure the difference between the inner products from the simulator and the inner product from the NISQ device, the root mean square error (RMSE) was calculated. The value of RMSE was 0.039 (3.9%) and 0.075 (7.5%) for 8 and 16 feature counts, respectively (Fig.Β 1). Therefore, the fidelities of the quantum circuits on the quantum cloud device were estimated 96% and 92.5% for the 8 and 16 feature counts, respectively. For more details of the experiment see Appendix G of the Supplementary material.
The depolarizing noise model represents a linear relationship between the ideal (simulator) and the noisy (experiment) values of the inner products based on Eq.Β (14) in βMethodsβ. Nevertheless, the slope of the fit lines in Fig.Β 1 show that the depolarizing noise model cannot estimate the true value of probability of error rate (\(\lambda\)). This is due to gate errors35 that are originated from miscalibration of quantum Hardware, not being covered by the depolarizing noise model.
Discussion
In this study, we aimed to investigate the effect of two encoding strategies in various quantum machine learning-built clinical prediction models. Next to prior quantum machine learning approaches, we also proposed two methods specifically designed for the \({\mathrm{log}}_{2}N\) encoding approach.
Our results demonstrate that the \({\mathrm{log}}_{2}N\) encoding in combination with low-complexity quantum machine learning approaches provides comparable or better results than the \(N\) encoding approach with previously-proposed quantum machine learning methods. This advantage was demonstrated not only in a simulator environment, but also utilizing NISQ devices. The low algorithmic quantum complexity also aims towards building prediction models that may be easier to interpret in the future, especially in light of the high complexity of classic machine learning approaches36. In contrast, it is important to emphasize, that the proposed quantum machine learning processes are also applicable in big data, given, that calculating the inner product of quantum states in NISQ devices can be done efficiently with the \({\mathrm{log}}_{2}N\) encoding approach21,22. The \({\mathrm{log}}_{2}N\) data encoding is also more robust against noise compared to the \(N\) data encoding, since it uses less number of noisy qubits of the NISQ device to estimate the inner product of quantum states10.
After encoding data from classical Euclidean space into quantum Hilbert space, the distance between data points may increase or decrease, which has implications in case of kernel methods20. The \({g}_{CQ}\) score can represent, whether distances between data points would increase or decrease after data encoding. For further explanations see Supplemental Appendix E.
When feature count increases, \({g}_{CQ}\) increases as well, because quantum state vectors of input features become closer due to the high dimensionality property of the Hilbert space. Higher feature count significantly influences performance in a positive way if \({g}_{CQ}\) isβ<β1 (e.g.β+β5β6% AUC in the Pediatric bone marrow dataset). It has been shown that classical ML models are competitive or outperform quantum ML approaches when \({g}_{CQ}\) is small20. Nevertheless, we demonstrated that when \({g}_{CQ}>1\), higher feature count does not contribute much to the performance increase (e.g. 1% difference in the Wisconsin breast cancer dataset). It is important to point out that a high \({g}_{CQ}\) (>β1) alone does not mean that the dataset is not ideal for kernel-based quantum machine learning. Specifically, the highest AUC of 0.93 was achieved in the 16 feature counts Wisconsin breast cancer dataset, while it also demonstrated the highest \({g}_{CQ}\), which also confirms prior findings20. In contrast, the same dataset in the classic SVM resulted in 0.89 AUC. We hypothesize that this phenomenon is due to the high sample count of the Wisconsin breast cancer dataset (Mβ=β569). In general, the imbalance ratio of the datasets did not appear to be correlated with predictive performance. The \({\mathrm{log}}_{2}N\) increased AUC with up to 2% compared to the \(N\) encoding when comparing the execution of the quantum machine learning approaches using simulation environment. This behavior was also identifiable with executions in NISQ devices, in case of kernel methods and the qDC. We hypothesize that lower AUC performance for the \(N\) encoding method in the simulator environment and NISQ device is due to higher number of qubits which likely lead to lower value of inner products. This is in line with the findings in20.
In general, the qKSVM demonstrated 2β5% higher AUC compared to the sqKSVM. The relative performance increase of the qKSVM was in relation to sample count and feature count. Specifically, the qKSVM showed an average 2% higher AUC with small sample count (Heart failure and Pediatric bone datasets), while it had 5% higher AUC in the Wisconsin breast cancer dataset. Nevertheless, both the qKSVM and the sqKSVM increased its AUC with double feature counts in the small Pediatric bone marrow dataset. This level of performance increase was not identifiable in the larger Wisconsin breast cancer dataset. Classic SVM demonstrated similar properties in relation to higher feature counts in small datasets20, while it was outperformed by the qKSVM in the large Wisconsin breast cancer dataset.
In conclusion, quantum SVM approaches benefit from higher feature count in general, where the qKSVMβdue to relying on optimizationβhas a particular benefit compared to the sqKSVM. In contrast, the sqKSVM algorithm reduces the time complexity of the qKSVM algorithm significantly, which may be advantageous in case of large datasets on NISQ devices. In the large Wisconsin breast cancer dataset, the qDC demonstrated higher performance compared to the sqKSVM, especially in small feature counts (0.91 AUC vs 0.87 AUC in the qDC and the sqKSVM respectively in 8 features). The qDC resulted in the highest AUC of 0.60 across all other quantum (0.50β0.51 AUC) and classic machine learning (0.53β0.58 AUC) approaches in the Heart failure dataset. We hypothesize that this is due to the distribution characteristics of the samples belonging to the two subclasses in the feature space, which challenges classification with kernel methods. Generally, the performance of the executed quantum and classic machine learning approaches are comparable within the collected cohorts (Table 3).
According to our findings, quantum distance approaches can provide high performance with small feature and sample counts, which is particularly ideal for NISQ devices. In contrast, quantum kernel methods appear to provide high performance with high feature and sample counts. We demonstrated that the \({\mathrm{log}}_{2}N\) encoding strategy allows to execute quantum ML algorithms for highly dimensional clinical datasets on low qubit count NISQ devices. In general, quantum machine learning benefits from utilizing the \({\mathrm{log}}_{2}N\) encoding strategy, as it increases predictive performance and reduces execution time in NISQ devices, while keeping model complexity lower. Our experiments also pointed out an important implication of how noise shall be estimated. As such, the depolarizing noise model cannot cover gate errors.
We consider our findings of high importance in relation to building future quantum ML prediction models in NISQ devices for clinically-relevant cohorts and beyond.
Methods
All experiments of this study were performed in accordance with the respective guidelines and regulations of the open-access data sources this study relied on. For details, see section βAccessβ.
Estimation of the inner product \(\langle {\varvec{u}}|{\varvec{v}}\rangle\) and \({\left|\langle {\varvec{u}}|{\varvec{v}}\rangle \right|}^{2}\)
FigureΒ 2 shows the quantum circuit for estimation of the real part of \(\langle u|v\rangle\) with the Hadamard Test.
To estimate the real part of \(\langle u|v\rangle\) on the quantum computer with the Hadamard Test, the training and test data needs to be prepared in a quantum state as
where \(\left|u\rangle \right.\) and \(\left|v\rangle \right.\) are the quantum states for the train and test datasets, respectively.
Then the Hadamard gate on the ancilla qubit interferences the training vector \(\left|u\rangle \right.\) with the test vector \(\left|v\rangle \right.\)
Finally, the measuring quantum state given in Eq.Β (8) in the computational basis \({\left|0\rangle \right.}_{a}\) gives probability as
where \(\mathrm{Pr}\) is the value of the probability of measurement on the \({\left|0\rangle \right.}_{a}\) state of Eq.Β (8) and \(\langle u|u\rangle =\langle v|v\rangle =1\). Since our datasets are real values \(Re\left(\langle u|v\rangle \right)=\langle u|v\rangle\). See Appendix H of the Supplementary material for details of the estimation of the inner product on the IBMQ Melbourne machine with the Hadamard Test.
The inner product \(\langle u|v\rangle\) can also be estimated on a quantum computer with the Swap Test (see Fig.Β 3). The Hadamard gate is applied on the ancilla qubit to create a superposition of \(\left|u\rangle \left|v\rangle \right.\right.\), i.e.
The application of the single-controlled swap gates on the state given in Eq.Β (10) entangles the ancilla qubit with \(\left|u\rangle \left|v\rangle \right.\right.\). The resulted entangled quantum state is \(\frac{1}{\sqrt{2}}\left({\left|0\rangle \right.}_{a}\left|u\rangle \left|v\rangle +{\left|1\rangle \right.}_{a}\left|v\rangle \left|u\rangle \right.\right.\right.\right.\right)\). Then, another Hadamard gate interferences the product state of the state vectors of the training and the test i.e.
Measuring the quantum state given in Eq.Β (11) in the computational basis yields \({\left|0\rangle \right.}_{a}\) with the probability
where \(\mathrm{Pr}\) is the value of the probability of measurement on the \({\left|0\rangle \right.}_{a}\) state of Eq.Β (11). See Appendix D of the Supplementary material for details of the estimation of the inner product on the IBMQ Melbourne machine with the Hadamard Test.
Simplified quantum kernel support vector machine
The quantum Support Vector Machine algorithm is proposed in37 for big data classification. They show exponential speedup for their algorithm via quantum mechanically access to data. Nevertheless, this approach is not ideal for NISQ devices9. To date, two separate qKSVM approaches are proposed for data classification via classical access to data6,7. In these approaches, the quantum circuits must run twice on the quantum computer and a cost function needs to be optimized on the classical computer to compute the support vector7. We propose a simplified version qKSVM called sqKSVM as shown in Fig.Β 4.
Software and hardware
For classical machine learning algorithms, we use classical machine learning (CML) libraries of scikit-learn38. Pennylane-Qiskit33 is used for quantum circuit simulation and quantum experiment for designing quantum computing programs. Pennylane-Qiskit 0.13.0 plugin integrates the Qiskit quantum computing framework to the Pennylane simulator.
For executing quantum algorithms on existing quantum computers, this study relied on IBMβs remote quantum machines (https://quantum-computing.ibm.com/) that can run quantum programs with noisy qubits. Since IBM quantum computers only support single-qubit gate and two-qubit \(CNOT\) gate operations, complex gate operations must be decomposed into elementary supported gates before mapping the quantum circuit on noisy hardware. Owing to the specific architecture of IBM quantum computers, all two-qubit \(CNOT\) gate operations must satisfy the constraints imposed by the coupling map39, i.e., if \({q}_{i}\) is the control qubit and \({q}_{j}\) is the target qubit, \(CNOT({q}_{i}, {q}_{j})\) can only be applied if there is coupling between \({q}_{i}\) and \({q}_{j}\). In case of running the QML algorithms on the quantum computer, we choose the 15-qubits IBMQ Melbourne machine with the supported gates \(I\), \({U}_{3}\), and \(CNOT\), where \(I\) is identity single-qubit gate, \({U}_{3}\) is single-qubit arbitrary rotation gates with \(CNOT\) as two-qubit gate. FigureΒ 5 shows the coupling map of the IBMQ Melbourne with its gate error rates.
Depolarizing noise model
A simple model to describe incoherent noise is the depolarizing noise model. For a \(n\) qubit pure quantum state \(\left|u\rangle \right.\), the depolarizing noise operator (channel) leads to a loss of information with probability \(\lambda\) and with probability \(\left(1-\lambda \right)\) the system is left untouched40. The state of the system after this noise is
where \({\epsilon }_{\lambda }\) denotes the noise channel, \(\rho =\left|u\rangle \right.\langle \left.u\right|\) is a density matrix, \(\lambda\) is the probability of error rate that depends on NISQ devices, the gate operations, and the depth of the quantum circuit, and \(I\) is the (\({2}^{n}\times {2}^{n}\)) identity matrix.
The expectation value of observable \(O\) for a state represented by a density matrix \(\rho\) is given by
where \(\overline{\langle O\rangle }\) is the noisy expectation value and \(\langle O\rangle\) is the noiseless expectation value41.
Data availability
The source code of the implemented quantum algorithms can be accessed by the link: https://github.com/sassan72/Quantum-Machine-learning. For classical machine learning executions, this study relied on the scikit-learn library38: https://scikit-learn.org/stable/. All included open-access datasets are accessible through the following links: Pediatric bone marrow transplant dataset27: https://archive.ics.uci.edu/ml/datasets/Bone+marrow+transplant%3A+children. Wisconsin breast cancer dataset28: https://archive.ics.uci.edu/ml/datasets/Breast+Cancer+Wisconsin+(Diagnostic). Heart failure dataset29: https://www.kaggle.com/andrewmvd/heart-failure-clinical-data.
References
Suzuki, Y. et al. Amplitude estimation without phase estimation. Quant. Inf. Process. 19, 75 (2020).
Tanaka, T. et al. Amplitude estimation via maximum likelihood on noisy quantum computer. Quant. Inf. Process. 20, 293 (2021).
Grinko, D., Gacon, J., Zoufal, C. & Woerner, S. Iterative quantum amplitude estimation. NPJ Quant. Inf. 7, 52 (2021).
Aaronson, S. & Rall, P. Quantum approximate counting. Simplified. https://doi.org/10.1137/1.9781611976014.5 (2019).
Bouland, A., van Dam, W., Joorati, H., Kerenidis, I. & Prakash, A. Prospects and challenges of quantum finance. (2020).
Schuld, M. & Killoran, N. Quantum machine learning in feature Hilbert spaces. Phys. Rev. Lett. 122, 040504 (2019).
HavlΓΔek, V. et al. Supervised learning with quantum-enhanced feature spaces. Nature 567, 209β212 (2019).
Wiebe, N., Kapoor, A. & Svore, K. Quantum algorithms for nearest-neighbor methods for supervised and unsupervised learning. (2014).
Preskill, J. Quantum computing in the NISQ era and beyond. Quantum 2, 79 (2018).
Johri, S. et al. Nearest centroid classification on a trapped ion quantum computer. NPJ Quant. Inf. 7, 122 (2021).
Saggio, V. et al. Experimental quantum speed-up in reinforcement learning agents. Nature 591, 229β233 (2021).
Steinbrecher, G. R., Olson, J. P., Englund, D. & Carolan, J. Quantum optical neural networks. NPJ Quant. Inf. 5, 60 (2019).
Peters, E. et al. Machine learning of high dimensional data on a noisy quantum processor. (2021).
Liu, Y., Arunachalam, S. & Temme, K. A rigorous and robust quantum speed-up in supervised machine learning. Nat. Phys. 17, 1013β1017 (2021).
Hubregtsen, T. et al. Training Quantum Embedding Kernels on Near-Term Quantum Computers. (2021).
Papp, L., Spielvogel, C. P., Rausch, I., Hacker, M. & Beyer, T. Personalizing medicine through hybrid imaging and medical big data analysis. Front. Phys. 6, 2 (2018).
Giovannetti, V., Lloyd, S. & Maccone, L. Architectures for a quantum random access memory. Phys. Rev. A 78, 052310 (2008).
Schuld, M. & Petruccione, F. Supervised Learning with Quantum Computers (Springer International Publishing, 2018). https://doi.org/10.1007/978-3-319-96424-9.
MΓΆttΓΆnen, M., Vartiainen, J. J., Bergholm, V. & Salomaa, M. M. Quantum circuits for general multiqubit gates. Phys. Rev. Lett. 93, 130502 (2004).
Huang, H.-Y. et al. Power of data in quantum machine learning. Nat. Commun. 12, 2631 (2021).
Schuld, M., Bocharov, A., Svore, K. M. & Wiebe, N. Circuit-centric quantum classifiers. Phys. Rev. A 101, 032308 (2020).
Schuld, M., Fingerhuth, M. & Petruccione, F. Implementing a distance-based classifier with a quantum interference circuit. EPL Europhys. Lett. 119, 60002 (2017).
Blank, C., Park, D. K., Rhee, J.-K.K. & Petruccione, F. Quantum classifier with tailored quantum kernel. NPJ Quant. Inf. 6, 41 (2020).
Park, D. K., Petruccione, F. & Rhee, J.-K.K. Circuit-based quantum random access memory for classical data. Sci. Rep. 9, 3949 (2019).
Zhang, Z. Introduction to machine learning: k-nearest neighbors. Ann. Transl. Med. 4, 218β218 (2016).
Chang, C.-C. & Lin, C.-J. LIBSVM. ACM Trans. Intell. Syst. Technol. 2, 1β27 (2011).
Sikora, M., WrΓ³bel, Ε & GudyΕ, A. GuideR: A guided separate-and-conquer rule learning in classification, regression, and survival settings. Knowl.-Based Syst. 173, 1β14 (2019).
Dua, D. & Graff, C. {UCI} Machine Learning Repository. University of California, Irvine, School of Information and Computer Sciences http://archive.ics.uci.edu/ml (2017).
Chicco, D. & Jurman, G. Machine learning can predict survival of patients with heart failure from serum creatinine and ejection fraction alone. BMC Med. Inform. Decis. Mak. 20, 16 (2020).
Krajnc, D. et al. Breast tumor characterization using [18F]FDG-PET/CT imaging combined with data preprocessing and radiomics. Cancers 13, 1249 (2021).
Papp, L. et al. Supervised machine learning enables non-invasive lesion characterization in primary prostate cancer with [68Ga]Ga-PSMA-11 PET/MRI. Eur. J. Nucl. Med. Mol. Imaging https://doi.org/10.1007/s00259-020-05140-y (2020).
Schuld, M. Supervised quantum machine learning models are kernel methods. (2021).
Bergholm, V. et al. PennyLane: Automatic differentiation of hybrid quantum-classical computations. (2018).
Luque, A., Carrasco, A., MartΓn, A. & de las Heras, A.,. The impact of class imbalance in classification performance metrics based on the binary confusion matrix. Pattern Recognit. 91, 216β231 (2019).
Leymann, F. & Barzen, J. The bitter truth about gate-based quantum algorithms in the NISQ era. Quantum Sci. Technol. 5, 044007 (2020).
Carvalho, D. V., Pereira, E. M. & Cardoso, J. S. Machine learning interpretability: A survey on methods and metrics. Electronics 8, 832 (2019).
Rebentrost, P., Mohseni, M. & Lloyd, S. Quantum support vector machine for big data classification. Phys. Rev. Lett. 113, 130503 (2014).
Pedregosa, F. et al. Scikit-learn: Machine Learning in Python. (2012).
Wille, R., Hillmich, S. & Burgholzer, L. Efficient and Correct Compilation of Quantum Circuits. in 2020 IEEE International Symposium on Circuits and Systems (ISCAS) 1β5 (IEEE, 2020). https://doi.org/10.1109/ISCAS45731.2020.9180791.
Nielsen, M. A. & Chuang, I. L. Quantum Computation and Quantum Information (Cambridge University Press, 2010). https://doi.org/10.1017/CBO9780511976667.
Urbanek, M. et al. Mitigating depolarizing noise on quantum computers with noise-estimation circuits. (2021).
Funding
This work was funded by Medizinische UniversitΓ€t Wien, FocusXL-2019Β grant βFoundations of a Quantum Computing Lab at the CMPBMEβ.
Author information
Authors and Affiliations
Contributions
All co-authors of this paper were involved in the conceptualization of the study, in writing, in the critical review as well as final approval to submit this paper. Specific author contributions are as follows: S.M.: Quantum machine learning design, implementation and execution. C.B.: Development environment support, algorithm design. C.S.: Classic machine learning comparative analysis. D.K.: Data preparation, cross-validation scheme, performance evaluation. S.H., R.W.: Simulator environment establishment, support in algorithm optimization and execution. W.D., L.P.: Study design, supervision.
Corresponding author
Ethics declarations
Competing interests
The authors declare no competing interests.
Additional information
Publisher's note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary Information
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Moradi, S., Brandner, C., Spielvogel, C. et al. Clinical data classification with noisy intermediate scale quantum computers. Sci Rep 12, 1851 (2022). https://doi.org/10.1038/s41598-022-05971-9
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41598-022-05971-9
This article is cited by
-
Quantum Machine-Based Decision Support System for the Detection of Schizophrenia from EEG Records
Journal of Medical Systems (2024)
-
Advances of machine learning in materials science: Ideas and techniques
Frontiers of Physics (2024)
-
A machine learning model based on MRI for the preoperative prediction of bladder cancer invasion depth
European Radiology (2023)
-
Error mitigation enables PET radiomic cancer characterization on quantum computers
European Journal of Nuclear Medicine and Molecular Imaging (2023)
Comments
By submitting a comment you agree to abide by our Terms and Community Guidelines. If you find something abusive or that does not comply with our terms or guidelines please flag it as inappropriate.