Next Article in Journal
A Quantum Model of Trust Calibration in Human–AI Interactions
Next Article in Special Issue
Security of the Decoy-State BB84 Protocol with Imperfect State Preparation
Previous Article in Journal
Community-Based Matrix Factorization (CBMF) Approach for Enhancing Quality of Recommendations
Previous Article in Special Issue
Proposal for Trapped-Ion Quantum Memristor
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Quantum and Quantum-Inspired Stereographic K Nearest-Neighbour Clustering

1
Theoretical Quantum System Design Group, Chair of Theoretical Information Technology, Technical University of Munich, 80333 Munich, Germany
2
Institute for Communications Engineering, TUM School of Computation, Information and Technology, Technical University of Munich, 80333 Munich, Germany
3
Optical and Quantum Laboratory, Munich Research Center, Huawei Technologies Düsseldorf GmbH, Riesstr. 25-C3, 80992 Munich, Germany
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
Entropy 2023, 25(9), 1361; https://doi.org/10.3390/e25091361
Submission received: 1 May 2023 / Revised: 12 September 2023 / Accepted: 13 September 2023 / Published: 20 September 2023

Abstract

:
Nearest-neighbour clustering is a simple yet powerful machine learning algorithm that finds natural application in the decoding of signals in classical optical-fibre communication systems. Quantum k-means clustering promises a speed-up over the classical k-means algorithm; however, it has been shown to not currently provide this speed-up for decoding optical-fibre signals due to the embedding of classical data, which introduces inaccuracies and slowdowns. Although still not achieving an exponential speed-up for NISQ implementations, this work proposes the generalised inverse stereographic projection as an improved embedding into the Bloch sphere for quantum distance estimation in k-nearest-neighbour clustering, which allows us to get closer to the classical performance. We also use the generalised inverse stereographic projection to develop an analogous classical clustering algorithm and benchmark its accuracy, runtime and convergence for decoding real-world experimental optical-fibre communication data. This proposed ‘quantum-inspired’ algorithm provides an improvement in both the accuracy and convergence rate with respect to the k-means algorithm. Hence, this work presents two main contributions. Firstly, we propose the general inverse stereographic projection into the Bloch sphere as a better embedding for quantum machine learning algorithms; here, we use the problem of clustering quadrature amplitude modulated optical-fibre signals as an example. Secondly, as a purely classical contribution inspired by the first contribution, we propose and benchmark the use of the general inverse stereographic projection and spherical centroid for clustering optical-fibre signals, showing that optimizing the radius yields a consistent improvement in accuracy and convergence rate.

1. Introduction

Quantum Machine Learning (QML), using quantum algorithms to learn quantum or classical systems, has attracted much research in recent years, with some algorithms possibly gaining an exponential speedup [1,2,3]. Since machine learning routines often push real-world limits of computing power, an exponential improvement to algorithm speed would allow for such systems with vastly greater capabilities [4]. Google’s ‘Quantum Supremacy’ experiment [5] showed that quantum computers can naturally solve specific problems with complex correlations between inputs that can be incredibly hard for traditional (“classical”) computers. Such a result suggests that machine learning models executed on quantum computers could be more effective for specific applications. It seems quite possible that quantum computing could lead to faster computation, better generalisation on less data, or both, for an appropriately designed learning model. Hence, it is of great interest to discover and model the scenarios in which such a “quantum advantage” could be achieved. A number of such “Quantum Machine Learning” algorithms are detailed in papers such as [2,6,7,8,9]. Many of these methods claim to offer exponential speedups over analogous classical algorithms. However, some significant gaps exist between theoretical prediction and implementation on the path from theory to technology. These gaps result in unforeseen technological hurdles and sometimes misconceptions, necessitating more careful case-by-case studies such as [10].
It is known from the literature that the k-nearest-neighbour clustering algorithm (kNN) can be applied to solve the problem of phase estimation in optical fibres [11,12]. A quantum version of this kNN has been developed in [2], promising an exponential speedup. However, the practical usefulness of this algorithm is under debate [4]. The encoding of classical data into quantum states has been proven to be a complex task which significantly reduces the advantage of known quantum machine learning algorithms [4]. There are claims that the speedup is reduced to only polynomial once the quantum version of the algorithm takes into account the time taken to prepare the necessary quantum states. Furthermore, for near-intermediate scale quantum (NISQ) [3] applications, we should not expect the availability of QRAM, as this assumes reliable memories and operations which are still several milestones out of reach [13]. For this reason, it is not currently possible to use the fully quantum clustering algorithm and thus we resort to using hybrid quantum-classical kNN algorithms. Any classical implementation of kNN clustering involves, among other steps, repeated evaluations of a dissimilarity and and a loss function; changing the dissimilarity leads to a different clustering. A hybrid quantum classical kNN clustering algorithm utilizes quantum methods only to estimate the dissimilarity, eliminating the need for long-lasting quantum memories. However, reproducing the dissimilarity of a classical kNN algorithm using quantum methods can be prohibitively restrictive. The quantum dissimilarity also depends on the embedding (how the classical data are encoded in quantum states) and might only approximate the classical one, introducing fundamental deviations from the classical kNN algorithm. In [10], we applied a hybrid quantum-classical algorithm with modified angle embedding to the problem of k-means clustering for 64-QAM (Quadrature Amplitude Modulation) optical-fibre data (a well-known technical problem in signal processing through optical-fibre communication links) provided by Huawei [14], and show that this does not currently yield an advantage due to both the embedding and the current speed and noise of quantum devices.
In this work, we use the same problem and datasets to bring two main but independent contributions using the generalised inverse stereographic projection. First, we embed classical 2-dimensional data by computing the ISP onto the 3-dimensional sphere, and use the resulting normalised vector as the Bloch vector to produce a pure quantum state of one qubit, which we call stereographic embedding. The resulting quantum dissimilarity directly translates into the cosine dissimilarity, thus making the quantum algorithm mathematically closer to the classical k-means algorithm. This means that no inherent limitation is introduced by the embedding and any loss in performance of this hybrid algorithm can be compensated for by improving the noise level and the speed of the quantum device. We thus propose stereographic embedding as an improved quantum embedding that may lead to improvement in several quantum machine learning algorithms (although there might not still be a practical quantum time advantage).
The second contribution comes from the benchmarking of the hybrid stereographic quantum mentioned above. Since, as already mentioned, the resulting hybrid clustering algorithm is mathematically equivalent to a classical ‘quantum-inspired’ kNN algorithm, in order to assess its performance in the absence of noise, we simply test the equivalent classical quantum-inspired kNN algorithm. This algorithm is the result of first computing the ISP of the data and then performing clustering using a novel ‘quantum’ centroid update. We observe an increase in accuracy and convergence performance over k-means clustering on the 2-dimensional optical-fibre data. This suggests, as a purely classical second main contribution, that an advantage in decoding 64-QAM optical-fibre data is achieved by performing clustering in the inverse stereographically projected sphere and by using the spherical centroid.
This paper is structured as follows. In the remainder of this introduction, we discuss related works and our contribution to it. In Section 2, we introduce the experimental setup generating the 64-QAM optical-fibre transmission data and define clustering, the stereographic projection and the necessary quantum concepts for the hybrid protocols. Next, Section 3 introduces the developed Stereographic Quantum kNN (SQ-kNN), while Section 4 defines the developed quantum-inspired 2D Stereographic Classical kNN (2DSC-kNN) algorithm and proves its equivalence to the SQ-kNN quantum algorithm. In Section 5, we describe the various experiments for testing the algorithms, present the obtained results, and discuss their conclusions. We conclude the main text in Section 6, proposing some directions for future research, some of which are further discussed in Appendix D.

1.1. Related Work

A unifying overview of several quantum algorithms is presented in [15] in a tutorial style. An overview targeting data scientists is provided in [16]. The idea of using quantum information processing methods to obtain speedups for the k-means algorithm was proposed in [17]. In general, neither the best nor even the fastest method for a given problem and problem size can be uniquely ascribed to either the class of quantum or classical algorithms, as observed in the detailed discussion presented in [9]. The advantages of using local (classical) processing units alongside quantum processing units in a distributed fashion are discussed in [18]. The accuracy of (quantum) k-means has been demonstrated experimentally in [19] and in [20], while quantum circuits for loading classical data into a quantum computer are described in [21].
An algorithm is proposed in [2] that solves the problem of clustering N-dimensional vectors to M clusters in O ( log ( M N ) ) time on a quantum computer, which is exponentially faster than the O ( poly ( M N ) ) time for the (then) best known classical algorithm. The approach detailed in [2] requires querying the QRAM [22] for preparing a ‘mean state’, which is then used to find the inner product between the centroid (by default, the mean point) using the SWAP test [23,24,25]. However, there exist some significant caveats to this approach. Firstly, this algorithm achieves an exponential speedup only when comparing the bit-to-bit processing time with the qubit-to-qubit processing time. If one compares the bit-to-bit execution times of both algorithms, the exponential speedup disappears, as shown in [4,26]. Secondly, since stable enough quantum memories do not exist, a hybrid quantum-classical approach must be used in real-world applications. Namely, all the information must be stored in classical memories, and the states to be used in the algorithm are prepared in real time. The process of preparing quantum states from classical data is known as ‘Data Embedding’ since we are embedding the classical data into quantum states. This, as mentioned before [4,26], slows down the algorithm to only a polynomial advantage over classical k-means. However, we propose an approach whereby this step of embedding can be treated as a data pre-processing step, allowing us to achieve some advantages in accuracy and convergence rate, and taking a step towards making the quantum approach more viable. Instead of using a quantum algorithm, classical alternatives mimicking their behaviour, collectively known as quantum-inspired algorithms, have shown much promise in classically achieving some types of advantage that are demonstrated by quantum algorithms [4,26,27,28], but as [9] remarks, the massive increase in runtime with rank, condition number, Frobenius norm, and error threshold make the algorithms proposed in [4,26] impractical for matrices arising from real-world applications. This observation is supported by [29].
Recent works such as [26] suggest that even the best QML algorithms, without state preparation assumptions, fail to achieve exponential speedups over their classical counterparts. In [4], it is pointed out that most QML algorithms are incomparable to classical algorithms since they take quantum states as input and output quantum states, and that there is no analogous classical model of computation where one could search for similar classical algorithms. In [4], the idea of matching state preparation assumptions with 2 -norm sampling assumptions (first proposed in [26]) is implemented by introducing a new input model, sample and query access (SQ access). In [4], the Quantum k-means algorithm described in [2] is ‘de-quantised’ using the ‘toolkit’ developed in [26], i.e., a classical quantum-inspired algorithm is provided that, with classical SQ access assumptions replacing quantum state preparation assumptions, matches the bounds and runtime of the corresponding quantum algorithm up to the polynomial slowdown. From the works [4,26,30], we can conclude that the exponential speedups of many quantum machine learning algorithms that are under consideration arise not from the ‘quantumness’ of the algorithms but instead from strong input assumptions, since the exponential part of the speedups vanish when classical algorithms are provided analogous assumptions. In other words, in a wide array of settings, these algorithms do not provide exponential speedups but rather yield polynomial speedups on classical data.
The fundamental aspect that allowed for the exponential speedup in [26] is exemplified by the problem of recommendation systems. The philosophy of classical recommendation algorithms before this breakthrough was to estimate all the possible preferences of a user and then suggest one or more of the most preferred objects. A quantum algorithm in [8] promised an exponential speedup but provided a recommendation without estimating all the preferences; namely, it only provided a sample of the most preferred objects. This process of sampling, along with state preparation assumptions, was, in fact, what gave the quantum algorithm an exponential advantage. The new classical algorithm also obtains comparable speedups by only providing samples rather than solving the whole preference problem. In [4], it is argued that the time taken to create the quantum state should be included for comparison since the time taken is not insignificant; it is also claimed that for every such linear algebraic quantum machine learning algorithm, a polynomially slower classical algorithm can be constructed by using the binary tree data structure described in [26]. Since then, more sampling algorithms have shown that multiple quantum exponential speedups are not due to the quantum algorithms themselves but due to the way data are provided to the algorithms and how the quantum algorithm provides the solutions [4,29,30,31]. Notably, in [31], it is argued that there exist competing classical algorithms for all linear algebraic subroutines and thus for many quantum machine learning algorithms. However, as pointed out in [9] and proven in [29], significant caveats exist to these aforementioned results of quantum-inspired algorithms. The polynomial factor in these algorithms often contains a very high power of the rank and condition number, making them suitable only for sparse low-rank matrices. Matrices of real-world data are often relatively high in rank and hence unfavourable for such sampling-based quantum-inspired approaches. Whether such sampling algorithms can be used also highly depends on the specific application and whether or not samples of the solution instead of the complete data are suitable. It should be pointed out that quantum machine learning algorithms generally do not provide an advantage if such complete data are needed.
The method of encoding classical data into quantum states contributes to the complexity and performance of the algorithm. An extensive analysis and testing of the hybrid quantum-classical implementation of the quantum k-means algorithm using angle embedding can be found in [10]. In this work, the use of the ISP is proposed. Others have explored this procedure [32,33,34] as well; however, the motivation, implementation, and use vary significantly, as well as the procedure for embedding data points into quantum states. There has also been no extensive testing of the proposed methods, especially not in an industry context. In our method, we exclusively use pure states from the Bloch sphere since this reduces the complexity of the application. Lemma 3 assures us that our method with existing quantum techniques is applicable for nearest neighbour clustering. In contrast, the density matrices of mixed states and the normalised trace distance between the density matrices are used for the binary classification in [32,33]. A crucial thing to consider here is to distinguish the contribution of the ISP from the quantum effects. We will see in Section 5 that the ISP seems to be the most important contributing factor. In [35], it is also proposed to encode classical information into quantum states using the ISP in the context of quantum generative adversarial networks. Their motivation for using the ISP is due to the fact that it is injective and can hence be used to uniquely represent every point in the 2D plane without any loss of information. On the other hand, angle embedding loses all amplitude information due to the normalisation of all points. A method to transform an unknown manifold into an n-sphere using ISP is proposed in [36]—here, however, the property of their concern was the conformality of the projection since subsequent learning is performed upon the surface. In [37], a parallelised version of [2] is developed using the FF-QRAM procedure [38] for amplitude encoding and the ISP to ensure a injective embedding.
In the method of Spherical Clustering [39], the nearest neighbour algorithm is explored based on the cosine similarity measure (Equation (21) and Lemma 2). The cosine similarity is used in cases of information retrieval, text mining, and data mining to find the similarity between document vectors. It is used in those cases because the cosine similarity has low complexity for sparse vectors since only the non-zero co-ordinates need to be considered. For our case as well, it is in our interest to study Equations (16)–(18) with the cosine dissimilarity. This approach becomes particularly relevant once we employ stereographic embedding to encode the data points into quantum states.

1.2. Contribution

In this work, we first develop generalised stereographic embedding for hybrid quantum-classical kNN clustering as a better encoding that allows the quantum algorithm (Section 3) to outperform the accuracy and convergence of classical k-means algorithm in the absence of noise; in contrast, angle embedding introduces fundamental limitations to the accuracy not due to quantum noise. To validate this statement, we simulate this algorithm classically, which translates into an equivalent classical quantum-analogous stereographic kNN clustering algorithm (Section 4). One must note that we do not demonstrate that running the stereographic quantum kNN algorithm is more practical than the classical k-means algorithm in the NISQ context. We show that stereographic quantum kNN clustering converges faster and is more accurate than other hybrid quantum-classical kNN algorithms with angle or amplitude embedding. In parallel, the benchmarking of the classical stereographic kNN algorithm lets us claim that for the problem of decoding 64-QAM optical-fibre signals, the generalised ISP and spherical centroid can allow for better accuracy and convergence.
The extensive testing upon the real-world, experimental QAM dataset (Section 2.1) revealed some significant results regarding the dependence of accuracy, runtime, and convergence performance upon the radius of projection, number of points, noise in the optical-fibre, and stopping criterion—described in Section 5. Noteworthy, we observe the existence of a finite optimal radius for the ISP (not equal to 1). To the best of our knowledge, no other work has considered a generalised projection radius for quantum embedding or studied its effect. Through our experimentation, we have verified that there exists an ideal radius greater than 1 for which accuracy performance is maximised. The advantageous implementation of the algorithm upon experimental data shows that our procedure is quite competitive. The fact that the developed quantum algorithm has an entirely classical analogue (with comparable time complexity to the classical k-means algorithm) is a distinct advantage in terms of in-field deployment, especially compared to [2,9,17,32,33,34,37]. The developed quantum algorithm also has another advantage in the context of Noisy Intermediate-Scale Quantum (NISQ) realisations—it has the least circuit depth and circuit width among all candidates [2,9,34,37]—making it easier to implement with the current quantum technologies. Another significant contribution is our generalisation of the dissimilarity for clustering; instead of Euclidean dissimilarity (distance), we consider other dissimilarities which might be better estimated by quantum circuits (Appendix E). A somewhat similar approach was developed in parallel by [40] in the context of amplitude embedding. All previous approaches [2,9,34,37] only try to estimate the Euclidean distance. We also make the contribution of studying the relative effect of ‘quantumness’ and the ISP, something completely overlooked in previous works. We show that the quantum ‘advantage’ in accuracy performance touted by works such as [32,33,34,37] is in reality quite suspect and achievable through classical means. In Appendix D, we describe a generalisation of the stereographic embedding—the Ellipsoidal embedding, which we expect to provide even better results in future works.
Other secondary contributions of our work include:
  • The development of a mathematical formalism for the generalisation of kNN to indicate the contribution of various parameters such as dissimilarities and dataspace (Section 2.4);
  • Presenting the procedure and circuit for stereographic embedding using the Bloch embedding procedure, which consumes only O ( 1 ) in time and resources (Section 3.1).

2. Preliminaries

In this section, for completeness, we touch upon some concepts and background information required to understand this paper. These concepts range from small general statements on quantum states (Bloch sphere, fidelity, and Bell-state measurement in Section 2.2 and Section 2.3) to the mathematical formalism of kNN (Section 2.4, Section 2.5 and Section 2.6), and stereographic projection (Section 2.7). We begin by first describing the optic-fibre experimental setup used to collect the 64-QAM dataset, upon which the clustering algorithms were tested and benchmarked.

2.1. Optical-Fibre Setup

M-ary Quadrature Amplitude Modulation (M-QAM) is a simple and popular protocol for digital data transmission through analog communication channels. It is widely used in optical-fibre communication networks, and the decoding process of the received data often uses the k-nearest-neighbour algorithm to cluster nearby points. More details, including the description of the model used in the experiments, can be found in Appendix A. We now describe the experimental setup used to collect the dataset that is used for benchmarking the clustering algorithms.
The dataset contains a launch-power (laser-power feed into the fibre) sweep (four datasets collected at four different launch powers) of 80 km fibre transmission of coherent dual polarization (DP)-64QAM with a gross data rate of 960 × 10 9 bits/s. For the dataset, we assumed 15% overhead for forward error correction (FEC) and used 3.47% overhead for pilots and training sequences; thus, the net bit rate is 800 × 10 9 bits/s. Note that the pilots and training sequences are removed after the MIMO equalizer. An overview of the experimental setup [10,14] to capture this real-world database is shown in Figure 1. Four 120 × 10 9  Samples/s digital-to-analog converters (DACs) generate an electrical signal amplified by four 60 GHz 3dB-Bandwidth amplifiers. A tunable 100 kHz external cavity laser (ECL) source generates a continuous wave signal that is modulated by a 32 GHz DP-I/Q modulator. The receiver comprises an optical 90 -hybrid and four 100 GHz balanced photodiodes. The electrical signals are digitized using four 10-bit analog-to-digital converters (ADCs) with 256 × 10 9  Samples/s and 110 GHz. Subsequently, the raw signals are pre-processed by the receiver digital signal processing (DSP) blocks.
The datasets were collected in a very short time, corresponding to the memory size of the oscilloscope, which is limited. This is referred to as offline processing. At the receiver, the signals were normalised to fit the alphabet. The average launch power in watts can be calculated as follows:
P ( W ) = 1 W · 10 ( P ( d B m ) ) / 10 ) / 1000 = 10 ( P ( d B m ) 30 ) / 10
There are four sets of published data with different launch powers, corresponding to different levels of non-linear distortions during transmission: 2.7 dBm, 6.6 dBm, 8.6 dBm, and 10.7 dBm. Each dataset consists of the ‘alphabet’ (initial analog transmission values), the error-corrected received analog values, and the true labels of the transmitted points. The data have been explained and visualised in detail in the Appendix A. To quantify the system performance of an amplified coherent optical communication system, one uses either a launch power sweep or an OSNR (Optical Signal to Noise Ratio) sweep. While the OSNR metric is used when the system is operating in the linear region, the launch power is the preferred metric to show the performance degradation in the nonlinear region since the induced nonlinear effects are directly proportional to the launch power. The signal-to-noise ratio ( SNR ) for each launch power can be computed using the following expression, where z are the received noisy signals and x are the noiseless target symbols (the launched signals):
SNR = 10 log 10 mean ( z 2 ) mean ( z x 2 ) .
After obtaining this noisy real-world dataset, our task is to decode the received analog values into bit-strings. The kNN is the candidate of choice since it classifies datasets into clusters by associating an ‘average point’ (centroid) to each cluster. In our method, the objective of the clustering algorithm is first to identify, using the set of received signals, a given number M of centroids (one for each cluster) and then to assign each signal to the ‘nearest’ centroid. The second step is classification. This creates the clusters, which can then be decoded into bit signals through the process of demapping. Demapping consists of mapping the original transmission constellation (alphabet) to the current centroids and then assigning the bit-string label associated with that initial transmission point to all the points in the cluster of that centroid. This process completes the final step of the QAM protocol, translating the analog values to bit-strings read by the receiver. The size M of the constellation is known since we know beforehand which QAM protocol is being used. We also know the “alphabet”, i.e., the initial and ideal points at which the signals were transmitted.

2.2. Bloch Sphere

It is well known that all the qubit pure states can be obtained from the zero state using the unitary U [41]
| ψ ( θ , ϕ ) = U ( θ , ϕ ) | 0 = c o s ( θ / 2 ) | 0 + e i ϕ sin ( θ / 2 ) | 1
where
U ( θ , ϕ ) : = cos θ 2 sin θ 2 e i ϕ sin θ 2 e i ϕ cos θ 2 .
These are unit vectors in the unit sphere of C 2 , but it is also well known that the corresponding density matrices are uniquely represented by the Bloch vectors
a ( θ , ϕ ) : = ( sin θ cos ϕ , sin θ sin ϕ , cos θ )
as points in the unit sphere S 2 ( 1 ) R 3  [41] (the Bloch sphere) through the relation
ρ ( θ , ϕ ) = ψ ( θ , ϕ ) ψ ( θ , ϕ ) = cos 2 θ 2 e i ϕ cos θ 2 sin θ 2 e i ϕ cos θ 2 sin θ 2 sin 2 θ 2 = 1 2 1 + cos θ e i ϕ sin θ e i ϕ sin θ 1 cos θ = 1 2 1 + a ( θ , ϕ ) · σ
where 1 is the identity matrix and σ = ( σ x , σ y , σ z ) is the vector of Pauli matrices
σ x = 0 1 1 0 σ y = i 0 1 1 0 σ x = 1 0 0 1 .
Regarding mixed states, notice that Equation (3) is linear and thus, convex combinations of density matrices translate to convex combinations of Bloch vectors, meaning that the interior of the sphere represents the mixed states. Namely, the most general qubit quantum states can be represented by
ρ a ρ ( a ) = 1 2 1 + a · σ , a 2 1 .
Finally, since the Pauli matrices are orthogonal operators under the Hilbert-Schmidt inner product, this inner product is easily computed as
Tr ρ a 1 ρ a 2 = 1 2 1 + a 1 · a 2 .
which for pure states coincides with the fidelity.
Using the Bloch sphere representation of qubit quantum states also makes it easy to find orthogonal states and compute diagonalizations. Indeed, let a be a unit vector ( a = a · a = 1 ), thus representing the pure state ρ a = 1 2 1 + a · σ , then the orthogonal state to ρ a is simply the antipodal point
ρ a = 1 2 1 a · σ
which can be shown by computing the inner product as in Equation (5)
Tr ( ρ + a ρ a ) = 1 4 1 + a · ( a ) = 0 .
Hence, the Bloch eigenvectors for any Bloch vector a are ± a a , the two antipodal points where the line of a intersect the Bloch sphere. Namely, for any mixed quantum state corresponding to the Bloch vector a , we can decompose the quantum state as
1 2 1 + a · σ = p 1 2 1 + a a · σ + ( 1 p ) 1 2 1 a a · σ = 1 2 1 + ( 2 p 1 ) a a · σ
with
2 p 1 = a p = 1 2 ( 1 + a ) .
In the next section, we discuss how we use the Bell-state measurement to estimate the fidelity between quantum states and exposit when this should be chosen over the SWAP test.

2.3. Bell-State Measurement and Fidelity

We use the Bell-state measurement to estimate the fidelity between two pure states. The Bell-state measurement is defined as the von-Neumann measurement of the maximally entangled basis
| ϕ i j : = C N O T ( H 1 ) | i j ,
which by construction is equivalent to a standard basis measurement after ( H 1 ) C N O T as displayed in Figure 2. This measurement can be used to estimate the fidelity as follows.
Lemma 1.
Let | ψ and | χ be two qubit pure states and let | ϕ 11 : = C N O T ( H 1 ) | 11 (the singlet Bell state). Then
ϕ 11 | | ψ | χ 2 = 1 2 ( 1 | ψ | χ | 2 ) .
Proof. 
Let us write the states as
| ψ = ψ 0 | 0 + ψ 1 | 1 | χ = χ 0 | 0 + χ 1 | 1 ,
Then, the state before the standard-basis measurement is
| ψ out = H 1 C N O T | ψ | χ = 1 2 ψ 0 χ 0 + ψ 1 χ 1 ψ 0 χ 1 + ψ 1 χ 0 ψ 0 χ 0 ψ 1 χ 1 ψ 0 χ 1 ψ 1 χ 0
and in particular, the probability of outcome i j = 11 (i.e., simultaneous measurement of both qubits yields value ‘1’ on each qubit) can be written as
ϕ 11 | | ψ | χ 2 = | 11 | ψ o u t | 2 = 1 2 | ψ 0 χ 1 ψ 1 χ 0 | 2 .
The fidelity is obtained now by adding and subtracting ψ 0 ψ 0 χ 0 χ 0 + ψ 1 ψ 1 χ 1 χ 1 and computing
ϕ 11 | | ψ | χ 2 = 1 2 1 ψ 1 χ 0 ψ 0 χ 1 ψ 0 χ 1 ψ 1 χ 0 ψ 0 ψ 0 χ 0 χ 0 ψ 1 ψ 1 χ 1 χ 1 = 1 2 ( 1 | ψ 0 χ 0 + ψ 1 χ 1 | 2 ) = 1 2 ( 1 | ψ | χ | 2 ) ,
concluding the proof.    □
Lemma 1 is used to construct the quantum clustering algorithm in Section 3. We will use the quantum circuit of Figure 2 for the fidelity estimation in the developed quantum algorithm.
Remark 1.
Since we are only interested in the i j = 11 outcome and we are measuring qubits, the course-grained projective measurement defined by ϕ 11 ϕ 11 and 1 ϕ 11 ϕ 11 is sufficient for computing the inner product. The non-destructive version of this measurement is known as the SWAP test [24,25], first described in [23]. This test has been used extensively for overlap estimation in quantum algorithms [2]. The SWAP test requires one to only measure an ancilla qubit instead of the two input qubits, leaving them in the post-measurement state, which can be used later for other purposes. However, given the current limitations of NISQ technologies, storing quantum information for reuse is quite impractical; therefore, we prefer the destructive measurement version for overlap estimation. Namely, we use the Bell-state measurement instead of the SWAP test because the post-measurement state is unnecessary.

2.4. Nearest-Neighbour Clustering Algorithms

Clustering is a simple, powerful and well-known machine-learning algorithm that has been extensively used throughout the literature. In this section, we summarise some standard and basic notions introduced by clustering and define this class of heuristic algorithms precisely so that we can make clear the difference between regular clustering and the quantum and quantum-inspired clustering algorithms introduced in this paper. We first define the involved variables needed for the kNN.
Definition 1
(Clustering State). We define a k-Nearest-Neighbour Clustering State, or clustering state for short, as a collection ( D , c ¯ , D , d ) where
  • D is a space called dataspace with elements called points.
  • D D is a subset called dataset consisting of points called datapoints.
  • c ¯ = ( c 1 c 2 c k ) D k is a list (of size k) of points called centroids
  • d : D × D R is a lower bounded function called dissimilarity function, or dissimilarity for short.
Note that d does not have to be a distance metric. We now define the basic steps that are repeated in the clustering algorithm.
Definition 2
(Clusters and Centroid update). Let ( D , c ¯ , D , d ) be a clustering state. We define the clusters of the state as, for each j = 1 , , k , the set
C j ( c ¯ ) = p D | d ( p , c j ) d ( p , c ) = 1 , , k , p < j C ( c ¯ ) .
We now define the possible new centroids of a subset C D as the set
P ( C ) : = argmin x D p C d ( x , p )
of all points minimising the total (and thus the average) dissimilarity. Then, we call a centroid update any function c update : P ( D ) D (where P denotes the power set) of clusters such that c update ( C ) is a possible new centroid, namely such that c update ( C ) P ( C ) , for all j = 1 , , k . We then define the following short-hand notation for the centroid update of c ¯ , namely the new list of centroids
c ¯ update ( c ¯ ) = c update ( C 1 ( c ¯ ) ) , , c update ( C k ( c ¯ ) ) .
We now define the general k-nearest-neighbour clustering algorithm.
Definition 3
(K-Nearest-Neighbour Clustering Algorithm (kNN)). Finally, we define a K-Nearest-Neighbour clustering algorithm (kNN) as a pair of clustering state and centroid update ( D , c ¯ 1 , D , d , c ¯ update ) . The kNN algorithm defines a sequence of clustering states ( D , c ¯ i , D , d ) via c ¯ i + 1 = c ¯ update ( c ¯ i ) for all i N which we call the iterations (of the algorithm).
A point of note is that Equation (17) implies that the new centroid is one of the points x in the dataspace that minimises the total (and hence the average) dissimilarity with all the points p in the cluster. Moreover, notice that this definition requires one to initialise or populate the list c ¯ with initial values, i.e., the initial centroids c ¯ 1 must be defined as a starting point for the clustering algorithm. The initial centroids can be assigned randomly or defined as a function of parameters such as the dataset.
Another comment about Equation (17): in our case, we will see later in Section 3.2 that all choices of points from the set P j will be equivalent. As in our algorithm, this freedom of choice can be exploited to reduce the amount of computation or for other optimisations.
Notice that Equations (17) and (18) implies that centroids are generally not part of the original dataset; however, according to Equations (17) and (18), they must be restricted to the space in which the dataset is defined. Definitions involving centroids for which c ¯ D k are possible but are not used in this work.
One can observe that any kNN can be broken down into two steps that keep alternating until a stopping condition (a condition which, when true, forces the algorithm to terminate) is met: a cluster update which updates the points associated with the newly calculated centroid, and then a centroid update which recalculates the centroid based upon the new points associated to it through its cluster. For the cluster update, the value of the centroid calculated in the previous iteration is taken, and its cluster set is constructed by collecting all the points in the dataset that are ‘closer’ to it than any other centroid. The ‘closeness’ is computed by using a pre-defined dissimilarity. In the next step, the centroids are updated by searching in the dataspace, and for each updated cluster, a new point for which the sum of dissimilarities between that point and all points in the cluster is minimised.
This procedure will lead to different results if one changes the dissimilarity or the space of data points or both. In this paper, we explore the effects of changing this dissimilarity as well as the space of data points, and we shall explain it in the context of quantum states.

2.5. Euclidean Dissimilarity and Classical Clustering

It can be observed from the centroid update in Equations (16) and (17) that the dissimilarity plays a central role in the clustering algorithm. The nature of this function directly controls the first step of the cluster update since the dissimilarity is used to compute the ‘closeness’ between any two points in the dataspace. It is also apparent that if the dissimilarity is changed in the centroid update, the points at which the minimum is achieved could also change.
The Euclidean dissimilarity d e : R n × R n R is defined simply as the square of the Euclidean distance between the points:
d e ( a , b ) = a b 2 .
For a finite subset C R n , the minimisation of Equations (17) and (18) yields a unique point, reducing to the average point of the cluster:
c e update ( C ) : = argmin x R n p C d e ( x , p ) = 1 | C | p C p ,
which we call the Euclidean centroid update. This is the most typical case of the centroid update, where the new centroid is updated as the mean point of all points in the cluster. This corresponds to the classic k-means clustering algorithm [42], which now can be defined as follows.
Definition 4
(n-Dimensional Euclidean Classical kNN (nDEC-kNN)). An n-dimensional classical Euclidean kNN algorithm is any clustering algorithm with dataspace R n , Euclidean dissimilarity and cluster update as the average point, as in Equation (20). Namely, any clustering algorithm of the form ( D , c ¯ , R n , d e , c e update ) .
The computation of the centroid through Equation (20) instead of Equations (17) and (18) reduces the complexity of the centroid update step; such a reduced expression is used to compute the updated centroids rather than searching the entire dataspace for the minimising points during the centroid update.

2.6. Cosine Dissimilarity

In this work, we project the collected two-dimensional dataset (described in Section 2.1) into a sphere via the ISP. After this projection, the calculation of the centroids according to Equation (20) would generally yield centroids which lie inside the sphere instead of on the S 2 ( r ) surface due to the convex nature of the sphere’s surface.
In our work, to use qubit pure states, we restrict the dataspace D to the sphere surface S 2 ( r ) , forcing the centroids to lie on the surface of a sphere. This naturally leads to the question of what the proper reformulation of Equations (17) and (18) is, and whether a computationally inexpensive formula similar to Equation (20) exists for this case as well. This question will be answered in Lemma 3. For this purpose, it is useful to first define the cosine dissimilarity [43] and see how it relates to the Euclidean dissimilarity.
Definition 5
(Cosine Dissimilarity). For two points, a and b in an inner-product space D , the cosine dissimilarity, is defined as:
d s ( a , b ) = 1 a · b a b ,
where a · b is the inner product between the two points expressed as vectors from the origin, a is the norm of a induced by the inner product.
This is called cosine dissimilarity because when a , b R n the cosine dissimilarity d s ( a , b ) reduces to 1 cos ( α ) , where α is the angle between a and b . The cosine dissimilarity is also known sometimes as cosine distance (although it is not a distance), while a · b a b is well known as cosine similarity. This quantity, by construction, only depends on the direction of the vectors and not their magnitude. Said otherwise, we have
d s ( a , b ) = d s ( c a , b ) = d s ( a , c b )
for any positive constant c > 0 . We also note that the cosine dissimilarity of Equation (21) can be related to the Euclidean dissimilarity of Equation (19) if a and b lie on the n-sphere S n ( r ) : = s R n + 1 | s 2 = r of radius r, as stated by the following lemma.
Lemma 2.
Let d s and d e be the cosine and Euclidean dissimilarities, respectively. Let s 1 , s 2 S n ( r ) be points on the n-sphere of radius r, then
d e ( s 1 , s 2 ) = 2 r 2 d s ( s 1 , s 2 ) .
Proof. 
Assuming s 1 , s 2 S n ( r ) , Equation (21) reduces to:
d s ( s 1 , s 2 ) = 1 1 r 2 s 1 · s 2 ,
then
2 r 2 d s ( s 1 , s 2 ) = 2 r 2 2 s 1 · s 2 = s 1 2 + s 2 2 2 s 1 · s 2 = s 1 s 2 2 = d e ( s 1 , s 2 ) ,
concluding the proof.    □
From this, we can expect that the minimiser of the centroid update equation (Equation (17)) computed using the cosine dissimilarity will closely relate to the Euclidean centroid update. However, the derivation is not straightforward since the Euclidean centroid update does not lie on the same sphere, but lies inside at a smaller radial distance. This is shown in the following lemma.
Lemma 3.
Let C S n ( r ) be a finite set, then
c s update ( C ) : = argmin x S n ( r ) p C d s ( x , p ) = r p C p p C p .
We call this the cosine or spherical centroid update. In particular, thus
c s update ( C ) = r c e update ( C ) c e update ( C )
where c e update ( C ) = 1 | C | p C p is the Euclidean centroid update of Equation (20).
Proof. 
The second claim is trivial; we thus have to prove only the first claim. Given that C S n ( r ) , then, according to Lemma 2, the cosine dissimilarity given in Equation (21) reduces for all a , b C to:
d s ( a , b ) = 1 a · b a b = 1 1 r 2 a · b .
The minimisation in Equation (17) can then be calculated for the cosine dissimilarity with a Lagrangian (see Equation (31)) that satisfies Equations (17) and (18) at the minimising point. Namely, we have to find x R n that minimises
f ( x ) = p C d ( x , p ) ,
subject to the restriction condition that assures that x S n ( r ) , that is
g ( x ) = x 2 r 2 = 0 .
Such a Lagrangian is expressed as
L ( x , λ ) = f ( x ) λ g ( x ) = p C d s ( x , p ) λ ( x 2 r 2 ) ,
where λ is the Lagrangian multiplier. We then calculate the centroid update by employing the derivative criteria to Equation (31).
0 = p C 1 1 r 2 x · p λ x 2 + r 2 = 1 r 2 p C p 2 λ x
Therefore, the following holds:
x = 1 2 λ r 2 p C p .
Substituting Equation (33) into the restriction in Equation (30), we obtain the multiplier λ  as:
| λ | = 1 2 r 3 · p C p .
Therefore, the critical point and minimising point c s is written as
c s = r p C p p C p ,
as claimed.    □
We can observe that Lemma 3 implies that the minimiser obtained by restricting the point to lie on the surface of the sphere is the projection (from the origin) of the minimiser of the Euclidean dissimilarity into the sphere’s surface.
Corollary 1.
Let C S n ( r ) be a finite set, and the possible new centroids of C under cosine dissimilarity in R 3 are
P s ( C ) : = argmin x R n p C d s ( x , p ) = r p C p : r > 0 .
We call these the cosine possible new centroids.
Proof. 
We have
c s update : = argmin x R n { 0 } p C d s ( x , p ) = argmin x S n ( r ) , r > 0 p C d s ( x , p ) = r p C p p C p : r > 0 = r p C p : r > 0 ,
where the last equality follows from Equation (22), namely
d s r p C p p C p , p = d s p C p , p
for all r and p ; thus making all the points r p C p , r > 0 equivalent possibilities for the centroid update.    □

2.7. Stereographic Projection

The inverse stereographic projection (ISP), shown in Figure 3, is a bijective mapping
s r 1 : R n S n ( r ) { N }
from the Euclidean space R n into an n-sphere S n ( r ) R n + 1 without the north pole N.
This mapping is interesting because of the natural equivalence between the 3D unit sphere S 2 ( 1 ) and the Bloch sphere of qubit quantum states. In this case, as displayed in Figure 3, the ISP maps a two-dimensional point p = ( p x , p y ) R 2 into a three-dimensional point s r 1 ( p ) = ( s x ( p ) , s y ( p ) , s z ( p ) ) S 2 ( r ) { ( 0 , 0 , r ) } through the following set of transformations:
s x ( p ) = p x · 2 r 2 p x 2 + p y 2 + r 2 s y ( p ) = p y · 2 r 2 p x 2 + p y 2 + r 2 s z ( p ) = r · p x 2 + p y 2 r 2 p x 2 + p y 2 + r 2 = p x · 2 r 2 p 2 + r 2 = p y · 2 r 2 p 2 + r 2 = r · p 2 r 2 p 2 + r 2 .
The polar and azimuthal angles of the projected point are given by the expressions:
ϕ ( p ) = tan 1 p y p x θ ( p ) = 2 · tan 1 r p
This information, particularly Equation (41), will allow us to associate each point in R 2 to a unique quantum state through the Bloch sphere. Still, the inverse stereographic projection does not need to be bound to the preparation of quantum states and can be used as a transformation between classical kNN algorithms. Indeed, we can stereographically project and then perform classical clustering on the 3D data, and namely perform 3DEC-kNN as defined in Definition 4.
Definition 6
(Three-Dimensional Stereographic Classical kNN (3DSC-kNN)). Let s r 1 be an ISP, and let ( D , c ¯ , R 2 , d e ) be a clustering state (recall, D is the dataset and c ¯ are the initial centroids). We then define the 3D Stereographic Classical kNN (3DSC-kNN) as ( s r 1 ( D ) , s r 1 ( c ¯ ) , R 3 , d e , c e update ) .
Here, we apply s r 1 elementwise and thus s r 1 ( c ¯ ) = ( s r 1 ( c 1 ) , , s r 1 ( c k ) ) for any list of centroids c ¯ and s r 1 ( D ) = s r 1 ( p ) : p C for any set of points C, and where C j are the clusters as defined in Equation (16).
Remark 2.
Derivations and further observations can be found in Appendix C. Of particular note, as explained in more detail in Appendix C.2, is that changing the plane’s distance from the centre of the sphere is equivalent to a change of radius. Therefore, we can limit our analysis to projections where the centre of the plane is also the centre of the sphere without a loss of generality.

3. Stereographic Quantum Nearest-Neighbour Clustering (SQ-kNN)

This section proposes and describes the quantum kNN algorithm using stereographic embedding. In Section 4, we demonstrate an equivalent quantum-inspired (classical) version. Section 3.1 defines the method to convert the classical data into quantum states. In what follows, we describe how these states are manipulated so that we obtain an output that can be used to perform clustering, using the circuit of Section 2.3 for the dissimilarity estimation. Section 3.2 defines the quantum algorithm in terms of Definitions 1–3, and Section 3.3 discusses the complexity and scalability of the algorithm. Section 3.4 discusses the SQ-kNN algorithm in the context of mixed states.

3.1. Stereographic Embedding, Bloch Embedding and Quantum Dissimilarity

For quantum algorithms to work on classical data, the classical data must be converted into quantum states. This process of encoding classical data into quantum states is also called embedding. The embedding of classical data into quantum states is not unique, and each technique’s pros and cons must be weighed in the context of a specific application. The process of data embedding is an active field of research. More details on existing embedding can be found in Appendix B.
Here, we propose the stereographic embedding as an improved embedding of classical vector p R 2 into a quantum state using its stereographic projection. We can split stereographic embedding into two steps: inverse stereographic projection and Bloch embedding. We define Bloch embedding, a variation of angle embedding, as follows.
Definition 7
(Bloch embedding). Let P R 3 . We define the Bloch embedded quantum state, or Bloch embedding for short, of P as the quantum state
ψ P : = 1 2 1 + P P · σ
which is simply the pure state obtained using P / P as the Bloch vector.
At this point, we define this general embedding for general three-dimensional points since this general form will yield the quantum dissimilarity defined next. We will also define this embedding in the context of the ISP in Definition 9, below.
To obtain ψ P , the state can be encoded as explained in the preliminaries in Section 2.2, through Equations (2) and (3). For Bloch embedding, the θ and ϕ of Equations (2) and (3) would be the polar and azimuthal angles of P , respectively. We now define the quantum dissimilarity, as follows.
Definition 8
(Quantum Dissimilarity). For any two points P 1 , P 2 R 3 , we define the quantum dissimilarity as
d q ( P 1 , P 2 ) : = 1 2 ( 1 Tr ( ψ P 1 ψ P 2 ) ) ,
where ψ P is the Bloch embedding of P .
Notice that, as per this definition, the classical two-dimensional points are embedded in pure states only. In Section 3.4, we consider Bloch embedding to be the centroids into mixed states as well, showing that this does not provide an advantage in our framework. This quantum dissimilarity can be obtained either with the SWAP test or with the Bell state measurement on ψ P 1 ψ P 2 as described in Section 2.3. In our application, we use the Bell state measurement (depicted in Figure 2), as we do not need the extra resources of the SWAP test that allow us to keep the post-measurement state. For more details, see Lemma 1.
By Equation (5), the quantum dissimilarity is proportional to the cosine dissimilarity (this might not be true for other definitions of quantum dissimilarity, as in Section 3.4 where we redefine it to include embedding into mixed states)
d q ( P 1 , P 2 ) = 1 2 ( 1 Tr ( ψ P 1 ψ P 2 ) ) = 1 4 1 P 1 P 1 · P 2 P 2 = 1 4 d s ( P 1 , P 2 )
It is also proportional to the Euclidean distance for points on the same sphere (points with the same magnitude) as per Lemma 2. Namely, if s 1 , s 2 S 2 ( r ) then:
d q ( s 1 , s 2 ) = 1 8 r 2 d e ( s 1 , s 2 ) .
We can finally define stereographic embedding as follows.
Definition 9
(Stereographic Embedding). We define the stereographic embedding of a classical vector p R 2 as
1. 
Projecting the 2D point p into a point on the sphere of radius r in a 3D space through the ISP:
s : = s r 1 ( p ) S 2 ( r ) R 3 ;
2. 
Bloch embedding s into ψ s = ψ s r 1 ( p ) .
Comparing the distance estimate of the Stereographic embedding procedure (Equations (45) and (A18)) with that for the hybrid quantum-classical k-means with angle embedding (the ‘distance loss function’ described in [10]), we can observe that the theoretical performance has been improved, since the estimate has been much improved with respect to the closeness to Euclidean distance. This leads us to expect a performance improvement of the SQ-kNN algorithm over the hybrid quantum-classical implementation of quantum k-means with angle embedding.
A very time-consuming computational step of kNN involves the repeated calculations of distances between the dataset points meant to be classified and each centroid. In the case of the quantum kNN in [2], since angle embedding is not injective, many steps must be spent after estimating the fidelity to calculate the distance between the points using the norms. Even in [32,33,37], the norms of the points have to be stored classically, leading to much computational expense. Our method has the clear benefit of calculating the cosine dissimilarity directly through fidelity estimation. No further calculations are required due to all stereographically projected points having the same norm r in the sphere, and the existence of a bijection between the ISP and the original 2D datapoints, thus saving computational time and resources. In summary, Equations (44) and (45) portray a method to measure a dissimilarity that leads to consistent clustering involving pure states.
As one can observe, in the case of stereographically projected points, d q is directly proportional to the Euclidean dissimilarity between them. Since all the points after the projection into the sphere have equal modulus r, and each projected point corresponds to a unique 2D data point, we can directly compare the probability of obtaining outcome i j = 11 on the Bell-state measurement circuit for cluster assignment. This eliminates extra steps needed during computation to account for the different moduli of points on the two-dimensional plane.

3.2. The SQ-kNN Algorithm

We now have all the building blocks to define the quantum clustering algorithm. The quantum part will be the dissimilarity estimation d q , obtained by embedding the data into quantum states as described in Section 3.1 and then feeding it into the quantum circuit described in Section 2.3 and Figure 2 to estimate an outcome probability. The finer details of distance estimation are further described in Appendix E. We can now formally define the developed algorithm building on the definition of clustering state (Definition 1), of clustering algorithm and cluster update provided by Definitions 2 and 3, of ISP as defined in Section 2.7, and of quantum dissimilarity d q from Definition 8.
Definition 10
(Stereographic Quantum kNN (SQ-kNN)). Let s r 1 be the ISP, let d q be the quantum dissimilarity, and let ( D , c ¯ , R 2 , d e ) be a clustering state (where D and c ¯ are two-dimensional datasets and initial centroids). We then define the Stereographic Quantum kNN (SQ-kNN) as the kNN clustering algorithm
s r 1 ( D ) , s r 1 ( c ¯ ) , R 3 , d q , c ¯ q update
where c q update : = p C p .
The complete process of performing SQ-kNN in practice can be described in detail, as follows.
  • First, prepare to embed the classical data and initial centroids into quantum states using the ISP: project the two-dimensional datapoints and initial centroids (in our case, the alphabet) into a sphere of radius r, and calculate the polar and azimuthal angles of the points. This first step is executed entirely on a classical computer.
  • Cluster Update: The calculated angles are used to create the states using Bloch embedding (Definition 7). The dissimilarity between the centroid and point is then estimated using the Bell-state measurement. Once the dissimilarities between a point and all the centroids have been obtained, the point is assigned to the cluster of the ‘closest’ centroid. This is repeated for all the points that have to be classified. The quantum circuit and classical controller handle this step entirely. The controller feeds in the classical values at the appropriate times, stores the results of the various shots and classifies the point to the appropriate cluster.
  • Centroid Update: Since any non-zero point on the subspace of c s (see Corollary 1, Figure 4) is an equivalent choice, to minimise the computational expense, the centroids are updated as the sum point of all points in the cluster—as opposed to the average, for example, which minimises the Euclidean dissimilarity (Equation (20)).
Once the centroids are updated, Step 2 (Cluster Update) is repeated, followed once again by Step 3 (Centroid Update) until a decided stopping condition is fulfilled.
Compared to 2D quantum kNN clustering with angle or amplitude embedding, the differences with the SQ-kNN algorithm lie in the embedding and the post-processing after the inner-product estimation.
  • The stereographic embedding of the 2D datapoints is conducted by theinverse stereographic projecting the point into a sphere of a chosen radius and then producing the quantum state obtained by rescaling the sphere to radius one.
    In contrast, in angle embedding, the coefficients of the vectors are used as the angles of the Bloch vector (also known as dense angle embedding [44]), while in amplitude embedding, they are used as the amplitudes in the standard basis. For 2D vectors, amplitude embedding allows one to encode only one coefficient (instead of two) in one qubit, and sometimes angle embedding would also encode only one coefficient by using a single rotation (standard angle embedding [45]). Both angle and amplitude embeddings require the lengths of the vectors to be stored classically beside the quantum state, which is not needed in Bloch embedding.
  • No post-processing is needed after the overlap estimation of stereographically embedded data, as the obtained estimate is already a linear function of the inner product, as opposed to standard approaches using angle or amplitude encoding. Amplitude embedding also requires non-trivial computational time in the state preparation process. In contrast, in angle embedding, though the state preparation time is constant, recovering a useful dissimilarity (e.g., Euclidean) may involve many post-processing steps.
In short, stereographic embedding has the advantage of angle over amplitude embedding of being able to encode all values of a vector and low state preparation time, and the advantage of amplitude versus angle embedding in the recovery of the dissimilarity.

3.3. Complexity Analysis and Scaling

Let the ISP of a d-dimensional point p = [ x 1 x 2 x d ] into S d ( r ) , using the point ( r , 0 , 0 , , 0 ) (the ‘North pole’) as the projection point, be the point s = [ s 0 s 1 s d ] . It is known that the Cartesian coordinates of s are given by:
s 0 = r p 2 r 2 p 2 + r 2 s i = 2 r 2 x i p 2 + r 2 i = 1 , , d
where p 2 = j = 1 d x j 2 . Hence one can observe that the time complexity of the projection for a single d -dimensional point is O ( d ) , provided we only need the Cartesian coordinates. However, for the Stereographic Embedding procedure, one would need to calculate the angles made by s with the axes of the space, making the time complexity of projection O ( p o l y ( d ) ) . Therefore, the total time complexity of the Stereographic Embedding for a d -dimensional dataset D of size | D | = N and k centroids is given by O ( ( k + N ) p o l y ( d ) ) . We now specify two strategies for scaling our algorithm for higher dimensional datapoints.

3.3.1. Using Qubit-Based System

We consider the case where we have two d-dimensional vectors p 1 , p 2 R d , and we want to compute the quantum dissimilarity of vectors. If we have a qubit-based system and use dense angle encoding for encoding the stereographically projected point, we would encode the d calculated angles using d 2 qubits. Namely, for the ( d + 1 ) -dimensional projection of a d-dimensional point p 1 , one would obtain d angles [ θ 1 θ 2 θ d ] that specify the projected point s 1 on S d ( r ) . We then encode this vector using the same unitary, as follows:
| ψ 1 = j ( 1 , 3 , , d 1 ) | ψ 1 j = j o d d ( d ) U ( θ j , θ j + 1 , 0 ) | 0 .
If d is odd, we can pad [ θ 1 θ 2 θ d ] with an extra 0 to make it even. The other point s 2 = s r 1 ( p 2 ) will be encoded into the state
| ψ 2 = j ( 1 , 3 , , d 1 ) | ψ 2 j = j o d d ( d ) U ( θ j , θ j + 1 , 0 ) | 0 .
Now, to find the overlap between the states, one would have to perform the Bell-state measurement (Section 2.3) pairwise using | ψ 1 j and | ψ 2 j as inputs, i.e.,
| ϕ 11 | | ψ 1 j | ψ 2 j | 2 = 1 2 ( 1 | ψ 1 j | ψ 2 j | 2 )
In the common practical case of the vectors being expressed in an orthogonal basis, one would only have to find the overlap for j = j . We would then have the quantum dissimilarity by adding up the individual probabilities
d q ( p 1 , p 2 ) = j o d d ( d ) | ϕ 11 | | ψ 1 j | ψ 2 j | 2
This procedure has a time complexity of O ( d ) . It is important point to note that this quantum dissimilarity will no longer correspond directly to the inner product or Euclidean distance between either p 1 and p 2 or s 1 and s 2 . With the strategy of pairwise overlap estimation, we observe that if the number of shots to estimate | ϕ 11 | ( | ψ 1 j | ψ 2 j | 2 is kept constant, the error in estimation will be d . Hence, taking into account the increase in the number of shots to estimate the quantum dissimilarity with a given total error ϵ , the time complexity of this qubit implementation of overlap estimation between two points using SQ-kNN scales is O ( ϵ 1 p o l y ( d ) ) . Hence, for all points and clusters, the time complexity would be O ( ϵ 1 k N p o l y ( d ) ) .
It is shown in [46] that collective measurements are a better strategy than repeated individual measurements for overlap estimation. Although this is shown in [46] for estimating overlap between two states given the availability of multiple copies of the same states, similar collective measurement strategies could be applied in this case for better results. In conclusion, the time complexity of SQ-kNN for qubit-based implementation is
O ( ϵ 1 k N p o l y ( d ) )

3.3.2. Using Qudit-Based System

Consider
| 1 : = i 0 , , d 1 | i i .
Then, for any two real vectors | ψ = ψ i | i and | ϕ = ϕ i | i , namely when ψ i , ϕ i R , we have
1 | ( | ψ | ϕ ) = ψ i ϕ i = ϕ | ψ = ϕ | ψ .
Now, we make a qudit Bell measurement, which can be obtained as in Figure 2 by replacing the Hadamard with the Fourier transform and the qubit C N O T with the qudit C N O T i i i + j mod d j (if we have multiple qubits instead of qudits, then the solution is even simpler: perform a qubit Bell measurement with each pair of qubits because the tensor product of maximally entangled states is still a maximally entangled state). Then, one of the basis states of this von Neumann measurement will be
| Φ | Φ d = 1 d | 1 .
Thus, the inner product between two real vectors can still be measured with a Bell measurement, but the resulting probability of measuring outcome | Φ scales as
| Φ | ( | ψ | ϕ ) | 2 = 1 d | ϕ | ψ | 2 ;
meaning that, as the inner product remains constant going to higher dimensions, the number of shots needed to estimate the inner product with constant precision scales polynomially in the dimension. In contrast, such complexity for the SWAP test remains constant because the contribution of the fidelity to the outcome probability is not divided by the dimension d. This is why the SWAP test is usually considered for inner product estimation, even if in the case of qubits the Bell measurement is a simpler solution [25].

3.4. SQ-kNN and Mixed States

Instead of estimating the quantum dissimilarity, we can use the datapoints produced by the ISP to perform classical kNN clustering on the 3D-projected data. We called this the 3DSC-kNN (3D Stereographic Classical kNN) in Definition 6. This algorithm produces centroids that are inside the sphere. As previously pointed out, when computing the Euclidean 3D centroid on the data projected on the sphere, the result is a point inside the sphere rather than on the sphere itself.
In the Bloch sphere, internal points are mixed states, namely states with added noise. In contrast, the quantum algorithm (SQ-kNN) always produces pure centroids, namely points on the surface of the sphere. The only noiseless states are the pure states on the surface of the sphere, and thus the intuition is that arguably mixed states should not help. However, this is not immediately clear from the algorithm. Comparing 3DEC-kNN to SQ-kNN, it is thus natural to ask whether embedding the centroids into mixed states inside the Bloch sphere improves the accuracy.
Here, we show that the intuition is correct, namely that projecting into the pure state centroid is a better option. The reason is that while the quantum dissimilarity is proportional to the Euclidean dissimilarity for states in the same sphere, the same is not true for Bloch vectors with different lengths.
To allow for mixed state embedding, we can modify the definition of quantum dissimilarity (Equation (45)) to produce mixed states whenever the 3D vector has a length of less than one. This results in the following new quantum dissimilarity.
Definition 11
(Noisy Quantum Dissimilarity). Let B 2 ( 1 ) = P R 3 | P 1 be the ball of radius 1. We define the noisy quantum dissimilarity as the function d ˜ q : B 2 ( 1 ) × B 2 ( 1 ) R
d ˜ q ( P 1 , P 2 ) : = 1 2 1 Tr ( ρ P 1 ρ P 2 )
where ρ P is the quantum state of the Bloch vector P as in Equation (4).
Now suppose we have a convex combination of pure states; namely, suppose we have
ρ ¯ = i p i ρ P i P i = 1 i .
where p i is a probability distribution ( p i > 0 and p i = 1 ). By linearity, we have
ρ ¯ = ρ p i P i = : ρ ( P ¯ ) P ¯ = p i P i ,
By convexity, P ¯ will always lie in the sphere. Namely, we have P ¯ 1 and thus by linearity
d ˜ q P ¯ , P = p i d ˜ q P i , P = p i d q P i , P .
The result in Equation (57) can be interpreted as another two-step process: first, repeatedly performing the Bell-state measurement of each state  ρ P i that makes up the cluster and ρ P corresponding to the datapoint, to estimate each individual dissimilarity; and then, taking the weighted average of the dissimilarities according to the composition of the mixed state centroid. This procedure is clearly impractical experimentally and also no longer correlates to the cosine dissimilarity for mixed states.
Computing the diagonalization of ρ ¯ as per Equation (8)
ρ = p ρ P P + ( 1 p ) ρ P P p = 1 2 ( 1 + P ¯ ) = p ψ P ¯ + ( 1 p ) ψ P ¯
(where ψ is the Bloch embedding) makes the estimation more practical by reducing it to two estimations of d q , namely
d ˜ q P ¯ , P = p d ˜ q P ¯ P ¯ , P + ( 1 p ) d ˜ q P ¯ P ¯ , P   = p d q P ¯ , P + ( 1 p ) d q P ¯ , P
The implementation portrayed at Equation (59) simplifies the measurement procedure of the mixed state. Furthermore, instead of estimating d q ( ± P ¯ , P ) separately, the estimation can be performed directly by preparing ψ ( P ) with probability p and ψ ( P ) with probability 1 p , and finally collecting all the outcomes in a single estimation, which requires a larger number of shots to achieve the same precision of estimation. Another issue is that the points P ¯ , P ¯ have to be computed, which is quite time-consuming. This is true even for Equation (57); however, a number of shots proportional to the number of Bloch vectors P i in the cluster is needed for an accurate estimation. Regardless, linearity and convexity make it clear that using mixed states can only increase the quantum dissimilarity.
Namely, while in Euclidean dissimilarity, points inside the sphere can reduce the dissimilarity, the quantum dissimilarity is proportional to the Euclidean dissimilarity only for unit vectors and actually increases for points inside the Bloch sphere. Hence, we conclude that the behaviour of 3DSC-kNN does not carry over to SQ-kNN.

4. Quantum-Inspired Stereographic Nearest-Neighbour Clustering (2DSC-kNN)

We have detailed the developed quantum algorithm in the previous Section 3. This section develops the classical analogue to this quantum algorithm—the ‘quantum-inspired’ classical algorithm. A table summarising all the algorithms discussed in this paper, including the next one, can be found in Table 1. We begin by defining this analogous classical algorithm in terms of the clustering state (Definition 1), deriving a relationship between the Euclidean and spherical centroids given datapoints that lie on a sphere, and then proving our claim that the defined classical algorithm and previously described stereographic quantum kNN are indeed equivalent.
Recall from Lemma 3 that
c s update ( C ) : = argmin x S n ( r ) p C d s ( x , p ) = r p C p p C p .
Definition 12
(Two-Dimensional Stereographic Classical kNN (2DSC-kNN)). Let s r 1 be the ISP, and let ( D , c ¯ , R 2 , d e ) be a 2D euclidean clustering state. We define the 2D Stereographic Classical kNN (2DSC-kNN) as
s r 1 ( D ) , s r 1 ( c ¯ ) , S 2 ( r ) , d s , c ¯ s update .
Remark 3.
Notice that due to the cluster update being cosine ( c ¯ s update ) and Lemma 3, we can equivalently substitute d s with d e , namely we can substitute it without changing the outcome of the cluster update. In our implementation, we use the Euclidean dissimilarity for simplicity of coding.
To expand upon Definition 12, for the quantum-inspired/classical analogue stereographic kNN, the steps of execution are as follows:
  • Stereographically project all the 2-dimensional data and initial centroids into the sphere S 2 ( r ) of radius r. Notice that the initial centroids will lie on the sphere by construction.
  • Cluster Update: Form the clusters using the method defined in Equation (16), i.e., form all C j ( c ¯ i ) . Here, D = S 2 ( r ) and dissimilarity d = d s ( p , c ) = 1 2 r 2 d e ( p , c ) (Definition 5 and Lemma 2).
  • Centroid Update: A closed-form expression for the centroid update was calculated in Equation (35) c s updated = r p C p p C p . This expression recalculates the centroid once the new clusters have been formed. Once all the centroids are updated, Step 2 (cluster update) is then repeated, and so on, until a stopping condition is met.

4.1. Equivalence

We now want to show that the 2DSC-kNN algorithm of Definition 12 is equivalent to the previously defined quantum algorithm using stereographic embedding (Definition 10). For that, we first define the equivalence of two clustering algorithms.
Definition 13
(Equivalence of Clustering Algorithms). Let K = ( D , c ¯ 1 , D , d , c ¯ update ) and K = ( D , c ¯ 1 , D , d , c ¯ update ) be two clustering algorithms. They are said to be equivalent if there exists a transformation t : D D such that it maps the data, initial centroids and centroid update, and clusters of K to the data, initial centroids and centroid update, and clusters of K ; namely if
1. 
D = t ( D ) ,
2. 
c ¯ 1 = t ( c ¯ 1 ) and c ¯ update = t c ¯ update ,
3. 
C j ( t ( c ¯ ) ) = t ( C j ( c ¯ ) ) for all j = 1 , , k and any c ¯ D k .
where we apply t elementwise and thus t ( c ¯ ) = ( t ( c 1 ) , , t ( c k ) ) for any list of centroids c ¯ and t ( C ) = t ( p ) : p C for any set of datapoints C, and where C j are the clusters as defined in Equation (16).
Theorem 1.
SQ-kNN (Definition 10) and 2DSC-kNN (Definition 12) are equivalent.
Proof. 
By definition, let ( D , c ¯ 1 , R 2 , d e ) be the 2D clustering state, thus giving us the SQ-kNN algorithm as
K = S , s ¯ 1 , R 3 , d q , c ¯ q update
and the 2DSC-kNN clustering algorithm as
K = S , s ¯ 1 , S 2 ( r ) , d s , c ¯ s update
where
S : = s r 1 ( D ) s ¯ 1 : = s r 1 ( c ¯ 1 )
Let us use the notation p ^ : = p p and define the transform t : R 3 S 2 ( r ) as t ( p ) = r p ^ , which rescales any vector to have length r. Observe that trivially for all p S 2 ( r ) , t ( p ) = p and thus t s r 1 = s r 1 . Therefore
t ( S ) = S , t ( s ¯ 1 ) = s ¯ 1 .
Moreover, the equivalence of centroids is obtained since
t ( c q update ( C ) ) = t p C p = r p C p p C p = c s update ( C ) .
For the clusters, we prove the equivalence of the cluster updates as follows. We will use d s ( a , b ) = 4 · d q ( a , b ) (Equation (44)) and the fact that d s and d q are invariant under t, namely d s t = d s and d q t = d q , or more explicitly
d s ( t ( a ) , b ) = d s ( a , b ) = d s ( a , t ( b ) ) , d q ( t ( a ) , b ) = d q ( a , b ) = d q ( a , t ( b ) ) .
Let now s ¯ ( R 3 ) k . Then, using the above equations and that t ( s ) = s , t ( S ) = S , we have
C j ( t ( s ¯ ) ) = p S | d s ( p , t ( s j ) ) d s ( p , t ( s ) ) = 1 , , k , p < j C ( t ( s ¯ ) ) = p S | d q ( p , s j ) d q ( p , s ) = 1 , , k , p < j C ( s ¯ )
where the change in the dissimilarity inequality has also transformed the calculation of C ( s ¯ ) into the calculation of C ( s ¯ ) . We are now finished, since t ( s ) = s for s S and thus
C j ( t ( s ¯ ) ) = t ( p ) t ( S ) | d q ( p , s j ) d q ( p , s ) = 1 , , k , p < j C ( s ¯ )   = ( C j ( s ¯ ) )
This concludes the proof    □
The following discussion provides a visual intuition of Theorem 1. In Figure 4, the sphere with centre origin (O) and radius r is the stereographic sphere into which the two-dimensional points are projected, while the sphere with centre O and radius 1 is the Bloch sphere. The points p 1 , p 2 , , p n are the stereographically projected points defining a cluster, corresponding to the previously used labels s 1 , s 2 , , s n . The centroid c e is obtained with the euclidean average in R 3 . In contrast, the centroid c s is restricted to be in S 2 ( r ) and equal c e rescaled to lie on this sphere. The quantum states | ψ p 1 , | ψ p 2 , , | ψ p n are obtained after Bloch embedding the stereographically projected points p 1 , p 2 , , p n , and | ψ c is the quantum state obtained after Bloch embedding the centroid. The points marked on the Bloch sphere in Figure 4 are the Bloch vectors of the quantum states | ψ p 1 , | ψ p 2 , , | ψ p n and | ψ c .
One can observe from Definition 7 that the origin, any point p on the sphere, and | ψ p , are collinear. Hence, it can be observed that in the process of SQ-kNN clustering, the points on the stereographic sphere are projected radially into the sphere of radius 1. Once the labels were assigned in the previous iteration, the new centroid is computed, giving an integer multiple of the average point c e , which lies within the stereographic sphere. Crucially, when we embed this new centroid into the quantum state for the quantum dissimilarity calculation of the next step, since we only use the polar and azimuthal angle of the point for embedding (see Definition 7), the prepared quantum state is also projected into the surface of the Bloch sphere—or, in other words, a pure state is prepared ( | ψ c ). Hence, we can observe that all the dissimilarity calculations in the SQ-kNN will take place between points on the surface of the Bloch sphere, even though the calculated quantum centroid is contained outside the stereographic sphere. This argument also illustrates why any point on the ray O c e c s can be used for the centroid update step of the stereographic quantum kNN; any chosen point on the ray, when embedded into a quantum state for dissimilarity calculations will reduce to | ψ c .
In short, we know from Lemma 3 that O , c e , and c s lie on a straight line. Therefore, one can observe that if the Bloch sphere is scaled by r, the point on the Bloch sphere corresponding to | ψ c will transform to c s , i.e., 0 , | ψ c , c e and c s are all collinear. Equation (45) shows that SQ-kNN clustering clusters points on a sphere as per Euclidean dissimilarity; that implies that simply scaling the sphere makes no difference to the clustering. Therefore, we conclude that clustering on the surface of the stereographic sphere S 2 (2DSC-kNN) is equivalent to the quantum algorithm with stereographic embedding (SQ-kNN).

4.2. Complexity Analysis and Scaling

As we showed in Section 3.2, the time complexity of ISP for calculating Cartesian coordinates of a d -dimensional vector is O ( d ) . Hence, the total time complexity of projection for the 2DSC-kNN will be O ( ( k + N ) d ) , where N = | D | is the total number of points, and k is the total number of centroids. Since the cluster update step uses Euclidean dissimilarity, it will take O ( k N d ) time in total ( O ( d ) for each distance calculation, which is to be conducted for each pair of N points and k centroids). The centroid update expression (Equation (35)) can be calculated in O ( N d ) , making the total time for this step O ( k N d ) since we have k centroids. Hence, we have
Time complexity of 2 DSC - k NN algorithm = O ( k N d ) ,
on par with the classical k-means clustering algorithm, and at least polynomially faster than the stereographic quantum kNN (Equation (53)) or Lloyd’s quantum clustering algorithm (taking into account input assumptions) [2,4,9].

5. Experiments and Results

We defined the procedure for SQ-kNN in Section 3. Section 3.1 introduces our idea for state preparation—projecting the two-dimensional data points into a higher dimension. Section 3.1 details the hybrid quantum-classical method used for our process and then proves that the output of the quantum circuit is not only a valid but also an excellent metric that can be used for distance estimation between two points. Section 4 describes the quantum-inspired classical algorithm analogous to the quantum algorithm (2DSC-kNN). In this section, we test and compare the quantum-inspired rather the quantum algorithm for two main reasons:
  • The hardware performance and availability of quantum computers (NISQ devices) is currently so much worse than that of classical computers that no advantage can likely be obtained with the quantum algorithm.
  • The goal of this paper is not to show a “quantum advantage” in time complexity over the classical k-means in the NISQ context—it is to show that stereographic projection can lead to better learning for classical clustering and be a better embedding for quantum clustering. In particular, the equivalence between 2DSC-kNN and SQ-kNN proves that noise is the only limitation for the stereographic quantum algorithm to achieve the accuracy of the quantum-inspired algorithm.
All the experiments were carried out on a server with the following specifications: 2 Intel Xeon E5-2687W v4 chips clocked at 3.0 GHz (24 cores/48 threads), 128 GB RAM. All experiments are performed on the real-world 64-QAM data provided by Huawei (see Section 2.1 and Appendix A). Due to the extensive nature of testing and the large volume of analysis generated, we do not present all the figures in the following sections. Figures which sufficiently demonstrate general trends and observations have been included here. An exhaustive collection of all figures and other such analysis results, as well as the source code, real-world input data, and collected output data, can be accessed at [47].
The terminology used is as follows:
  • Radius: the radius of the stereographic sphere into which the two-dimensional points are projected.
  • Number of points: the number of points upon which the clustering algorithm was performed. For every experiment, the selected points were a random subset of all the 64-QAM data (of a specific noise) with cardinality equal to the required number of points. The random subset is created using the numpy.random.sample() from the Python Numpy library.
  • Number of runs: Since for each choice of parameters for each experiment we select a subset of points at random, we repeat each of the experiments many times to remove bias from the random choice and obtain stable averages and standard deviations for the collected performance parameters (described in another list below). This number of repetitions is the “number of runs”.
  • Dataset Noise: As explained in Section 2.1, data were collected for four different input powers. Data are divided into four datasets labelled with powers 2.7, 6.6, 8.6, and 10.7 dBm.
  • Natural endpoint: The natural endpoint of a clustering algorithm occurs when
    C j ( c ¯ i + 1 ) = C j ( c ¯ i ) j = 1 , , k
    i.e., when all the clusters remained unchanged (stay the same) even after the centroid update. It is the natural endpoint since if the clusters do not change, the centroids will not change either in the next iteration, leading to the same clusters (Equation (71)) and centroids for all future iterations.
The algorithms that we test are:
  • 2DSC-kNN: The quantum-analogue algorithm of Definition 12, the classical equivalent of the SQ-kNN and the most important candidate for our testing.
  • 2DEC-kNN: The standard classical kNN of Definition 4 implemented upon the original 2-dimensional dataset ( n = 2 ), which serves as a baseline for performance comparison.
  • 3DSC-kNN: The standard classical kNN, but implemented upon the stereographically projected 2-dimensional dataset, as defined in Definition 6. We again emphasise that in contrast to the 2DSC-kNN, the centroid lies within the sphere, and in contrast to the 2DEC-kNN, the clustering takes place in R 3 . This algorithm serves as another control, to gauge the relative impacts of stereographically projecting the dataset versus restricting the centroid to the surface of the sphere. It is an intermediate step between the 2DSC-kNN and the 2DEC-kNN algorithms.
From these algorithms, we measure the following performance parameters (or KPIs, Key Performance Indicators):
  • Accuracy: Since we have the true labels of the datapoints available, we can measure the accuracy of the algorithm as the percentage of points that have been given the correct label, i.e., symbol accuracy rate. All accuracies are recorded as a percentage.
  • Symbol or Bit error rate: As mentioned in Appendix A, due to Gray encoding, the bit error rate is approximately 1 6 of the symbol error rate, which in turn is simply one minus the accuracy. Although error rates are the standard performance parameter in channel coding, we decided to measure the accuracy instead, which is the standard performance parameter in machine learning.
  • Accuracy gain: The gain is calculated as (accuracy of candidate algorithm minus accuracy of two-dimensional classical k-means clustering algorithm), i.e., it is the increase in accuracy of the algorithm over the baseline, defined as the accuracy of the classical k-means clustering algorithm acting on the 2D dataset for those number of points.
  • Number of iterations: One iteration of the clustering algorithm occurs when the algorithm performs the cluster update followed by the centroid update (the algorithm must then perform the cluster update again). The number of times the algorithm repeats these two steps before stopping is the number of iterations. We use the number of iterations the algorithm requires to reach its ‘natural endpoint’ as a proxy for convergence performance. The lesser the number of iterations performed, the faster the algorithm’s convergence. The number of iterations does not directly correspond to time performance since the time taken for one iteration differs between all algorithms.
  • Iteration gain: The gain in iterations is defined as (the number of iterations of 2D k-means clustering algorithm minus the number of iterations of candidate algorithm), i.e., the gain is how many fewer iterations the candidate algorithm took than the 2DEC-kNN algorithm to reach its natural endpoint.
  • Execution time: The amount of time taken for a clustering algorithm to provide the final output (the final centroids and clusters) given the two-dimensional data points as input, i.e., the time taken end to end for the clustering process. All times in this work are recorded in milliseconds (ms).
  • Execution time gain: This gain is calculated as (the execution time of 2DEC-kNN k-means clustering algorithm minus the execution time of candidate algorithm).
  • Overfitting Parameter: The difference in testing and training accuracy.
With these algorithms and variables, we perform two main experiments:
  • The Overfitting Test: The dataset is divided into a ‘training’ and a ‘testing’ set, to characterise the clustering and classification performance of the algorithms.
  • The Stopping Criterion Test: The iterations and other performance parameters are varied, to test whether and what kind of stopping criterion is required.
We observe that the tested algorithms display some very promising and interesting results. We manage to obtain improvements in accuracy and convergence performance almost across the board, and we discover the very important optimisation parameters of the radius of projection and the stopping criterion.

5.1. Experiment 1: Overfitting

Here, the datasets were divided into training and testing data. First, a random subset of cardinality equal to the number of points was chosen from the dataset, and then 80% of the selected points were assigned as ‘Training Data’, while the other 20% was assigned as ‘Testing Data’.
In the training phase, the algorithms were first run on the training data with the maximum possible iterations set to 50 to keep an acceptable running time. The stopping criterion for all algorithms was chosen as the natural endpoint—the algorithm stopped either when the number of iterations hit 50, or when the natural endpoint was reached (whichever happened first). The final centroid co-ordinates ( c ¯ last iteration ) were recorded in the training phase, to be used for the testing phase, along with several performance parameters. The recorded performance parameters were the algorithm’s accuracy, the number of iterations taken, and the execution time.
Once the training was over, the centroids calculated at the end of training were used as the initial centroids for the testing set datapoints, and the algorithm was run with the maximum number of iterations set to 1, i.e., the calculated centroids were then used to classify the remaining points as per the dissimilarity and dataspace of each algorithm. The recorded performance parameters were the algorithm’s accuracy and execution time. Once both the testing and training accuracy had been recorded, the overfitting parameter was also recorded.
For each set of input variables (just the number of points for 2DEC-kNN clustering, the radius and number of points for the 2DSC-kNN and 3DSC-kNN clustering), the entire experiment (training and testing) was repeated 10,000 times in batches of 100 to calculate reasonable standard deviations for every performance parameter.
There are several reasons for this choice of experiment:
  • It exhaustively covers all the parameters that can be used to quantify the performance of the algorithms. We were able to observe very important trends in the performance parameters with respect to the choice of radius and the effect of the number of points (affecting the choice of when one should trigger the clustering process on the collected received points).
  • It avoids the commonly known problem of overfitting. Though this approach is not usually used in testing the kNN due to its iterative nature, we felt that from a machine learning perspective, it is useful to know how well the algorithms perform in a classification setting as well.
  • Another reason that justifies the training and testing approach (clustering and classification) is the nature of the real-world application setup. When transmitting QAM data through optical-fibre, the receiver receives only one point at a time and has to classify the received point to a given cluster in real-time using the current centroid values. Once a number of data points have accumulated, the kNN algorithm can be run to update the centroid values; after the update, the receiver will once again perform classification until some number of points has been accumulated. Hence, we can observe that in this scenario, the clustering and the classification performance of the chosen method become important.

5.1.1. Results

We begin the presentation of the results of this experiment by first showing the characterisation of the 2DSC-kNN algorithm with respect to the input variables.
Figure 5 characterises the testing and training accuracy of the 2DSC-kNN algorithm acting upon the 2.7 dBm dataset, i.e., classification and clustering performance, respectively. Figure 6 portrays the same results in the form of a heat map, with a focus on the region of interest of the algorithm. These figures are representative of the trends of all four datasets.
Figure 7 characterises the convergence performance of the quantum algorithm—it shows how the number of iterations required to reach the natural endpoint of the 2DSC-kNN algorithm varies as the number of points and radius of projection changes. Once again, the figures for all the other datasets follow the same pattern as the included figures.
We then compare the performance of the 2DSC-kNN algorithm with that of the 3DSC-kNN and 2DEC-kNN algorithms.
  • Accuracy Performance: Here, in all the following figures, the winner is chosen as the radius for which the maximum accuracy is achieved for the given number of points. Figure 8 depicts the variation in testing accuracy with the number of points for all three algorithms along with error bars. As mentioned before, this characterises the performance of the algorithms in a ‘classification’ mode, that is, when the received points must be decoded in real-time. Figure 9 portrays the trend in training accuracy with the number of points for all three algorithms along with error bars. This characterises the performance of the algorithms in ‘clustering’ mode, that is, when the received points must be used to update the centroid for future re-classification or if the received datapoints are stored and decoded in batches. Figure 8 and Figure 9 also plot the gain in testing and training accuracies respectively for the 3DSC-kNN and 2DSC-kNN algorithms. The label of the points in these figures is the radius of ISP for which that accuracy gain was achieved.
  • Iteration Performance: Here, in all the following figures, the winner is chosen as the radius for which the minimum number of iterations is achieved for the given number of points. Figure 10 shows how the required number of iterations for all three algorithms varies as the number of points increases. Figure 10 also displays the gain of the 2DSC-kNN and 3DSC-kNN algorithms in the number of iterations to reach their natural endpoints. The label of the points in these figures is the radius of ISP for achieving that iteration gain.
  • Time Performance: Here, in all the following figures, the winner is chosen as the radius for which the minimum execution time is achieved for the given number of points. Figure 11 puts forth the dependence of testing execution time upon the number of points for all three algorithms along with error bars. As mentioned before, these times are effectively the amount of time the algorithm takes for one iteration. This characterises the performance of the algorithms when performing direct classification decoding of the received points in real time.
    This figure reveals the trend in training execution time with the number of points for all three algorithms, along with error bars. This characterises the time performance of the algorithms when performing clustering—that is, when the received points must be used to update the centroid for future re-classification or if the received datapoints are stored and decoded in batches. Figure 12 plots the gains in testing and training execution times for the 3DSC-kNN and 2DSC-kNN algorithms.
  • Overfitting Performance: Figure 13 exhibits how the overfitting parameter for the 2DEC-kNN k-means clustering, 3DSC-kNN, and 2DSC-kNN algorithms vary as the number of points changes.

5.1.2. Discussion and Analysis

From Figure 5, we can observe that there is an ideal radius > 1 for which maximum accuracy is achieved. This ideal radius is usually between two and five for our datasets. For a good choice of radius (>1), the accuracy increases monotonically (with an upper bound) with the number of points. In contrast, for a poor choice of radius (<1), the accuracy nosedives as the number of points increases. This is due to the clusters getting squished together near the North pole of the stereographic sphere (the point ( 0 , 0 , r ) ). If one is dealing with a large number of points, the accuracy becomes even more sensitive to the choice of radius, as the decline in accuracy for a bad radius is much steeper as the number of points increases. These observations hold for both training and testing accuracy (classification and clustering), regardless of the noise in the dataset. These observations are also well-reflected in the heatmaps, where one can observe that the best training and testing performance is for r = 2 to 3 and the maximum number of points. It would seem that choosing too large of a radius is not too harmful. This might hold true for the classical algorithms, but when the quantum algorithm is deployed, all the points will be clustered around the South pole of the Bloch sphere and even minimal noise in the quantum circuit will degrade performance. Hence, there is a sweet spot of the radius to be chosen.
Figure 7 also shows that there is an ideal radius > 1 for which one needs the minimum number of iterations to reach the natural endpoint. This ideal radius is once again between two and five for our datasets. As the number of points increases, the number of iterations always increases. The increase is minimal for a good choice of radius, while for a bad choice, the convergence is very slow. For our experiments, we chose the maximum iterations as 50, hence the observed plateau is at 50 iterations. If one is dealing with a large number of points, the convergence becomes more sensitive to the choice of radius. The increase in iterations for a poor choice of radius is much steeper. 2DSC-kNN algorithm and 3DSC-kNN algorithm display near-identical performance.
From Figure 8, we can observe that both 2DSC-kNN algorithm and 3DSC-kNN algorithm perform better in accuracy than the 2DEC-kNN algorithm for all datasets. The advantage becomes more definitive as the number of points increases, as the increase in accuracy moves beyond the error bar. We observe the highest increase in accuracy for the 2.7 dBm dataset.
In Figure 9, one can observe the noticeably better performance of the 2DSC-kNN algorithm and 3DSC-kNN algorithm over the 2DEC-kNN algorithm for all datasets than in the testing case (classification mode). Once again, the 2.7 dBm dataset shows the maximum increase. The advantage again becomes more definitive as the number of points increases as the increase in accuracy moves beyond the error bar. The 2DSC-kNN algorithm and 3DSC-kNN algorithm show an almost identical performance.
From Figure 8 and Figure 9, we can also observe that almost universally for both algorithms, the gain is greater than 0, i.e., we beat the 2DEC-kNN algorithm in nearly every case. We can also observe that the best radius is almost always between two and five. Another observation is that the gain in training accuracy increases with the number of points. The figures further display how similarly the 3DSC-kNN algorithm and 2DSC-kNN algorithm perform in terms of accuracy, regardless of noise.
From Figure 10, it can be concluded that for low noise datasets, since the number of iterations is already quite low, there is not much gain or loss; all three algorithms perform almost identically. For high-noise datasets, however, both the 3DSC-kNN algorithm and 2DSC-kNN algorithm show significant performance improvement, especially for a higher number of points. For a high number of points, the improvement is beyond the error bars and hence very significant. It can be noticed that the ideal radius for minimum iterations is once again between two and five. Here, also, the 3DSC-kNN algorithm and 2DSC-kNN algorithm perform similarly, with the 2DSC-kNN algorithm performing better in certain cases.
One learns from Figure 11 that most importantly, the 2DSC-kNN algorithm and 2DEC-kNN algorithm take nearly the same amount of time for execution in classification mode, and the 2DSC-kNN algorithm in most cases beats the 3DSC-kNN algorithm. Here too, the gain is significant, since it is much beyond the error bar. The execution time increases linearly with the number of points, as expected. These conclusions are supported by Figure 12. Since the 2DSC-kNN algorithm takes almost the same time, and provides greater accuracy, it is an ideal candidate to replace the 2DEC-kNN algorithm for classification applications.
Figure 11 also shows that all three algorithms take almost the same amount of time for training, i.e., in clustering mode. The 3DSC-kNN and 2DSC-kNN algorithms once again perform almost identically, almost always slightly worse than 2DEC-kNN clustering. Figure 12 supports these observations. Here, execution time increases linearly with the number of points as well, as expected. In Figure 13, all three algorithms have nearly identical performance. As expected, the overfitting decreases with an increase in the number of points.

5.2. Experiment 2: Stopping Criterion

Based on the results obtained from the first experiment, we performed another experiment to see how the accuracy of the algorithms varies iteration by iteration. It was observed that the natural endpoint of the algorithm was rarely the ideal endpoint in terms of performance. Hence, we wished to observe the performance of each algorithm as the number of iterations progressed.
In this experiment, the entire random subset of datapoints was used for the clustering algorithm. The algorithms were run on the dataset, and the accuracy of the algorithms at each iteration as well as the iteration number of the natural endpoint was recorded. The maximum number of iterations was once again 50. By repeating this 100 times for each number of points (and radius, if applicable), we obtained the general performance variation of each algorithm with the iteration number. The input variables were the number of points, the radius of the stereographic sphere and the iteration number; the recorded performance parameters were the accuracy and probability of stopping.
This experiment revealed that the natural endpoint was indeed a poor choice of stopping criterion and that the endpoint should be chosen per some “loss function”. It also revealed some important trends in the performance parameters which not only emphasised the importance of the choice of radius and number of points but also provided greater insight into the disadvantages and advantages of each algorithm.

5.2.1. Results

  • Characterisation of the 2DSC-kNN algorithm: Figure 14 depicts the dependence of the accuracy of the 2DSC-kNN algorithm upon the iteration number and projection radius for the 2.7 dBm dataset. The figures for the rest of the datasets follow the same trends and are nearly identical in shape.
    Figure 15 shows the dependence of the probability of the 2DSC-kNN algorithm reaching its natural endpoint versus the radius of projection and iteration number for the 10.7 dBm dataset with 51,200 points and for the 2.7 dBm dataset with 640 points. Once again, the figures for the rest of the datasets follow the same trends and their shape can be extrapolated from the presented Figure 15.
  • Comparison with 2DEC-kNN and 3DSC-kNN Clustering: Figure 16 portrays the gain of the 2DSC-kNN and 3DSC-kNN algorithms in the number of iterations to reach maximum accuracy for the 2.7 and 10.7 dBm datasets. In these figures, a gain of ‘g’ means that the algorithm took ‘g’ fewer iterations than the classical k-means acting upon the 2D dataset did to reach maximum accuracy.
    Figure 17 plots the gain of the 2DSC-kNN and 3DSC-kNN algorithms in the maximum achieved accuracy for the 2.7 and 10.7 dBm datasets. Here, a gain of ‘g’ means that the algorithm was g % more accurate than the maximum accuracy of the classical k-means acting upon the 2D dataset.
    Lastly, Figure 18 illustrates the maximum accuracies achieved by the 2DSC-kNN, 3DSC-kNN, and 2DEC-kNN algorithms for the 2.7 and 10.7 dBm datasets.

5.2.2. Discussion and Analysis

Figure 14 shows that once again, there is an ideal radius for which maximum accuracy is achieved. The ideal projection radius is larger than one; in particular, it seems to be between two and five. Most importantly, there is an ideal number of iterations for maximum accuracy, beyond which the accuracy reduces. As the number of points increases, the sensitivity of the accuracy to radius increases significantly. For a bad choice of radius, accuracy only falls with an increase in the number of iterations and stabilises at a very low value. For a good radius, accuracy increases to a point as iterations proceed, and then stabilises at a slightly lower value. If the allowed number of iterations is restricted, the choice of radius to achieve the best results becomes extremely important. With a good radius one can achieve nearly the maximum possible accuracy with very few iterations. As mentioned before, this holds for all dataset noises. As the dataset noise increases, the iteration number at which the maximum accuracy is achieved also expectedly increases. Since accuracy always falls after a point, choosing a stopping criterion is essential rather than waiting for the algorithm to reach its natural endpoint. An idea for the stopping criterion is to record the sum of the average dissimilarity for each centroid at each iteration and stop the algorithm if that quantity increases.
Figure 15 portrays that for a good choice of radius, the 2DSC-kNN algorithm approaches convergence much faster. For r < 1 , the algorithm converges much slower or never converges. As the number of points increases, the convergence rate for the poor radius falls dramatically. For a radius greater than the ideal radius as well, the convergence rate is lower. As one would expect, the algorithm takes longer to converge as the dataset noise increases. As mentioned before, if the number of iterations is severely limited, the choice of radius becomes very important. The algorithm can reach its ideal endpoint in very few iterations if the radius is chosen well.
Through Figure 16, we observe that for lower values of noise, both algorithms do not produce much advantage in terms of iteration gain, regardless of the number of points in the dataset. However, both algorithms significantly outperform the classical one at higher noise in the dataset and a high number of points. This effect is especially significant for the 2DSC-kNN algorithm. For the highest noise and all the points, it saves over 20 iterations compared to the 2DEC-kNN algorithm—an advantage of over 50%. One of the reasons for this is that at low noises, the algorithms already perform quite well, and it is at high noise with a high number of points that the algorithm is stressed enough to reveal the difference in performance. It should be noted that these gains are much higher than when the algorithms are allowed to reach their natural endpoint, suggesting another reason for choosing an ideal stopping criterion.
Figure 17 shows that for all datasets and numbers of points, the two algorithms perform better than 2DEC-kNN clustering. The 3DSC-kNN algorithm and 2DSC-kNN algorithms perform nearly the same, and the accuracy gain seems to stabilise with an increase in the number of points. Figure 18 supports these conclusions.

5.3. Overall Observations

5.3.1. Overall observations from Experiment 1:

  • The ideal projection radius is greater than one and between two and five. At this ideal radius, one achieves maximum testing and training accuracy, and minimum iterations.
  • In general, the accuracy performance is the same for 3DSC-kNN and 2DSC-kNN algorithms—this shows a significant contribution of the ISP to the advantage as opposed to ‘quantumness’. This is a significant distinction, not made by any previous work.
  • The 2DSC-kNN and 3DSC-kNN algorithms lead to an increase in the accuracy performance in general, with the increase most pronounced for the 2.7 dBm dataset.
  • The 2DSC-kNN algorithm and 3DSC-kNN algorithm provide more iteration performance gain (fewer iterations required than 2DEC-kNN) for high noise datasets and for a large number of points.
  • Generally, increasing the number of points favours the 2DSC-kNN and 3DSC-kNN algorithms, with the caveat that a good radius must be carefully chosen.

5.3.2. Overall observations from Experiment 2:

  • These results further stress the importance of choosing a good radius (two to five in this application) and a better stopping criterion. The natural endpoint is not suitable.
  • The results justify the fact that the developed 2DSC-kNN algorithm has significant advantages over 2DEC-kNN k-means clustering and 3DSC-kNN clustering.
  • The 2DSC-kNN algorithm performs nearly the same as the 3DSC-kNN algorithm in terms of accuracy, but for iterations to achieve this max accuracy, the 2DSC-kNN algorithm is better (especially for high noise and a high number of points).
  • The developed 2DSC-kNN algorithm and 3DSC-kNN algorithm are better than the 2DEC-kNN algorithm in general—in terms of accuracy and iterations to reach that maximum accuracy.
  • The supremacy of the 2DSC-kNN algorithm over the 2DEC-kNN algorithm implies that a fully quantum SQ-kNN algorithm would have an advantage over the fully quantum k-means algorithm of [2].

6. Conclusions and Further Work

This work considers the practical case of performing kNN on experimentally acquired 64-QAM data. This work described the problem in detail and explained how the SQ-kNN and its classical analogue, the 2DSC-kNN clustering algorithm, can be used. The proposed processes and circuits, as well as the theoretical justification for the SQ-kNN quantum algorithm and the 2DSC-kNN classical algorithm, were described in detail. Finally, the simulation results on the real-world datasets were presented, along with a relevant analysis. From the analysis, one can observe that the classical analogue of the stereographic quantum kNN, the 2DSC-kNN algorithm, is something that should be considered for industrial implementation—the experiments provide a proof of concept. It also shows the importance of choosing the projection radius and provides a very useful embedding for quantum machine learning algorithms—the generalised stereographic embedding. The theoretical advantage offered by the SQ-kNN algorithm over the hybrid quantum-classical quantum k-means algorithm was also demonstrated. Another important inference from the obtained results is that the SQ-kNN algorithm offers a way to achieve the same advantage compared to the fully quantum k-means that 2DSC-kNN has over 2DEC-kNN—by using the stereographically projected quantum states. These results warrant the practical implementation and testing of both quantum and classical algorithms.
Quantum and quantum-inspired computing has the potential to change the way certain algorithms are performed, with potentially significant advantages. However, as the field is still in relative infancy, finding where quantum and quantum-inspired computing fits in practice is a challenging problem. Here, we have observed that quantum and quantum-inspired computing can indeed be applied to signal-processing scenarios and could potentially work well in the noisy quantum era as clustering algorithms that are relatively robust to noise and inaccuracy.

Future Work

One of the most important directions of future work is to experiment with more diverse datasets. More experimentation may also lead to more sophisticated methods of selecting the radius for ISP. A more detailed analysis of how to choose a radius of the projection through analytical methods is another important direction for future work. A differential geometric analysis of the effects of the ISP on a square grid provides a rough intuition of why one needs an appropriate radius. A comparison with amplitude embedding is also warranted. The ellipsoidal projection (Appendix D) is another promising and novel idea that is to be explored further. In this project, two different stopping criteria for the algorithm were proposed and revealed a change in its performance; yet there is plenty of room to explore more possible stopping criteria.
Further directions of study include improved overlap estimation methods [46] and communication scenarios where the dimensionality of the data points is greatly increased. For example, this happens when multiple carriers experience identical or at least systematically correlated phase rotations.
Another future work is to benchmark against sampling-based quantum-inspired algorithms. As part of a research analysis to evaluate the best possibilities for achieving a practical speed-up, we investigated the landscape of classical algorithms inspired by the sampling in quantum algorithms. Initially, we found that such algorithms have a theoretical complexity competing with quantum algorithms; however, only under arguably unrealistic assumptions on the structure of the classical data. As the performance of the quantum algorithms turns out to be extremely poor, this reopens the possibility that quantum-inspired algorithms can yield performance improvements while we wait for quantum computers with sufficiently low noise. Thus future work will also be a practical implementation of the quantum-inspired kNN [4], with the goal of testing the computational advantage over 2DEC-kNN, 3DSC-kNN, and 2DSC-kNN algorithms.

Author Contributions

Conceptualization, A.M. and A.V.J.; methodology, A.M. and A.V.J.; software, A.M. and A.V.J.; validation, A.M. and A.V.J.; formal analysis, A.M., A.V.J. and R.F.; investigation, A.M. and A.V.J.; resources, C.D. and F.F.; data curation, M.S.; writing—original draft preparation, A.M. and A.V.J.; writing—review and editing, A.M., A.V.J. and R.F.; visualization, A.M. and A.V.J.; supervision, R.F., C.D. and J.N.; project administration, F.F.; funding acquisition, C.D., J.N. and F.F. All authors have read and agreed to the published version of the manuscript.

Funding

This work was funded by the TUM-Huawei Joint Lab on Algorithms for Short Transmission Reach Optics (ASTRO). J.N. was funded from the DFG Emmy-Noether program under grant number NO 1129/2-1 and Munich Center for Quantum Science and Technology (MCQST). C.D., and J.N. were funded by the Federal Ministry of Education and Research of Germany in the joint project 6G-life, project identification number: 16KISK002. C.D., J.N., and A.V.J. were funded the Munich Quantum Valley (MQV) which is supported by the Bavarian state government with funds from the Hightech Agenda Bayern Plus. C.D., J.N., and R.F. were funded by the Bavarian State Ministry for Economic Affairs, Regional Development and Energy in the project 6G and Quantum Technology (6GQT). A.M., and C.D. were funded by the Federal Ministry of Education and Research of Germany in the project QR.X with the project number 16KISQ028.

Data Availability Statement

All source codes and the complete set of generated graphs are available through [47]. The datasets are published and are a property of Huawei Technologies.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

A list of abbreviations used in this manuscript can be found in the following table.
ADCAnalog-to-digital converter
CDChromatic dispersion
CFOCarrier frequency offset
CPECarrier phase estimation
DACDigital-to-analog converter
DPDual Polarisation
ECLExternal cavity laser
FECForward error correction
GHzGigahertz
GBdGigabauds
GSa/s × 10 9 samples per second
DSPDigital signal processing
ISPInverse stereographic projection
MIMOMultiple input multiple output
M-QAMM-ary Quadrature Amplitude Modulation
QMLQuantum Machine Learning
QRAMQuantum Random Access Memory
TRTiming recovery
SQ accessSample and query access
kNNk-Nearest-Neighbour clustering algorithm(Definition 3)
FF-QRAMFlip flop QRAM
NISQNear-intermediate scale quantum
D Dataspace
DDataset
DTwo-dimensional dataset
c ¯ set of all M centroids
C ( c ) Cluster associated to a centroid c
d ( · , · ) Dissimilarity (measure function)
d e ( · , · ) Euclidean dissimilarity
d s ( · , · ) Cosine dissimilarity
s r 1 ISP into a sphere of radius r
S n ( r ) n-sphere of radius r
H 2 Hilbert space of one qubit
nDEC-kNNn-dimensional Euclidean Classical kNN (Definition 4)
SQ-kNNStereographic Quantum kNN (Definition 10)
2DSC-kNNTwo-dimensional Stereographic Classical kNN (Definition 12)
3DSC-kNNThree-dimensional Stereographic Classical kNN (Definition 6)

Appendix A. QAM and Data Visualisation

Quadrature amplitude modulation (QAM) conveys multiple digital bits with each transmission by mixing amplitude and phase variations in a carrier frequency. This is conducted by changing (modulating) the amplitudes of two carrier waves. The two carrier waves (of the same frequency) are out of phase with each other by 90 ; namely, they are the sine and cosine waves of a given frequency. This condition is known as orthogonality, and the two carrier waves as quadrature. The transmitted signal is created by adding together the two carrier waves (the sine and cosine components). The two waves can be coherently separated (demodulated) at the receiver because of their orthogonality. QAM is used extensively as a modulation scheme for digital telecommunication systems, such as in 802.11 Wi-Fi standards [48]. Arbitrarily, high spectral efficiencies can be achieved with QAM by setting a suitable constellation size, limited only by the noise level and linearity of the communications channel. QAM allows us to transmit multiple bits for each time interval of the carrier symbol. The term “symbol” means some unique combination of phase and amplitude [49].
In this work, each transmitted signal corresponds to a complex number s C :
s = | s | e i ϕ ,
where | s | 2 is the initial transmission power and ϕ is the phase of s. The case shown in Equation (A1) is ideal; however, in real-world systems, noise affects the transmitted signal, distorting and scattering it in the amplitude and phase space. For our case, the received and partially processed noisy signal can be modelled as follows:
s = | s | e i ϕ + N ,
where N C is a random noise affecting the overall value of the ideal amplitude and phase. This model motivates the use of nearest neighbour clustering for cases when the noise N causes the received signal to be scattered in the vicinity of the ideal signal s.

Appendix A.1. Description of 64-QAM Data

The various datasets we collected through the setup described in Section 2.1 are visualised in this section. As mentioned, there are four datasets with launch powers of 2.7 , 6.6 , 8.6 , 10.7 dBm, corresponding to various noise levels. Each dataset consists of three variables:
  • alphabet’: The initial analog values at which the data were transmitted, in the form of complex numbers, i.e., for an entry ( a + i b ), and the transmitted signal was of the form a sin ( θ ) + b cos ( θ ) . Since the transmission protocol is 64-QAM, there are 64 values in this variable. The transmission alphabet is the same irrespective of the non-linear distortions.
  • rxsignal’: The received analog values of the signal by the receiver. These data are in the form of a 52,124 × 5 matrix. Each datapoint was transmitted five times to the receiver; thus, each row contains the values detected by the receiver during the different instances of the transmission of the same datapoint. The values in different rows represent unique datapoint values detected by the receiver.
  • bits’: This is the true label for the transmitted points. These data are in the form of a 52,124 × 6 matrix. Since the protocol is 64-QAM, each analog point represents six bits. These six bits are the entries in each column, and each value in a different row represents the correct label for a unique transmitted datapoint value. The first three bits encode the column, and the last three bits encode the row—see Figure A3.
In Appendix A.1, we can observe the analog transmission values (alphabet) for all channels. In the subsequent figure (Figure A2), we can observe the transmission data for all the iterations for each channel. The first transmission data are represented as blue crosses, the second transmission as orange circles, the third transmission as yellow dots, the fourth transmission as purple stars, and the fifth transmission as green pluses. The de-mapping alphabet is depicted in Figure A3.
One can observe from these figures that as the noise in the channel increases, the points are scattered furthered away from the initial alphabet. In addition, the non-linear noise effects also increase, causing distortion of the ‘shape’ of the data, most clearly visible in Figure A2—especially near the ‘corners’. The birefringence phase noise also increases with an increase in the channel noise, causing all the points to be ‘rotated’ about the origin.
Once the centroids have been found and the data have been clustered, as mentioned before, we need to ‘de-map’ the analog centroid values and clusters to bit-strings. For this, we need a de-mapping alphabet which maps the analog values of the alphabet to the corresponding bit strings. The de-mapping alphabet is depicted in Figure A3. It can be observed from the figure that, as in most cases, the points are Gray coded, i.e., adjacent points differ in binary translation by only 1 bit. This helps minimise the number of bit errors per symbol error in case of misclassification or exceptionally high noise. In case a point is misclassified, with the most probability, it will be assigned to a neighbouring cluster. Since the surrounding clusters differ by only 1 bit, it minimises the bit error rate. Due to Gray coding, the bit error rate is approximately 1 6 of the symbol error rate.
Figure A1. The analog alphabet (initial analog transmission values) for the data transmitted in all the channels. The real part represents the amplitude of the transmitted sine wave, and the imaginary part represents the amplitude of the cosine wave.
Figure A1. The analog alphabet (initial analog transmission values) for the data transmitted in all the channels. The real part represents the amplitude of the transmitted sine wave, and the imaginary part represents the amplitude of the cosine wave.
Entropy 25 01361 g0a1
Figure A2. The four datasets detected by the receiver with various launch powers corresponding to different noise levels: 2.7 dBm (top left), 6.6 dBm (top right), 8.6 dBm (bottom left), and 10.7 dBm (bottom right). All five iterations of transmission are depicted together.
Figure A2. The four datasets detected by the receiver with various launch powers corresponding to different noise levels: 2.7 dBm (top left), 6.6 dBm (top right), 8.6 dBm (bottom left), and 10.7 dBm (bottom right). All five iterations of transmission are depicted together.
Entropy 25 01361 g0a2
Figure A3. The bit-string mapping and demapping alphabet.
Figure A3. The bit-string mapping and demapping alphabet.
Entropy 25 01361 g0a3

Appendix B. Data Embedding

One needs data in the form of quantum states for processing in a Quantum Computer. However, due to the instability of current qubits, data can only be stored for an extended period in a classical form. Hence, the need arises to convert classical data into a quantum form. NISQ devices have a minimal number of logical qubits, which are, in addition, stable for only a limited time before they lose their quantum information to decoherence. This makes the first step in Quantum Machine Learning to load classical data by encoding it into qubits. This process is called data encoding or embedding. Classical data encoding for quantum computation plays a critical role in the overall design and performance of quantum machine learning algorithms. Table A1 summarises the various forms of data embedding.
Table A1. Summary of embeddings [2,21,50].
Table A1. Summary of embeddings [2,21,50].
EmbeddingEncodingNum. Qubits RequiredGate Depth
Basis x i i = k m b i 2 i | b m b k l = k + m per data point O ( log 2 n )
Angle x i cos ( x i ) | 0 + s i n ( x i ) | 1 O ( n ) O ( 1 )
Amplitude X i = 0 n 1 x i | i log 2 n O ( 2 n ) gates
QRAM X n = 0 n 1 1 n | i | x i log 2 n + l O ( log 2 n ) queries

Appendix B.1. Angle Embedding

Angle encoding [44,45,51] is one of the most fundamental forms of encoding classical data into a quantum state. Each data point is represented as a separate qubit. The n-th classical real number is encoded into the rotation angle of the n-th qubit. In its most basic form, this encoding requires N qubits to represent N dimensional data. It is quite cheap to prepare in terms of complexity—all that is needed is one rotation quantum gate for each qubit. This is one of the forms of encoding we have used to implement quantum kNN clustering. It is generally useful for quantum neural networks and other such QML algorithms. Angle encoding encodes N features into the rotation angles of n qubits where N n .
The rotations can be chosen as either R X ( θ ) , R Y ( θ ) or R Z ( θ ) gates. As a first step, each input data point is normalised to the interval [ 0 , π ] . To encode the data points, a rotation around the y-axis is used. The angle of rotation depends on the value of the normalised data point. This creates the following separable state:
| ψ = R Y ( x 0 ) | 0 R Y ( x 1 ) | 0 R Y ( x n ) | 0 = cos x 0 sin x 0 cos x 1 sin x 1 cos x n sin x n
It can easily be observed that one qubit is needed per data point, which is not optimal. To load the data, the rotations on the qubits can be performed in parallel; thus, the depth of the circuit is optimal [45].
The main advantage of this encoding is that it is very efficient in terms of operations—only a constant number of parallel operations are needed regardless of how many data values need to be encoded. This is not optimal from a qubit point of view (the circuit is very wide), as every input vector component requires one qubit. Another related encoding, dense angle encoding, exploits an additional property of qubits (relative phase) to use only n / 2 qubits to encode n data points. QRAM can be used to generate the more compact quantum state | i R Y ( θ i ) | 0

Appendix C. Stereographic Projection

Appendix C.1. ISP for General Radius

In this appendix, the transformations for obtaining the Cartesian coordinates of the projected point on a sphere of general radius are derived, followed by the derivation of polar and azimuthal angles of the point on the sphere. First mentioned are three conditions that the point on the sphere must satisfy, and then it follows the rest of the derivation. Refer to Figure A4 for a better understanding of the conditions and calculations.
  • Azimuthal angle of the original point and the projected point must be the same, i.e., the original point, projected point, and the top of the sphere (the point from which all projections are drawn) lie on the same plane, which is perpendicular to the 2D plane.
    s y s x = p y p x
  • The projected point lies on the sphere.
    s x 2 + s y 2 + s z 2 = r 2
  • The triangle with vertices ( 0 , 0 , r ) , ( 0 , 0 , 0 ) and ( p x , p y , 0 ) is similar to the triangle with vertices ( 0 , 0 , r ) , ( 0 , 0 , s z ) and ( s x , s y , s z ) :
    p x 2 + p y 2 r = s x 2 + s y 2 r s z .
Using Equations (A3) and (A5), we obtain
s x = p x · 1 s z r and s y = p y · 1 s z r
Substituting in Equation (A4) we obtain,
s z = r p x 2 + p y 2 r 2 p x 2 + p y 2 + r 2
Hence, one obtains the set of transformations:
s x = p x 2 r 2 p x 2 + p y 2 + r 2
s y = p y 2 r 2 p x 2 + p y 2 + r 2
s z = r p x 2 + p y 2 r 2 p x 2 + p y 2 + r 2
Calculating the polar angle ( θ ) and azimuthal angle ( ϕ ):
ϕ = tan 1 s y s x ϕ = tan 1 p y p x tan π θ 2 = p x 2 + p y 2 r θ = 2 · tan 1 r p x 2 + p y 2
Figure A4. ISP for a sphere of radius r. In this figure, the plane is the plane perpendicular to the XY plane and has angle ϕ with respect to the x-axis, i.e., the plane containing the 2D point and its projection.
Figure A4. ISP for a sphere of radius r. In this figure, the plane is the plane perpendicular to the XY plane and has angle ϕ with respect to the x-axis, i.e., the plane containing the 2D point and its projection.
Entropy 25 01361 g0a4

Appendix C.2. Equivalence of Displacement and Scaling

Refer to Figure A5. Here, N is the north pole from which all the projections originate; O and O are the centres of the projection spheres of radius r and r , respectively; P is the point on the 2D plane to be projected; and s and S are the ISP of P on the spheres of radius r, centre O , radius r , and centre O , respectively.
| O N ¯ | = | O s ¯ | = r O N s = O s N = θ N O s = π 2 θ
Moreover,
| O N ¯ | = | O s ¯ | = r O N s = O s N = θ N O s = π 2 θ
Hence
N O s = N O s = π 2 θ
Since both s and s lie on the same plane, which is a vertical cross-section of the sphere (plane perpendicular to the data plane and passing through the centre of both stereographic spheres), the azimuthal angle of both points is equal ( ϕ = tan 1 p y p x ).
Hence, one can observe that the azimuthal and the polar angle generated by ISP on a sphere of radius r displaced above the 2D plane containing the points by ( 1 + δ ) r is the same as the azimuthal and the polar angle generated by ISP on a sphere of radius ( 1 + δ ) r centred at origin. This reduces the effective number of parameters that can be chosen for the embedding.
Figure A5. Two ISPs, one on a sphere with centre displaced above the plane (in blue) and a sphere centred at origin (in black) both with the same north pole N. The orange line is the common projection line between the two ISPs. From the figure it is clear that the resulting angle is the same.
Figure A5. Two ISPs, one on a sphere with centre displaced above the plane (in blue) and a sphere centred at origin (in black) both with the same north pole N. The orange line is the common projection line between the two ISPs. From the figure it is clear that the resulting angle is the same.
Entropy 25 01361 g0a5

Appendix D. Ellipsoidal Embedding

Here, we first derive the transformations for obtaining the Cartesian coordinates of the projected point on a general ellipsoid, followed by the derivation of polar and azimuthal angles for the point on the ellipsoid. First mentioned are three conditions that the point on the sphere must satisfy, and then it follows the rest of the derivation. Refer to Figure A6 for a better understanding of the conditions and calculations.
Figure A6. Ellipsoidal Projection—a generalisation of the ISP.
Figure A6. Ellipsoidal Projection—a generalisation of the ISP.
Entropy 25 01361 g0a6
  • The azimuthal angle is unchanged, thus
    s y s x = p y p x
  • The projected point lies on the ellipsoid, thus
    s x 2 a 2 + s y 2 b 2 + s z 2 c 2 = 1
  • The triangle with vertices ( 0 , 0 , c ) , ( 0 , 0 , 0 ) and ( p x , p y , 0 ) is similar to the triangle with vertices ( 0 , 0 , c ) , ( 0 , 0 , s z ) and ( s x , s y , s z ) , thus
    p x 2 + p y 2 c = s x 2 + s y 2 c s z .
From the above conditions, we have
s x = p x · 1 s z c and s y = p y · 1 s z c
Substituting as before, we obtain
s z = c · p x 2 a 2 + p y 2 b 2 1 p x 2 a 2 + p y 2 b 2 + 1
Hence, one obtains the set of transformations:
s x = p x 2 p x 2 a 2 + p y 2 b 2 + 1
s y = p y 2 p x 2 a 2 + p y 2 b 2 + 1
s z = c p x 2 a 2 + p y 2 b 2 1 p x 2 a 2 + p y 2 b 2 + 1
From Figure A6, one can observe that
tan ( π θ ) = s x 2 + s y 2 s z θ = tan 1 2 p x 2 + p y 2 c p x 2 a 2 + p y 2 b 2 1 Moreover , as before , by the same reasoning ϕ = tan 1 p y p x
Now that we have these expressions, we have two methods of encoding the datapoint. We can either encode it as before, using the unitary U ( θ , ϕ ) , which would correspond to projecting all the points on the ellipsoid to the surface of the sphere radially, or we could use mixed states to represent the points on the surface of the ellipsoid after rescaling it to lie within the Bloch sphere.

Appendix E. Distance Estimation Using Stereographic Embedding

In our proposed quantum algorithm (the SQ-kNN algorithm), we project the original two-dimensional points into the stereographic sphere before converting them into quantum states using angle embedding and then estimating the overlap with the Bell state measurement circuit. Due to the many steps of this procedure, it is insightful to calculate the final output in terms of the original input, the two-dimensional datapoints. It also serves as a helpful point of comparison with the 2DEC-kNN algorithm, where the Euclidean dissimilarity between the two-dimensional points is used for classification.
Recall Equation (40) from Section 2.7. Then concatenating the ISP with the cosine dissimilarity, we obtain a new dissimilarity, as follows.
As mentioned before, we begin with two 2D points p 1 = x 1 y 1 , p 2 = x 2 y 2 and compute
d s s r 1 ( p 1 , p 2 ) : = d s s r 1 ( p 1 ) , s r 1 ( p 2 ) .
To calculate this, we compute s r 1 ( p 1 ) · s r 1 ( p 2 ) using Equation (40):
s r 1 ( p 1 ) · s r 1 ( p 2 ) = 4 r 4 p 1 · p 2 ( p 1 2 + r 2 ) ( p 2 2 + r 2 ) + r 2 ( p 1 2 r 2 ) ( p 2 2 r 2 ) ( p 1 2 + r 2 ) ( p 2 2 + r 2 )
Combining, we have:
d s s r 1 ( p 1 , p 2 ) = 1 1 r 2 s r 1 ( p 1 ) · s 1 ( p 2 ) = ( p 1 2 + r 2 ) ( p 2 2 + r 2 ) 4 r 2 p 1 · p 2 ( p 1 2 + r 2 ) ( p 2 2 + r 2 ) ( p 1 2 r 2 ) ( p 2 2 r 2 ) ( p 1 2 + r 2 ) ( p 2 2 + r 2 ) = 2 r 2 p 1 2 + 2 r 2 p 2 2 4 r 2 p 1 · p 2 ( p 1 2 + r 2 ) ( p 2 2 + r 2 ) = 2 r 2 p 1 p 2 2 ( p 1 2 + r 2 ) ( p 2 2 + r 2 ) = d e ( p 1 , p 2 ) · 2 r 2 ( r 2 + p 1 2 ) ( r 2 + p 2 2 )
where d e is the Euclidean dissimilarity from Equation (19).
It is illustrative to pick the point ( 0 , 0 ) (origin) and observe how this function varies as the other point p varies. In this case, we have:
1 2 d s s r 1 ( 0 , p ) = r 2 p 2 ( r 2 + p 2 ) ( r 2 ) = p 2 r 2 + p 2 = 1 r 2 r 2 + p 2
For Figure A7, the radius of the stereographic sphere is assumed to be one. Hence, the quantum dissimilarity reduces to:
1 2 d s s r 1 ( 0 , p ) = 1 1 1 + p 2
For Figure A7, the radius of the stereographic sphere is assumed to be 2 and 0.5 , respectively.
Figure A7. r = 1 (left) r = 2 (middle) r = 0.5 (right).
Figure A7. r = 1 (left) r = 2 (middle) r = 0.5 (right).
Entropy 25 01361 g0a7

Appendix F. Rotation Gates and the UGate

The complete expression for the unitary UGate in Qiskit is as follows:
U ( θ , ϕ , λ ) : = cos θ 2 e i λ sin θ 2 e i ϕ sin θ 2 e i ( ϕ + λ ) cos θ 2
Using the | 0 state for encoding, we have
U ( θ , ϕ , λ ) | 0 = cos θ 2 e i λ sin θ 2 e i ϕ sin θ 2 e i ( ϕ + λ ) cos θ 2 1 0 = cos θ 2 e i ϕ sin θ 2
One can observe that this state is not dependent on λ . On the other hand, if we use the state | 1 for encoding, we have
U ( θ , ϕ , λ ) | 1 = cos θ 2 e i λ sin θ 2 e i ϕ sin θ 2 e i ( ϕ + λ ) cos θ 2 0 1 = e i λ sin θ 2 e i ( ϕ + λ ) cos θ 2 = e i λ sin θ 2 e i ϕ cos θ 2
One can observe that the λ term leads only to a global phase. A global phase will not affect the observable outcome of the SWAP test or Bell-state measurement (due to the modulus operator)—hence, once again, no information can be encoded into the quantum state using λ .
For constructing the point ( θ , ϕ ) on the Bloch sphere, we can use rotation gates as well:
( θ , ϕ ) : = R Z ( ϕ ) R Y ( θ ) | 0 = e i ϕ 2 0 0 e i ϕ 2 cos θ 2 sin θ 2 sin θ 2 cos θ 2 1 0 = e i ϕ 2 0 0 e i ϕ 2 cos θ 2 sin θ 2 = e i ϕ 2 cos θ 2 e i ϕ 2 sin θ 2 = e i ϕ 2 cos θ 2 e i ϕ sin θ 2 = e i ϕ 2 U ( θ , ϕ ) | 0
From Equation (A21) one can observe that R Z ( ϕ ) R Y ( θ ) | 0 and U ( θ , ϕ ) | 0 only differ by a global phase ( e i ϕ 2 ). Hence, the R Z ( ϕ ) R Y ( θ ) and U ( θ , ϕ ) operations can be used interchangeably for state preparation, since a global phase will not affect the observable result of the SWAP test.

References

  1. Harrow, A.W.; Hassidim, A.; Lloyd, S. Quantum Algorithm for Linear Systems of Equations. Phys. Rev. Lett. 2009, 103, 150502. [Google Scholar] [CrossRef] [PubMed]
  2. Lloyd, S.; Mohseni, M.; Rebentrost, P. Quantum algorithms for supervised and unsupervised machine learning. arXiv 2013, arXiv:1307.0411. [Google Scholar]
  3. Preskill, J. Quantum Computing in the NISQ era and beyond. Quantum 2018, 2, 79. [Google Scholar] [CrossRef]
  4. Tang, E. Quantum Principal Component Analysis Only Achieves an Exponential Speedup Because of Its State Preparation Assumptions. Phys. Rev. Lett. 2021, 127, 060503. [Google Scholar] [CrossRef]
  5. Arute, F.; Arya, K.; Babbush, R.; Bacon, D.; Bardin, J.; Barends, R.; Biswas, R.; Boixo, S.; Brandao, F.; Buell, D.; et al. Quantum Supremacy using a Programmable Superconducting Processor. Nature 2019, 574, 505–510. [Google Scholar] [CrossRef]
  6. Schuld, M.; Petruccione, F. Supervised Learning with Quantum Computers; Quantum Science and Technology; Springer International Publishing: Cham, Switzerland, 2018. [Google Scholar] [CrossRef]
  7. Schuld, M.; Sinayskiy, I.; Petruccione, F. An introduction to quantum machine learning. Contemp. Phys. 2015, 56, 172–185. [Google Scholar] [CrossRef]
  8. Kerenidis, I.; Prakash, A. Quantum Recommendation Systems. In Proceedings of the 8th Innovations in Theoretical Computer Science Conference (ITCS 2017), Berkeley, CA, USA, 9–11 January 2017; Leibniz International Proceedings in Informatics (LIPIcs). Papadimitriou, C.H., Ed.; Schloss Dagstuhl–Leibniz-Zentrum fuer Informatik: Dagstuhl, Germany, 2017; Volume 67, pp. 49:1–49:21. [Google Scholar] [CrossRef]
  9. Kerenidis, I.; Landman, J.; Luongo, A.; Prakash, A. q-means: A quantum algorithm for unsupervised machine learning. In Proceedings of the Advances in Neural Information Processing Systems (NeurIPS 2019), Vancouver Convention Center, Vancouver, BC, Canada, 8–14 December 2019; Wallach, H., Larochelle, H., Beygelzimer, A., d’Alché-Buc, F., Fox, E., Garnett, R., Eds.; Curran Associates, Inc.: New York, NY, USA, 2019; Volume 32. Available online: https://proceedings.neurips.cc/paper/2019/hash/16026d60ff9b54410b3435b403afd226-Abstract.html (accessed on 12 September 2023).
  10. Modi, A.; Jasso, A.V.; Ferrara, R.; Deppe, C.; Noetzel, J.; Fung, F.; Schaedler, M. Testing of Hybrid Quantum-Classical k-means for Nonlinear Noise Mitigation. arXiv 2023, arXiv:2308.03540. [Google Scholar]
  11. Pakala, L.; Schmauss, B. Non-linear mitigation using carrier phase estimation and k-means clustering. In Proceedings of the Photonic Networks; 16. ITG Symposium, Leipzig, Germany, 7–8 May 2015; VDE Verlag: Berlin, Germany, 2015; pp. 1–5. Available online: https://ieeexplore.ieee.org/document/7110093 (accessed on 12 September 2023).
  12. Zhang, J.; Chen, W.; Gao, M.; Shen, G. k-means-clustering-based fiber nonlinearity equalization techniques for 64-QAM coherent optical communication system. Opt. Express 2017, 25, 27570–27580. [Google Scholar] [CrossRef]
  13. Gambetta, J. IBM’s Roadmap for Scaling Quantum Technology. Available online: https://research.ibm.com/blog/ibm-quantum-roadmap-2025 (accessed on 12 September 2023).
  14. Diedolo, F.; Böcherer, G.; Schädler, M.; Calabró, S. Nonlinear Equalization for Optical Communications Based on Entropy-Regularized Mean Square Error. In Proceedings of the European Conference on Optical Communication (ECOC) 2022, Basel, Switzerland, 18–22 September 2022; Optica Publishing Group: Washington, DC, USA, 2022; p. We2C.2. Available online: https://ieeexplore.ieee.org/document/9979330 (accessed on 12 September 2023).
  15. Martyn, J.M.; Rossi, Z.M.; Tan, A.K.; Chuang, I.L. Grand unification of quantum algorithms. PRX Quantum 2021, 2, 040203. [Google Scholar] [CrossRef]
  16. Kopczyk, D. Quantum machine learning for data scientists. arXiv 2018, arXiv:1804.10068. [Google Scholar]
  17. Esma Aimeur, G.B.; Gambs, S. Quantum clustering algorithms. In Proceedings of the ICML’07: 24th International Conference on Machine Learning, Corvallis, OR, USA, 20–24 June 2007; pp. 1–8. Available online: https://icml.cc/imls/conferences/2007/proceedings/papers/518.pdf (accessed on 12 September 2023).
  18. Cruise, J.R.; Gillespie, N.I.; Reid, B. Practical Quantum Computing: The value of local computation. arXiv 2020, arXiv:2009.08513. [Google Scholar]
  19. Johri, S.; Debnath, S.; Mocherla, A.; Singh, A.; Prakash, A.; Kim, J.; Kerenidis, I. Nearest centroid classification on a trapped ion quantum computer. Npj Quantum Inf. 2021, 7, 122. [Google Scholar] [CrossRef]
  20. Khan, S.U.; Awan, A.J.; Vall-Llosera, G. k-means Clustering on Noisy Intermediate Scale Quantum Computers. arXiv 2019, arXiv:1909.12183. [Google Scholar]
  21. Cortese, J.A.; Braje, T.M. Loading classical data into a quantum computer. arXiv 2018, arXiv:1803.01958. [Google Scholar]
  22. Giovannetti, V.; Lloyd, S.; Maccone, L. Quantum Random Access Memory. Phys. Rev. Lett. 2008, 100, 160501. [Google Scholar] [CrossRef] [PubMed]
  23. Buhrman, H.; Cleve, R.; Watrous, J.; De Wolf, R. Quantum fingerprinting. arXiv 2001, arXiv:quant-ph/0102001. [Google Scholar] [CrossRef]
  24. Ripper, P.; Amaral, G.; Temporão, G. Swap Test-based characterization of decoherence in universal quantum computers. Quantum Inf. Process. 2023, 22, 220. [Google Scholar] [CrossRef]
  25. Foulds, S.; Kendon, V.; Spiller, T. The controlled SWAP test for determining quantum entanglement. Quantum Sci. Technol. 2021, 6, 035002. [Google Scholar] [CrossRef]
  26. Tang, E. A Quantum-Inspired Classical Algorithm for Recommendation Systems. In Proceedings of the STOC 2019—51st Annual ACM SIGACT Symposium on Theory of Computing, Phoenix, AZ, USA, 23–26 June 2019; Association for Computing Machinery: New York, NY, USA, 2019; pp. 217–228. [Google Scholar] [CrossRef]
  27. Chia, N.H.; Lin, H.H.; Wang, C. Quantum-inspired sublinear classical algorithms for solving low-rank linear systems. arXiv 2018, arXiv:1811.04852. [Google Scholar]
  28. Gilyén, A.; Lloyd, S.; Tang, E. Quantum-inspired low-rank stochastic regression with logarithmic dependence on the dimension. arXiv 2018, arXiv:1811.04909. [Google Scholar]
  29. Arrazola, J.M.; Delgado, A.; Bardhan, B.R.; Lloyd, S. Quantum-inspired algorithms in practice. Quantum 2020, 4, 307. [Google Scholar] [CrossRef]
  30. Chia, N.H.; Gilyén, A.; Li, T.; Lin, H.H.; Tang, E.; Wang, C. Sampling-based sublinear low-rank matrix arithmetic framework for dequantizing Quantum machine learning. In Proceedings of the 52nd Annual ACM SIGACT Symposium on Theory of Computing, Chicago, IL, USA, 22–26 June 2020. [Google Scholar] [CrossRef]
  31. Chia, N.H.; Gilyén, A.; Lin, H.H.; Lloyd, S.; Tang, E.; Wang, C. Quantum-Inspired Algorithms for Solving Low-Rank Linear Equation Systems with Logarithmic Dependence on the Dimension. In Proceedings of the 31st International Symposium on Algorithms and Computation (ISAAC 2020), Leibniz International Proceedings in Informatics (LIPIcs). Hong Kong, China, 14–18 December 2020; Cao, Y., Cheng, S.W., Li, M., Eds.; Schloss Dagstuhl–Leibniz-Zentrum für Informatik: Dagstuhl, Germany, 2020; Volume 181, pp. 47:1–47:17. [Google Scholar] [CrossRef]
  32. Sergioli, G.; Santucci, E.; Didaci, L.; Miszczak, J.A.; Giuntini, R. A quantum-inspired version of the nearest mean classifier. Soft Comput. 2018, 22, 691–705. Available online: https://link.springer.com/article/10.1007/s00500-016-2478-2 (accessed on 12 September 2023). [CrossRef]
  33. Sergioli, G.; Bosyk, G.M.; Santucci, E.; Giuntini, R. A quantum-inspired version of the classification problem. Int. J. Theor. Phys. 2017, 56, 3880–3888. Available online: https://link.springer.com/article/10.1007/s10773-017-3371-1 (accessed on 12 September 2023). [CrossRef]
  34. Subhi, G.M.; Messikh, A. Simple quantum circuit for pattern recognition based on nearest mean classifier. Int. J. Perceptive Cogn. Comput. 2016, 2. [Google Scholar] [CrossRef]
  35. Nguemto, S.; Leyton-Ortega, V. Re-QGAN: An optimized adversarial quantum circuit learning framework. arXiv 2022, arXiv:2208.02165. [Google Scholar] [CrossRef]
  36. Eybpoosh, K.; Rezghi, M.; Heydari, A. Applying inverse stereographic projection to manifold learning and clustering. Appl. Intell. 2022, 52, 4443–4457. [Google Scholar] [CrossRef]
  37. Poggiali, A.; Berti, A.; Bernasconi, A.; Del Corso, G.; Guidotti, R. Quantum Clustering with k-means: A Hybrid Approach. arXiv 2022, arXiv:2212.06691. [Google Scholar]
  38. de Veras, T.M.L.; de Araujo, I.C.S.; Park, D.K.; da Silva, A.J. Circuit-Based Quantum Random Access Memory for Classical Data With Continuous Amplitudes. IEEE Trans. Comput. 2021, 70, 2125–2135. [Google Scholar] [CrossRef]
  39. Hornik, K.; Feinerer, I.; Kober, M.; Buchta, C. Spherical k-means Clustering. J. Stat. Softw. 2012, 50, 1–22. [Google Scholar] [CrossRef]
  40. Feng, C.; Zhao, B.; Zhou, X.; Ding, X.; Shan, Z. An Enhanced Quantum K-Nearest Neighbor Classification Algorithm Based on Polar Distance. Entropy 2023, 25, 127. [Google Scholar] [CrossRef]
  41. Nielsen, M.A.; Chuang, I.L. Quantum Computation and Quantum Information; Cambridge University Press: Cambridge, UK, 2010. [Google Scholar] [CrossRef]
  42. Lloyd, S. Least squares quantization in PCM. IEEE Trans. Inf. Theory 1982, 28, 129–137. [Google Scholar] [CrossRef]
  43. Schubert, E.; Lang, A.; Feher, G. Accelerating Spherical k-means. In Similarity Search and Applications, Proceedings of the 14th International Conference, SISAP 2021, Dortmund, Germany, 29 September–1 October 2021; Reyes, N., Connor, R., Kriege, N., Kazempour, D., Bartolini, I., Schubert, E., Chen, J.J., Eds.; Springer International Publishing: Cham, Switzerland, 2021; pp. 217–231. [Google Scholar] [CrossRef]
  44. LaRose, R.; Coyle, B. Robust data encodings for quantum classifiers. Phys. Rev. A 2020, 102, 032420. [Google Scholar] [CrossRef]
  45. Weigold, M.; Barzen, J.; Leymann, F.; Salm, M. Expanding Data Encoding Patterns For Quantum Algorithms. In Proceedings of the 2021 IEEE 18th International Conference on Software Architecture Companion (ICSA-C), Stuttgart, Germany, 22–26 March 2021; pp. 95–101. [Google Scholar] [CrossRef]
  46. Fanizza, M.; Rosati, M.; Skotiniotis, M.; Calsamiglia, J.; Giovannetti, V. Beyond the Swap Test: Optimal Estimation of Quantum State Overlap. Phys. Rev. Lett. 2020, 124, 060503. [Google Scholar] [CrossRef] [PubMed]
  47. Viladomat Jasso, A.; Modi, A.; Ferrara, R.; Deppe, C.; Nötzel, J.; Fung, F.; Schädler, M. Stereographic Quantum Embedding Clustering, September 2023. Available online: https://github.com/AlonsoViladomat/Stereographic-quantum-embedding-clustering (accessed on 12 September 2023).
  48. IEEE Std 802.11-2020 (Revision of IEEE Std 802.11-2016); IEEE Standard for Information Technology—Telecommunications and Information Exchange between Systems—Local and Metropolitan Area Networks–Specific Requirements—Part 11: Wireless LAN Medium Access Control (MAC) and Physical Layer (PHY) Specifications. IEEE: Piscataway, NJ, USA, 2021; pp. 1–4379. [CrossRef]
  49. Frenzel, L.E., Jr. Electronics Explained: Fundamentals for Engineers, Technicians, and Makers; Newnes: Lithgow, Australia, 2018; Available online: https://www.sciencedirect.com/book/9780128116418 (accessed on 12 September 2023).
  50. Plesch, M.; Brukner, Č. Quantum-state preparation with universal gate decompositions. Phys. Rev. A 2011, 83, 032302. [Google Scholar] [CrossRef]
  51. Leymann, F.; Barzen, J. The bitter truth about gate-based quantum algorithms in the NISQ era. Quantum Sci. Technol. 2020, 5, 044007. [Google Scholar] [CrossRef]
Figure 1. Experimental setup over a 80 km G.652 fibre link at optimal launch power of 6.6 dBm. Chromatic disperion (CD) and carrier frequency offset (CFO) compensation, multiple-input multiple-output (MIMO) equalizer, timing recovery (TR) and carrier phase estimation (CPE) [10,14]. The red arrows distinguish the path of the laser from electrical signals.
Figure 1. Experimental setup over a 80 km G.652 fibre link at optimal launch power of 6.6 dBm. Chromatic disperion (CD) and carrier frequency offset (CFO) compensation, multiple-input multiple-output (MIMO) equalizer, timing recovery (TR) and carrier phase estimation (CPE) [10,14]. The red arrows distinguish the path of the laser from electrical signals.
Entropy 25 01361 g001
Figure 2. Quantum circuit of the Bell-state measurement. The measurement is obtained by first transforming the Bell basis into the standard basis with ( H 1 ) C N O T and then measuring in the standard basis.
Figure 2. Quantum circuit of the Bell-state measurement. The measurement is obtained by first transforming the Bell basis into the standard basis with ( H 1 ) C N O T and then measuring in the standard basis.
Entropy 25 01361 g002
Figure 3. Inverse Stereographic Projection (ISP) with radius r. The figure on the right is the cut of the figure in the left going through the N, p and the origin.
Figure 3. Inverse Stereographic Projection (ISP) with radius r. The figure on the right is the cut of the figure in the left going through the N, p and the origin.
Entropy 25 01361 g003
Figure 4. A diagram providing a visual intuition for how the stereographic quantum kNN (SQ-kNN) is equivalent to the 2DSC-kNN. The blue points are the projections of the 2-dimensional datapoints and their corresponding Euclidean centroid, the red points are their corresponding spherical centroid and the Bloch of its quantum state.
Figure 4. A diagram providing a visual intuition for how the stereographic quantum kNN (SQ-kNN) is equivalent to the 2DSC-kNN. The blue points are the projections of the 2-dimensional datapoints and their corresponding Euclidean centroid, the red points are their corresponding spherical centroid and the Bloch of its quantum state.
Entropy 25 01361 g004
Figure 5. Mean accuracy vs. Number of points vs. Projection radius for the 2DSC-kNN algorithm acting upon the 2.7 dBm dataset-Testing (top left), close up of testing (top right), Training (bottom left), close up of training (bottom right).
Figure 5. Mean accuracy vs. Number of points vs. Projection radius for the 2DSC-kNN algorithm acting upon the 2.7 dBm dataset-Testing (top left), close up of testing (top right), Training (bottom left), close up of training (bottom right).
Entropy 25 01361 g005
Figure 6. Heat map of Mean accuracy (%) vs. Number of points vs. Projection radius for the 2DSC-kNN algorithm acting upon the 2.7 dBm dataset. Testing on the (left) and training on the (right).
Figure 6. Heat map of Mean accuracy (%) vs. Number of points vs. Projection radius for the 2DSC-kNN algorithm acting upon the 2.7 dBm dataset. Testing on the (left) and training on the (right).
Entropy 25 01361 g006
Figure 7. Mean number of iterations in training vs. Number of points vs. projection radius for the 2DSC-kNN algorithm acting upon the 10.7 dBm dataset. Full data (top), close-up (bottom).
Figure 7. Mean number of iterations in training vs. Number of points vs. projection radius for the 2DSC-kNN algorithm acting upon the 10.7 dBm dataset. Full data (top), close-up (bottom).
Entropy 25 01361 g007
Figure 8. Maximum mean testing accuracy and testing accuracy gain among all tested radii vs. number of points-accuracy for 2.7 dBm (top left), accuracy for 10.7 dBm (top right), accuracy gain for 2.7 dBm (bottom left), accuracy gain for 10.7 dBm (bottom right).
Figure 8. Maximum mean testing accuracy and testing accuracy gain among all tested radii vs. number of points-accuracy for 2.7 dBm (top left), accuracy for 10.7 dBm (top right), accuracy gain for 2.7 dBm (bottom left), accuracy gain for 10.7 dBm (bottom right).
Entropy 25 01361 g008
Figure 9. Maximum mean training accuracy and training accuracy gain among all tested radii vs. number of points-accuracy for 2.7 dBm (top left), accuracy for 10.7 dBm (top right), accuracy gain for 2.7 dBm (bottom left), accuracy gain for 10.7 dBm (bottom right).
Figure 9. Maximum mean training accuracy and training accuracy gain among all tested radii vs. number of points-accuracy for 2.7 dBm (top left), accuracy for 10.7 dBm (top right), accuracy gain for 2.7 dBm (bottom left), accuracy gain for 10.7 dBm (bottom right).
Entropy 25 01361 g009
Figure 10. Mean training iterations and iteration gain among all tested radii vs. number of points-iterations for 2.7 dBm (top left), iterations for 10.7 dBm (top right), iteration gain for 2.7 dBm (bottom left), iteration gain for 10.7 dBm (bottom right).
Figure 10. Mean training iterations and iteration gain among all tested radii vs. number of points-iterations for 2.7 dBm (top left), iterations for 10.7 dBm (top right), iteration gain for 2.7 dBm (bottom left), iteration gain for 10.7 dBm (bottom right).
Entropy 25 01361 g010
Figure 11. Best mean execution time among all tested radii vs. number of points-testing time for 2.7 dBm (top left), testing time for 10.7 dBm (top right), training time for 2.7 dBm (bottom left), training time for 10.7 dBm (bottom right).
Figure 11. Best mean execution time among all tested radii vs. number of points-testing time for 2.7 dBm (top left), testing time for 10.7 dBm (top right), training time for 2.7 dBm (bottom left), training time for 10.7 dBm (bottom right).
Entropy 25 01361 g011
Figure 12. Best mean execution time gain among all tested radii vs. number of points-testing time gain for 2.7 dBm (top left), testing time gain for 10.7 dBm (top right), training time gain for 2.7 dBm (bottom left), training time gain for 10.7 dBm (bottom right).
Figure 12. Best mean execution time gain among all tested radii vs. number of points-testing time gain for 2.7 dBm (top left), testing time gain for 10.7 dBm (top right), training time gain for 2.7 dBm (bottom left), training time gain for 10.7 dBm (bottom right).
Entropy 25 01361 g012
Figure 13. Mean overfitting parameter vs. number of points for the 2.7 dBm (left) and 10.7 dBm (right) datasets.
Figure 13. Mean overfitting parameter vs. number of points for the 2.7 dBm (left) and 10.7 dBm (right) datasets.
Entropy 25 01361 g013
Figure 14. Maximum Accuracy vs. iteration number vs. Projection radius for the 2DSC-kNN algorithm acting upon the 2.7 dBm dataset. Close-up of the maximas at the bottom.
Figure 14. Maximum Accuracy vs. iteration number vs. Projection radius for the 2DSC-kNN algorithm acting upon the 2.7 dBm dataset. Close-up of the maximas at the bottom.
Entropy 25 01361 g014aEntropy 25 01361 g014b
Figure 15. The probability of stopping vs. projection radius vs. iteration number for 2DSC-kNN algorithm. (Top) 2.7 dBm dataset with the number of points = 640. (Bottom) 10.7 dBm dataset with the number of points = 51,200.
Figure 15. The probability of stopping vs. projection radius vs. iteration number for 2DSC-kNN algorithm. (Top) 2.7 dBm dataset with the number of points = 640. (Bottom) 10.7 dBm dataset with the number of points = 51,200.
Entropy 25 01361 g015
Figure 16. Gain in iteration number at maximum accuracy (number of iterations at maximum accuracy of 2DEC-kNN minus the number of iterations at maximum accuracy of 3DSC-kNN (blue) and 2DSC-kNN (red)) vs. number of points for the 2.7 dBm (left) and 10.7 dBm (right) datasets.
Figure 16. Gain in iteration number at maximum accuracy (number of iterations at maximum accuracy of 2DEC-kNN minus the number of iterations at maximum accuracy of 3DSC-kNN (blue) and 2DSC-kNN (red)) vs. number of points for the 2.7 dBm (left) and 10.7 dBm (right) datasets.
Entropy 25 01361 g016
Figure 17. Gain in maximum accuracy of 2DSC-kNN (red) and 3DSC-kNN (blue) algorithms vs. number of points for the 2.7 dBm (left) and 10.7 dBm (right) datasets.
Figure 17. Gain in maximum accuracy of 2DSC-kNN (red) and 3DSC-kNN (blue) algorithms vs. number of points for the 2.7 dBm (left) and 10.7 dBm (right) datasets.
Entropy 25 01361 g017
Figure 18. Maximum accuracy of 2DSC-kNN (red), 3DSC-kNN (blue) and 2DEC-kNN (yellow) algorithms vs. number of points for the 2.7 dBm (left) and 10.7 dBm (right) datasets.
Figure 18. Maximum accuracy of 2DSC-kNN (red), 3DSC-kNN (blue) and 2DEC-kNN (yellow) algorithms vs. number of points for the 2.7 dBm (left) and 10.7 dBm (right) datasets.
Entropy 25 01361 g018
Table 1. Summary of various kNN algorithms, where SC stands for “stereographic classical” (2D or 3D) and SQ for “stereographic quantum”. Here, D is the two-dimensional dataset, c ¯ are the 2-dimensional initial centroids (initial transmission points), s r 1 is the ISP into S 2 ( r ) , and the dissimilarities d e , d s and d q are defined in Equation (19) and Definitions 5 and 8, respectively. The option of using d e instead of d s in the 2DSC-kNN is due to Remark 3.
Table 1. Summary of various kNN algorithms, where SC stands for “stereographic classical” (2D or 3D) and SQ for “stereographic quantum”. Here, D is the two-dimensional dataset, c ¯ are the 2-dimensional initial centroids (initial transmission points), s r 1 is the ISP into S 2 ( r ) , and the dissimilarities d e , d s and d q are defined in Equation (19) and Definitions 5 and 8, respectively. The option of using d e instead of d s in the 2DSC-kNN is due to Remark 3.
AlgorithmReferenceDatasetInitial CentroidsDataspaceDissimilarityCentroid Update
D c ¯ 1 D d c update
2DEC-kNNDefinition 4D c ¯ R 2 d e 1 | C | p C p
3DSC-kNNDefinition 6 s r 1 ( D ) s r 1 ( c ¯ ) R 3 d e 1 | C | p C p
2DSC-kNNDefinition 12 s r 1 ( D ) s r 1 ( c ¯ ) S 2 ( r ) d s (or d e ) r p C p p C p
SQ-kNNDefinition 10 s r 1 ( D ) s r 1 ( c ¯ ) R 3 d q p C p
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Viladomat Jasso, A.; Modi, A.; Ferrara, R.; Deppe, C.; Nötzel, J.; Fung, F.; Schädler, M. Quantum and Quantum-Inspired Stereographic K Nearest-Neighbour Clustering. Entropy 2023, 25, 1361. https://doi.org/10.3390/e25091361

AMA Style

Viladomat Jasso A, Modi A, Ferrara R, Deppe C, Nötzel J, Fung F, Schädler M. Quantum and Quantum-Inspired Stereographic K Nearest-Neighbour Clustering. Entropy. 2023; 25(9):1361. https://doi.org/10.3390/e25091361

Chicago/Turabian Style

Viladomat Jasso, Alonso, Ark Modi, Roberto Ferrara, Christian Deppe, Janis Nötzel, Fred Fung, and Maximilian Schädler. 2023. "Quantum and Quantum-Inspired Stereographic K Nearest-Neighbour Clustering" Entropy 25, no. 9: 1361. https://doi.org/10.3390/e25091361

APA Style

Viladomat Jasso, A., Modi, A., Ferrara, R., Deppe, C., Nötzel, J., Fung, F., & Schädler, M. (2023). Quantum and Quantum-Inspired Stereographic K Nearest-Neighbour Clustering. Entropy, 25(9), 1361. https://doi.org/10.3390/e25091361

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop