1. Introduction
Quantum image processing (QIP), which is an emerging sub-discipline that seeks to extend traditional (i.e., classical or digital) image processing tasks to (or using resources from) the quantum computing realm [
1], has attracted a lot of attention from practitioners that come from divergent backgrounds, notably those from the quantum computing, computer science, and engineering fields [
2]. The first published material relating quantum computers to image processing can be traced to the work [
3] by Vlasov in 1996, wherein the use of “analogue” quantum computing hardware to describe image analysis was mooted. This was followed by efforts to design quantum algorithms based on unstructured picture (data set) in [
4] and [
5]. Optics-based descriptions for quantum imaging then became very prominent [
6], while the idea of a more general quantum signal processing framework was presented in [
7]. However, it is the exploratory work by Beach and Lomont [
8], in 2003, that is often credited as the bellwether of what has transmuted into today’s description of QIP, albeit with a slightly different acronym—quip. This was closely followed by Venegas–Andraca and Bose’s Qubit Lattice description for quantum images [
9]. Years later, Lattore [
10] proposed the Real Ket representation to encode quantum images. The third and final representation, completing the trio of representations that are collectively regarded as the pioneers of today’s version of the QIP sub-discipline, is Le et al.’s flexible representation for quantum images (FRQI) that was proposed in 2010 [
11] and later revised in 2011 [
12].
Since then, the impressive development of the QIP sub-area is widely attributed to the FRQI representation because most of the QIP literature are either focused on its use in applications such as quantum watermarking [
13], quantum movie [
14], image data base search [
15], etc.; or the extension and modification of the FRQI representation to encode different types of images [
16]. Overall, the objective of QIP is to utilise quantum computing technologies to capture, manipulate and recover quantum images in different formats and for different purposes. Detailed reviews on QIP can be found in [
17,
18,
19].
Most of the papers cited above are, however, focused on covering the latter two objectives of QIP, i.e., capturing and manipulating the images, while the former (i.e., encoding an image) is scantily covered [
20]. Even so, to the best of our knowledge, only [
16] provided readers with insights into likely technologies to visualise transformed images resulting from operations on quantum images [
17]. Meanwhile, interest in image and video security applications such as watermarking [
13,
21], encryption and decryption [
22] has continued to increase.
To the best of our knowledge, all the aforementioned applications of QIP utilise the digital peak-noise-to-signal-ratio (PSNR) image quality measure to benchmark and validate their approaches. This is mainly attributed to the absence of a quantum-based metric equivalent to the PSNR. The closest attempt to quantify the extent that two or more images are alike is Yan et al.’s probability-based image similarity score technique in [
15,
22].
Most of the researchers are content with adopting the classical PSNR image quality measure to assess likeness between two or more quantum images. The quantum mechanical nature of the quantum information carrier—the qubit—however, specifies a, sometimes confounding, procedure for processing quantum information, considering which, we argue, in this study, that these available classical metrics are insufficient and/or ill-suited to effectively quantify the fidelity between two or more quantum images. Therefore, the absence of an effective metric for evaluating congruity between quantum images had remained an unexplored research area within the QIP sub-discipline until now.
In recent work [
1], Iliyasu et al. proposed a wholly quantum-based image fidelity metric (or simply QIFM) to assess “likeness” between quantum images. Very much like the more mature classical PSNR image quality measure, the proposed QIFM metric was formulated as a pixel difference image quality tool but with the necessary attributes required for its execution on the quantum computing framework built into it. These two simple instances differentiate the QIFM from both the digital PSNR measure and Yan’s probability-based image similarity score [
23]. The only resources required to execute the QIFM metric are quantum computing resources (i.e., logic gates) and some useful data retrieved from the classical versions of the reference and test image pair.
In this study, we concentrate on this new metric and propose accretions tailored towards ensuring moderate computational requirements are used for its execution. In addition, we present a more detailed validation of the effectiveness of this new quantum image quality metric by deploying it as the metric to estimate congruity between images in an expanded database (in terms of both the size of images and the scope of applications) to include images of varying size, complexity, and applications [
13].
Throughout our appraisal of the QIFM’s efficiency we will compare its results alongside those obtained using the classical PSNR (P) measure and Yan’s probability-based image similarity score (S); hence, allowing us to utilise an FPS (fidelity, PSNR, and Score) analysis as the basis for validation.
As originally conceived in [
1] and validated in this study, the proposed QIFM metric has the potential to replace the PSNR as an image quality measure on the quantum computing framework.
In
Section 2, we enunciate the veridical properties of the strip representation used to encode 2
m-ending FRQI quantum images and adopt its use to encode our quantum register containing the reference and test images whose congruity is sought. To complete the background review required for our study, in latter parts of
Section 2, we review Yan et al.’s probability-based image similarity score and offer some motivation as to why it is inadequate to estimate congruity between two (or more) quantum images.
Section 3 is devoted to discussions on the recrudesced QIFM image quality metric, while
Section 4 focusses on the enhanced validation of this new image fidelity metric.
2. Strip Representation and Its Use in the Probability-Based Comparison of Images
The strip representation
was first formulated in [
14] as an array comprising 2
m flexible representation for quantum images (FRQI, [
12]) images each of
size in the form defined in Equation (1):
where:
with:
where
is the position of each image in the strip,
m is the number of qubits required to encode the images being compared,
is a FRQI image as defined in Equation (1) at position
,
and
encode the information about the colours and their corresponding positions in the image
. It is trivial to see that the state
is a normalised one capable of encoding 2
m quantum images using only
m + 2
n + 1 qubits.
As mentioned earlier, the strip representation was first conceived in [
14] where it was utilised as the backbone for the framework to represent and produce quantum movies. Later, it was used to compare similarity between images [
23], in database search [
15], and in quantum video encryption and decryption [
22].
Inspired by the utility of image search and retrieval on digital computers, the studies in [
15,
19,
23] explored the possibility of undertaking similar tasks on quantum computers.
Therein, similarity between two FRQI quantum images
and
(as a function of their pixel information
) was obtained by traversing every position in the image. In this context, similarity (
Sim), is defined in Equation (5):
where
.
Based on the definition in Equation (5), two special cases of the similarity between two quantum images can be deduced as follows:
if , then and the two images are considered as being totally different (or dissimilar), where is the joint pixel information retrieved from the reference and test image pair;
if , then and the two images are considered as exact matches, where , is the joint pixel information retrieved from the reference and test image pair.
The circuit structure to compare these two FRQI quantum images (shown in
Figure 1) requires a Hadamard gate,
H that maps the basis states
to
and
to
that are applied on the strip wire, which is a combination of a reference image
and a test image
. The final step in the circuit is the measurement operation
(always shown with two parallel lines exiting a small rectangular box), which is used to recover the final readout resulting from the operations.
Based on the foregoing, it is obvious that the result of the measurement operation
M0 depends on the disparities between
and
. The states
and
subsist on the strip wire
S0 at a certain probability. In accordance with the measurement postulate used in [
23], the probability of state
on this strip wire is:
In the same manner, the probability of state
on the same wire is:
The probabilities of these two states sum up to 1, i.e.,
, as they should [
22]. It is apparent that, arising from Equations (6) and (7), the pixel difference
is related to the probability of obtaining a readout of 1 from the strip wire
,
, in the measurement and that this probability
will increase when the pixel difference increases.
Furthermore, the similarity between two images, which is a function of the pixel differences at every position, depends on
as given in Equation (8) [
23]:
where
and
are the two images being compared,
is in the form defined in Equation (7), and
.
The similarity between the states and that are encoded in the strip agrees with the definition of similarity between two FRQI quantum images in Equation (5), therefore:
Figure 2 shows the generalised circuit required to compare similarity between multiple FRQI quantum images in a strip.
The circuit shows that
GS operations on the
m strip wires;
m (0 or 1 controlled operations that are depicted using the half-moon notation that is further explained in
Section 3); G
I operations that lead to geometric exchanges between 2
m images encoded in the strip; and zero (0) or one (1) 2
n–control condition Hadamard operations followed by
m measurement operations are needed in order to compare the similarity between multiple quantum images. It should be noted that the measurements
M0 to
Mm–1 that are depicted in
Figure 2 are standard measurement operations that lead to the collapse of the quantum state, and as such the images encoded on the strip are lost after each measurement cycle is executed [
14].
As highlighted in [
14], in the QIP context, complexity is analysed in terms of three parameters. These parameters are: firstly, cost, which is expressed in terms of the basic quantum gates used to accomplish any transformation on an image; secondly, the width of circuit, which depends on the size of the image and is usually fixed for any
N ×
N image; and thirdly, the depth of the circuit, which represents the number of layers into which the entire circuit can be efficiently partitioned. In other words, complexity is influenced by the dimension and types of control operations required to execute an operation.
As seen in
Figure 2, Yan et al.’s probability-based score method to compare similarity between quantum images could become prohibitively expensive because the depth of the circuit increases with the size of the strip (i.e., the number of images in the multi-image strip). This is more so when the comparison is being confined to a smaller proportion of images relative to the size of the strip or when the comparison is focused on predetermined regions of interest within two or more of the images in the strip.
In addition to the complexity of execution, all the measurement operations in [
23] are performed on the strip wire, which, as observed earlier, disrupts the “quantumness” of the system and the resulting read outs lead to a complete collapse of the system with which comes the loss of the hitherto quantum information encoded in the system. In many instances, such as quantum movie production or database search [
15], these losses could be too expensive to bear. The interim solution suggested in Yan et al.’s probability-based approach to estimating similarity between quantum images [
23], which requires multiple copies of the images be prepared, could also turn out to be even more prohibitively expensive. Therefore, in general, Yan’s approach [
23] is deemed imprudent.
Most QIP literature assumes operability (or execution) based on the circuit-model of quantum computation. In such models, classical versions of the images are used to prepare a compact unit in a quantum register and, following this, in order to execute any type of computation, the qubits in this register are actively manipulated by a network of logical gates [
14].
It has been noted, however, that the required control of these registers is very challenging to realise [
14]. In addition to this, the inherent quantum properties, principally, superposition and entanglement make the circuit-model of computation unsuitable for recovering or measuring content of certain transformed images, such as the movie reader in quantum movies [
14,
17]. This is because the “quantumness” of the quantum state is lost immediately after it is observed. However, oftentimes, it is also required that the content of the register be retained (or left) undisturbed in spite of the measurement.
Measurement-based quantum computation (MBQC) [
14,
24] was proposed as an alternative strategy to the circuit-model and it relies on the effects of measurement on an entangled multi-partite resource state to perform the required computation [
14,
17]. This strategy provides an astute pathway to overcome the perceived shortcomings of the circuit-model of quantum computation and aspects of it have been realised experimentally using single-qubit measurements that are considered “cheap” [
14,
17,
24].
All measurement-based models of quantum computing share the common feature that measurements are not performed on the qubits storing the data, because this will destroy the coherence essential for quantum computation. Rather, ancillary qubits are prepared and interacted with the information carrying qubits, so that the measurement operation is performed on them. By carefully choosing appropriate measurement and initial states of the ancillary qubits, the much sought coherence can be preserved [
24].
Figure 3A shows the notation for a single qubit projective (destructive) measurement whence the quantum information is lost immediately its content is accessed. On its left (i.e., in
Figure 3B) is a non-destructive (or ancilla-driven) measurement as used in MBQC, and in this particular case the ancilla-driven quantum computation (ADQC) model.
The ADQC [
19] is a newer type of MBQC, which was utilised for the movie reader in [
9], exploits the properties of single-qubit projective measurements and the entanglement-based interaction between the ancilla qubit and the register qubit (i.e., strip state encoding the 2
m-ending FRQI images). The measurements are performed subject to satisfying some predefined conditions [
24].
In the ADQC, the ancilla
A is prepared and then entangled to a register qubit (in our case the single qubit encoding the color information [
14]) using a fixed entanglement operator
E. A universal interaction between the ancilla and register is accomplished using the Controlled-Z (CZ) gate and a swap (
W) gate and then measured. An ADQC with such an interaction allows the implementation of any computation or universal state preparation [
14,
24]. This is then followed by single qubit corrections on both the ancilla and register qubits. Similar to all the measurement-based models, the standard ADQC uses a fully controlled ancilla qubit, which is coupled sequentially to one or at most two qubits of a register via a fixed entanglement operator
E. After each coupling, the ancilla is measured in a suitable basis, providing a back action onto the register [
24]. This implements both single- and two-qubit operations on the register qubits. The ADQC is also a hybrid scheme of the circuit model, since computation involves a sequence of single and two-qubit gates implemented on a register.
The description of
Figure 3B as provided via [
24] presents the ancilla-driven implementation of a single qubit rotation,
JR(β), on a register qubit
R, where the initial state of the total register can be a pure state,
, or accordingly for a mixed state. The ancilla and register qubits are first coupled with
EAR =
HA·
HR·
CZAR. The rotation
J(β) is then implemented on the fully controlled ancilla and transferred to the register qubit by measuring the ancilla in the z-basis. The result of the measurement,
j = 0, 1, determines if an
X correction appears on the register qubit.
4. Experimental Validation of the QIFM Metric
In the absence of physical quantum computing hardware, since its early use in [
11,
12,
13,
14], the use of Matlab-based classical simulations of the quantum images and circuitry to manipulate its content has become commonplace in QIP. The experiments reported in this section are based on a similar approach whereby a classical computer with Intel (R) Xeon (R) CPU E3–1225 3.20 GHz, 4.00 GB RAM equipped with Matlab R2016a program was used to perform the simulation-based experiments. Readers are referred to [
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23] for detailed descriptions of how FRQI-based QIP simulations are carried out.
4.1. FPS Analysis of Congruity between Quantum Images
In the introductory sections of this study we promised readers an analysis that compares congruity between two or more FRQI quantum images in terms of three metrics—our proposed QIFM fidelity metric (
F), the well-established classical PSNR (
P) image quality measure, and Yan’s probability-based image similarity score (
S) [
15]. The remainder of this section is devoted to this
F,
P and
S or simply FPS analysis.
In presenting the FPS analysis, we adopt simulation–based techniques similar to those in [
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23] in order to simulate a dataset comprising of 16 images, each of size 512 × 512 (presented in
Figure 8 below), that are widely used in image processing literature. Similar simulation-based approaches were used to simulate the circuitry to execute the
BCO and
BO sub-circuits that coalesce to execute the
FCO operation. Using this layout, the efficacy of our proposed QIFM metric as an acceptable measure to assess fidelity between two or more FRQI quantum images can be validated as presented and discussed in the sequel.
Using the labels assigned to the images in our database (
Figure 8), a notation
r-
T is used to indicate a
jth pair consisting of a reference image,
r, and its test image,
T. It is assumed that a strip (defined in
Section 2) consisting of the 16 images in our database has been captured, prepared and initialised as required in FRQI QIP [
17,
19]. For brevity, however, the images in our database are presented neither as a vertically nor as a horizontally-oriented strip [
9]. Instead, it is presented as a stack of images labelled from A to P (in
Figure 8).
Next, the 16 images that make up the strip are randomly paired into twenty-three (23) pairs each consisting of a reference and test image. The “likeness” between these pairs was evaluated in terms of the fidelity,
Fr,T; the PSNR,
Pr,T; and the Similarity,
Sr,T, between the pairing of each reference image (
r) and a test image (
T). These results are summarised in
Table 1 while the detailed FPS analysis is presented in the sequel.
The peak-signal-to-noise-ratio—PSNR (
P) is widely adopted as an image reconstruction and quality measure in both digital image processing and more recently in QIP. It is most easily defined via the mean squared error (
MSE), which for two
m ×
n monochrome images the original or input image,
I and its watermarked version
K in Equation (15):
Here,
MAXi is the maximum possible pixel value of the image [
13,
17,
25,
28].
4.2. QIFM (F) versus Yan’s Similarity Score (S)
As seen in
Table 1, Yan et al.’s probability-based image similarity score, which was defined in [
15,
24] and highlighted in
Section 2, produces results that are at best exiguous because the similarity scores for the 23 pairs (reported in
Table 1) vary within a narrow band in the range from 0.8 to 1.0.
This narrow range does not discernibly convey similarity and less so when two or more pairs are assessed. Actually, as defined in [
15,
23], this image similarity score produces a minimum value of 0.5 for any random pairing of a reference and test image. In other words, using the definition in [
15,
23] and Equation (8) it is guaranteed that even the most dissimilar pairing of a reference and test image will have a 50% similarity score, which appears very lopsided.
In contrast, our proposed QIFM fidelity metric has a smoother (i.e., wider-range) interval that is well spaced out in a manner that realistically conveys congruity between each of the reference and test images that were paired.
In order to widen the interval and expand the range of coverage for Yan et al.’s image similarity score (
S), here, we recalibrated the original formulation of the similarity score, which was defined in Equation (8) (and the discussions in [
23]), so that a wider range that is normalised between zero to one (i.e., an interval [0,1]) is obtained. We achieve this recalibration using Equation (16):
where
, which is the probability of obtaining a state
after measurement is made, retains its form as defined in Equation (7). The coefficient of ten (10) is added to ensure that a normalised value is obtained for the
SIM value between two or more images without significantly altering the requirements needed to realise the original results using Equation (8).
Consequently, the values in the sixth column of
Table 1 are the revised similarity values and, in order to distinguish it from the original notation for similarity (as defined in Equation (8)), we call this new measure the
SIM score between the pair
r and
T.
Using this recalibrated formulation of similarity, a wider range of values that more graphically convey the notion of congruity is obtained. This adjustment in the similarity scores (
S) better mimics the explicitness expected in descriptions of congruity between different pairings of images in
Figure 8.
One example that clearly stands out in showing the importance of this revised formulation for Yan’s probability-based image similarity score is the House–Bridge pairing in pair L–O. A simple visual inspection reveals that these two images that are glaringly discrepant. However, Yan’s similarity score (S) attributes a high similarity score of 0.918 for this pair but on the recalibrated similarity (SIM) scale this has a score of only 0.59.
Similar adjustments can be seen in the other reference and test images pairs, notably that in the Tree-House pair (i.e., pair P–L), which reduces from 0.843 using Yan’s image similarity score to only 0.21 using the revised SIM scale. All these adjustments suggest that the recalibrated SIM index is a better similarity scale than Yan’s original similarity score (S).
However, notwithstanding the aforementioned calibration to Yan’s original similarity score, the revised similarity score (i.e., the SIM values) attributes a similarity of 0.52 for the Lena–Inverted Lena reference and test image pair (i.e., pair A–B), which is obviously a poor correlation with explicit lexical descriptions of congruity (or likeness) between the two images). In contrast, to manifest itself as a better measure of congruity, the proposed QIFM metric attributes 62% fidelity between the pair, which is a better explicit lexical description of likeness between that pair.
Similar shortcomings of both the original formulation of Yan’s similarity score (S) and its recalibrated wider range, normalised version—the SIM score—are clearly seen in the Couple–Trees pair (i.e., pair J–P), which has high similarity scores of SJ,P = 0.9 and SIMJ,P = 0.7. These values seem illogical since a mere visual inspection reveals that the two images are clearly dissimilar.
In contrast, a fidelity value of
FJ,P = 52% that is obtained using the QIFM metric, which clearly infers less congruity (likeness) between the two images. Additionally, a 52% fidelity clearly resembles the expectations of the explicit lexical description of the congruity (or likeness) between these two images (i.e., the Couples and Trees images in
Figure 8).
As noted earlier in
Section 2, the most pertinent shortcoming of Yan’s similarity score (
S) is that it has high resource demands because its recovery (measurement) procedure is built entirely on the use of destructive measurements, which are carried out in each instance that the similarity between two images is assessed.
To clearly illustrate this high cost, let us assume that the measurements to quantify the similarity of the images in our strip are carried out in the order indicated by the serial numbers in
Table 1 (i.e., S/No. in the first column)—starting with the first pair (i.e., pair A–A in S/No. 1) and terminating with the Bridge–Aeroplane pair (i.e., pair O–K in S/No. 23)—it is instructive to note that the entire content of the strip will be lost upon the first measurement, i.e., after recovering the similarity between the first pair (i.e.,
SA,A in S/No. 1). In other words, to execute the experiments narrated in this section (and quantify the similarity scores for all the 23 pairings of the reference and test images in
Table 1) we need to prepare 23 versions of the same strip, i.e., one to recover the similarity of each reference and test image pair.
In cost terms, to execute Yan’s image similarity score a total of Cs = cn × 2mH resources will be required. Here, m is the number of qubits required to encode the images that make up the strip; c is the number of pairs whose similarity is being measured and H accounts for the cost of resources needed to prepare and initialise the 2m-images long the strip each of 2n × 2n size and executing the same number of measurements, i.e., 2m measurement operations to obtain the similarity between the reference and test images in the pair.
To illustrate this cost in terms of our experiments, quantifying Yan et al.’s probability-based similarity score (
S) for the 16-image long strip requires a total cost of
Cs = 23
8 × 2
3H. Here, we should further clarify that
H represents the cost associated with the peculiar needs of the technology (i.e., whether ion-trap, photonic, etc. hardware that is) used to prepare and initialise the 16-image long strip [
1].
In contrast to this, the proposed QIFM fidelity metric requires only 362n basic quantum gates and 26m Toffoli gates to implement. Additionally, the entire framework requires m + 1 + 2n qubits and m × 22(n−s)+4 ancillary measurement qubits in order to evaluate the fidelity of any predetermined 2s × 2s ROI in a strip consisting of 2m FRQI images, each of size N = 2n × 2n (as discussed earlier in this section).
It also merits emphasis to state that, in the quantum sense, the bizarre requirements needed to prepare and maintain a state in the format that supports meaningful computation is a daunting task. Moreover, it has been suggested that the cost attributed to producing multiple versions of a strip far outweighs the cost of the quantum computing resources required to manipulate and transform a quantum image [
14,
16,
17,
29]. Assigning a notation
e for this cost (of manipulating a quantum image), it then follows that
Cs >>
e.
Considering the QIFM metric, our experimental setup that consists of a 16-image long strip each of size 256 × 256 (i.e., m = 3 and n = 8) comes at a much cheaper cost (approximately, 2974 × H) where e = 362n + 26m. This cost is far less than the cost that was attributed to the execution of Yan’s similarity score (CS), which suggests that the QIFM fidelity metric is a cheaper and more efficient tool to assess congruity between FRQI quantum images.
4.3. PSNR versus Quantum Image Quality Metrics (QIFM (F), Similarity Score (S) and SIM Values)
The antinomy that pervades information processing in the digital and quantum realms imposes the need for a different set of rules must be used for quantum state preparation, transformation and recovery [
2,
29].
In QIP, once an image is prepared its transformation is only possible via a set of unitary operations as dictated by the algorithmic framework generating the required image application or task [
14,
16,
17]. The final step in QIP is quantum state recovery, which is perhaps the most delicate undertaking since, as noted in earlier sections of this paper, the content encoded in the image register could be lost if it is disturbed during any uncontrolled measurement. In addition to the possible collapse of the quantum state, such uncontrolled measurements have inescapable error terms integrated into them.
Without necessary refinement, using the digital PSNR image quality metric to assess congruity between two quantum images is both impracticable and non-executable because of the mismatch in terms of its present formulation with the procedures of quantum information processing. This makes the PSNR image quality measure an ill-suited and irrational choice for use in QIP.
Notwithstanding this, since, in QIP, classical (digital) copies of images used as references to prepare and initialize their quantum versions [
11,
12,
16,
17,
18,
19], it would suffice that the PSNR obtained from the measurement of likeness between classical cover and test image pair should serve as a reference or benchmark to assess fidelity between the quantum versions of the same image pair. Thence, a bivariate statistical analysis could be utilised in order to establish the strength of association or correlation between the quantum metrics (i.e., the QIFM (
F), Yan et al.’s probability-based score (
S) and recalibrated similarity (i.e., the
SIM values)) assessing the likeness between a cover and test images pair with their digital counterpart, i.e., the PSNR, obtained for the same pair of images.
In trying to establish this relationship, however, we must be mindful of the conflict in the measurement scales used in the PSNR and the trio of quantum image quality metrics that are used in our FPS analysis. While the quantum image quality metrics are measured on a normalised linear scale bounded in the range 0 to 1, the PSNR is measured on a logarithmic scale that spans an infinite length. Therefore, it is imperative that we must find some intermediate yet impartial platform for our correlation analysis.
As defined in Equation (15), the PSNR is composed of constant terms having an asymptotic relationship with the mean square error (
MSE) value with which it varies inversely and, is defined as the squared deviations of a prediction from its true values. This implies that the PSNR and
MSE are equivalent measures with asymptotic behaviour in terms of appropriated normalised or linear measurement scales. Moreover, since the
MSE can be normalised and reversed (
rNMSE) as presented in Equation (16) [
30], as an equivalence of the PSNR would provide an impartial platform for comparison with the other quantum image quality metrics.
where
rNMSE is the reversed
NMSE that fits in the interval of [0, 1] and in terms of quantitative inference a measurement of
rNMSE = 1 means that the two images being compared are exact matches. Using Equation (17),
rNMSE measurements for the cover and test image pairs in the first column of
Table 1 are presented in the rightmost (i.e., the last) column of the same table.
Having provided a more practical equivalence for the PSNR values, we have an impartial platform for our analysis between the digital image quality measures and their quantum counterparts. Now, we present a statistical analysis for the degree of association between the
rNMSE (
r) on one side and the trio of the QIFM (
F), Yan et al.’s probability-based score (
S) and its recalibrated version (
SIM) values on the other. For this task, we shall employ the widely used statistical correlation tests—i.e., the Spearman, Pearson and Kendal rank correlation coefficients—or simply the SPK correlation test. Details of these tests can be found in standard statistics texts and other available literature such as [
28].
Table 2 presents results of our SPK correlation test to ascertain the degree of association between the rNMSE digital image quality measure and the three quantum image quality metrics (i.e.,
F,
S and
SIM).
In interpreting the results in
Table 2, we see that the proposed QIFM metric (
F) has the best correlation with the
rNMSE than any of the other quantum image metrics (i.e., Yan et al.’s probability-based score (
S) and the
SIM values) for each of the correlation indices that we employed in the SPK test. In other words, using the classical versions of the quantum test and cover images in
Figure 8 (as paired in
Table 1) as benchmark, the assessment of congruity based on the QIFM fidelity metric (
F) offers the least correlation in terms of deviations (aka congruity) between the test and cover images in each pair than is obtainable via both Yan’s probability score (
S) and its recalibrated version (
SIM).
Table 2 allows us to make the deduction that in spite of the quantum mechanical requirements for its execution, the QIFM metric yields the best correlation with expectations from classical (digital) image quality measures.
In concluding our assessment using the PSNR, we revisit the lack of explicit lexicality in descriptions of congruity (or likeness) between images based on the PSNR as measure of image quality. To demonstrate this, we considered descriptions of congruity between the watermarked versions of the Lena and Blonde Lady images in
Figure 9D,E and their pristine versions in
Figure 9B,C (which were originally presented in
Figure 8A,C). Using the WaQI quantum image watermarking protocol [
13,
21,
27], we paired each of the two images with the PSAU logo in
Figure 9A as watermark signal in order to obtain the watermarked versions of Lena and Blonde Lady images (i.e., in
Figure 9D,E). The PSNR values and fidelity scores conveying the notion of congruity (or likeness) between the original and watermarked versions of the two images are also presented in
Figure 9.
As seen from the results, at 97% fidelity score the QIFM fidelity metric offers better explicitness in describing congruity (or likeness) between the original and watermarked versions of the images. From this description, it is trivial to deduce that the watermarked and original versions of the two images are somewhat the same, or that they have high fidelity. On the contrary, the PSNR values of 31.59 and 32.33 dB (for the Lena and Blonde Lady images respectively) present ambiguity in terms of descriptions of congruity between the original and watermarked images. Therefore, the quantum image metrics offer better lexical explicitness in their descriptions of congruity between two or more images than the PSNR in its standard form
5. Concluding Remarks
While remarkable progress has been made in the QIP sub-discipline over the last few years, it has still continued to rely on classical techniques, notations, and formulations, notably the PSNR image quality measure, to evaluate fidelity between quantum images. Incidentally, recent studies have proven that the PSNR metric as an unreliable image quality metric even in some digital applications.
Contending that such a metric does not conform with the intricacies inherent to quantum computing and QIP technologies, in this study, we have appraised the need for, and formulated a wholly quantum-based metric, the quantum image fidelity metric (QIFM), which could be used to efficiently assess the congruity between two or more quantum images.
The proposed QIFM metric is designed as a pixel difference-based image quality measure that can be effectively implemented using available QIP representations for quantum images; while, only basic quantum computing resources and some useful data collated from classical versions of the reference and test image pairs are required to manipulate and assess the fidelity between the images.
We simulated the implementation of our image quality metric (using classical computing resources) on a dataset of popularly used real images and our results suggest that in comparison with the traditional PSNR image quality measure the QIFM metric is a better tool to convey the notion of congruity.
Since it is conceived for use on the quantum computing framework, the QIFM metric is equipped to withstand the intricacies of such hardware, whilst the digital PSNR metric is ill–suited for such hardware implementation. Similarly, in comparison with the only known quantum image similarity technique (i.e., Yan et al.’s probability-based score) [
15,
23], the ancilla-driven attribute of our proposed QIFM makes it a more efficient and prudent image quality measurement tool. Additionally, by using a statistical analysis we established that the proposed QIFM metric had a better correlation with digital image quality assessments of congruity than the other quantum image quality measures.
Besides being a wholly quantum–based measure, the proposed QIFM image quality metric requires only 362n basic quantum gates and 26m Toffoli gates to implement. Additionally, the entire framework requires m + 1 + 2n qubits and m × 22(n−s)+4 ancillary measurement qubits in order to evaluate the fidelity of any predetermined 2s × 2s ROIs in a strip consisting of 2m FRQI images, each of size N = 2n × 2n. The formulation of the QIFM metric, in itself, is the first step in ensuring that applications sensitive to the peculiarities of quantum computation are formulated for effective quantum image processing (QIP).
In on-going work, we are assessing how to further reduce the overall cost of the proposed metric both in terms of its encoding (based on photonic quantum computing hardware implementation) and in terms of its overall execution.
In later work, we will utilise the resulting improved version of the QIFM metric to better assess the fidelity between different pairs of quantum images, such as those realised when a quantum image is watermarked, make-up frames required for quantum movie aggregation, etc. [
15,
22,
23].
The present formulation of the QIFM metric and its improved version, when realised, have the potential to improve the accuracy quantum image quality assessment and QIP applications in general.