Next Article in Journal
The Reducibility Concept in General Hyperrings
Next Article in Special Issue
Quantum Computing for Dealing with Inaccurate Knowledge Related to the Certainty Factors Model
Previous Article in Journal
Respondent Burden Effects on Item Non-Response and Careless Response Rates: An Analysis of Two Types of Surveys
Previous Article in Special Issue
DMO-QPSO: A Multi-Objective Quantum-Behaved Particle Swarm Optimization Algorithm Based on Decomposition with Diversity Control
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Quantum-Like Sampling

Department of Informatics, INESC-ID/IST-University of Lisboa, 1000-029 Lisboa, Portugal
Mathematics 2021, 9(17), 2036; https://doi.org/10.3390/math9172036
Submission received: 7 July 2021 / Revised: 11 August 2021 / Accepted: 14 August 2021 / Published: 24 August 2021
(This article belongs to the Special Issue Advances in Quantum Artificial Intelligence and Machine Learning)

Abstract

:
Probability theory is built around Kolmogorov’s axioms. To each event, a numerical degree of belief between 0 and 1 is assigned, which provides a way of summarizing the uncertainty. Kolmogorov’s probabilities of events are added, the sum of all possible events is one. The numerical degrees of belief can be estimated from a sample by its true fraction. The frequency of an event in a sample is counted and normalized resulting in a linear relation. We introduce quantum-like sampling. The resulting Kolmogorov’s probabilities are in a sigmoid relation. The sigmoid relation offers a better importability since it induces the bell-shaped distribution, it leads also to less uncertainty when computing the Shannon’s entropy. Additionally, we conducted 100 empirical experiments by quantum-like sampling 100 times a random training sets and validation sets out of the Titanic data set using the Naïve Bayes classifier. In the mean the accuracy increased from 78.84 % to 79.46 % .

1. Introduction

Quantum algorithms are based on different principles than classical algorithm. Here we investigate a simple quantum-like algorithm that is motivated by quantum physics. The incapability between Kolmogorov’s probabilities and quantum probabilities results from the different norms that are used. In quantum probabilities the length of the vector in l 2 norm representing the amplitudes of all events is one. Usually, Kolmogorov’s probabilities are converted in Quantum-like probabilities by the squared root operation, it is difficult to attach any meaning to the squared root operation. Motivated by the lack of interpretability we define quantum-like sampling, the l 2 sampling. The resulting Kolmogorov’s probabilities are not any more linear but related to the sigmoid function. In the following we will introduce the traditional l 1 sampling Then we will introduce the quantum-like sampling, the l 2 sampling. We will indicate the relation between the l 2 sampling, the sigmoid function and the normal distribution. The quantum-like sampling leads to less uncertainty. We fortify this hypothesis by empirical experiments with a simple Naïve Bayes classifier on the Titanic dataset.

2. Kolmogorovs Probabilities

Probability theory is built around Kolmogorov’s axioms (first published in 1933 [1]). All probabilities are between 0 and 1. For any proposition x,
0 p ( x ) 1
and
p ( t r u e ) = 1 , p ( f a l s e ) = 0 .
To each sentence, a numerical degree of belief between 0 and 1 is assigned, which provides a way of summarizing the uncertainty. The last axiom expresses the probability of disjunction and is given by
p ( x y ) = p ( x ) + p ( y ) p ( x y )
where do these numerical degrees of belief come from?
  • Humans can believe in a subjective viewpoint, which can be determined by some empirical psychological experiments. This approach is a very subjective way to determine the numerical degree of belief.
  • For a finite sample we can estimate the true fraction. We count the frequency of an event in a sample. We do not know the true value because we cannot access the whole population of events. This approach is called frequentist.
  • It appears that the true values can be determined from the true nature of the universe, for example, for a fair coin, the probability of heads is 0.5 . This approach is related to the Platonic world of ideas. However, we can never verify whether a fair coin exists.

Frequentist Approach and l 1 Sampling

Relying on the frequentist approach, one can determine the probability of an event x by counting. If Ω is the set of all possible events, p ( Ω ) = 1 and the cardinality card ( Ω ) determines the number of elements of a set Ω and card ( x ) is the number of elements of the set x and with x Ω
p ( x ) = card ( x ) card ( Ω ) .
This kind of sampling is the l 1 sampling. With n events x 1 , x 2 , . . . , x n that describe all possible events of the set Ω ,
0 card ( x i ) card ( Ω ) , i = 1 n card ( x i ) card ( Ω ) = 1 i = 1 n card ( x i ) = card ( Ω )
We can interpret card ( ω i ) card ( Ω ) as dimension of an n dimensional vector with the vector being
card ( x 1 ) card ( Ω ) card ( x 2 ) card ( Ω ) card ( x n ) card ( Ω ) .
with the l 1 norm
x 1 = | x 1 | + | x 2 | + + | x n |
it is as a unit-length vector in the norm l 1
card ( x 1 ) card ( Ω ) + card ( x 2 ) card ( Ω ) + + card ( x n ) card ( Ω ) = 1 .

3. Quantum Probabilities

Quantum physics evaluates a probability p ( x ) of a state x as the squared magnitude of a probability amplitude A ( x ) , which is represented by a complex number
p ( x ) = | A ( x ) | 2 = A ( x ) * · A ( x ) .
This is because the product of complex number with its conjugate is always a real number. With
A ( x ) = α + β · i
A ( x ) * · A ( x ) = ( α β · i ) · ( α + β · i )
A ( x ) * · A ( x ) = α 2 + β 2 = | A ( x ) | 2 .
Quantum physics by itself does not offer any justification or explanation beside the statement that it just works fine, see [2]. This quantum probabilities are as well called von Neumann probabilities. Converting two amplitudes into probabilities leads to an interference term 2 · ( A ( x ) · A * ( y ) ) ,
| A ( x ) + A ( y ) | 2 = | A ( x ) | 2 + | A ( y ) | 2 + 2 · ( A ( x ) · A * ( y ) ) | A ( x ) + A ( y ) | 2 = p ( x ) + p ( y ) + 2 · ( A ( x ) · A * ( y ) ) ,
making both approaches, in general, incompatible
| A ( x ) + A ( y ) | 2 p ( x ) + p ( y ) .
In other words, the summation rule of classical probability theory is violated, resulting in one of the most fundamental laws of quantum mechanics, see [2]. In quantum physics we interpret A ( x i ) as dimension of an n dimensional vector with the vector being
A ( x 1 ) A ( x 2 ) A ( x n ) .
In quantum physics the unit-length vector is computed in the in the norm l 2
x 2 = | x 1 | 2 + | x 2 | 2 + + | x n | 2
instead of the norm l 1 with a unit-length vector in the norm l 2
| A ( x 1 ) | 2 + | A ( x 2 ) | 2 + + | A ( x n ) | 2 = 1 .
By replacing the l 1 norm by the Euclidean l 2 norm in the classical probability theory (Kolmogorov probabilities), we obtain quantum mechanics [3] with all the corresponding laws. The incapability between the Kolmogorov’s probabilities and quantum probabilities results from the simple fact
x 2 x 1 .
This incapability is as well the basis for quantum cognition. Quantum cognition is motivated by clues from psychology indicate that human cognition is based on quantum probability rather than the traditional probability theory as explained by Kolmogorov’s axioms, see [4,5,6,7]. Empirical findings show that, under uncertainty, humans tend to violate the expected utility theory and consequently the laws of classical probability theory (e.g., the law of total probability [6,7,8]), In [4,9,10,11,12,13,14] leading to what is known as the “disjunction effect” which, in turn, leads to violation of the Sure Thing Principle. The violation results from an additional interference that influences the classical probabilities.

Conversion

The amplitude is the root of the belief multiplied with the corresponding phase [11,12,13] θ [ 0 , 2 · π )
A ( x ) = a ( x , θ ) = p ( x ) · e i · θ .
With l 1 sampling it is difficult to attach any meaning to card ( x ) ,
a ( x , θ ) = card ( x ) card ( Ω ) · e i · θ .
Motivated by the lack of interpretability we define the quantum-like sampling, the l 2 sampling with n events x 1 , x 2 , . . . , x n that describe all possible events
Φ = | card ( x 1 ) | 2 + | card ( x 2 ) | 2 + + | card ( x n ) | 2
p ( x i ) = | card ( x i ) | 2 Φ , i = 1 n | card ( ω i ) | 2 Φ = 1
and
a ( x , θ ) = card ( x ) card ( Φ ) · e i · θ .
with
Ω Φ Ω 2 .
The l 2 sampling leads to an interpretation of the amplitudes as a normalized frequency of occurrence of an event multiplied by the phase. Φ is dependent on the distribution of all values card ( x i ) . When developing empirical experiments that are explained by quantum cognition models l 2 sampling should be used rather than l 1 sampling. What is the interpretation of | card ( x i ) | 2 outside quantum interpretation?

4. Quantum-Like Sampling and the Sigmoid Function

To understand the difference between l 1 sampling and l 2 sampling (quantum-like sampling) we will analyze a simple binary event x
p ( x ) + p ( ¬ x ) = 1 .
For l 1 sampling p ( x ) and p ( ¬ x ) is defined as
p ( x ) = card ( x ) card ( Ω ) , p ( ¬ x ) = Ω card ( x ) card ( Ω )
and for l 2 sampling
p ( x ) = card ( x ) 2 card ( x ) 2 + ( Ω card ( x ) ) 2 , p ( ¬ x ) = ( Ω card ( x ) ) 2 card ( x ) 2 + ( Ω card ( x ) ) 2 .
with
ω = card ( x )
we can define the functions for binary l 1 and l 2 sampling as
f ( ω ) = ω Ω
and
g ( ω ) = ω 2 ω 2 + ( Ω ω ) 2 ,
In Figure 1  f ( ω ) = ω Ω is compared to g ( ω ) = ω 2 ω 2 + ( Ω ω ) 2 for Ω , (a) Ω = 4 , (b) Ω = 10 , (c) Ω = 100 and (d) Ω = 1000 . With growing size of Ω the function g ( ω ) converge to a continuous sigmoid function. In Figure 2 the sigmoid function g ( x ) with Ω = 10 is compared to the well known scaled to the logistic function.
The derivative of the sigmoid function g ( x ) is a bell-shaped function. For the continuous function g ( x ) , the derivative is similar to the Gaussian distribution. The central limit theorem states that under certain (fairly common) conditions, the sum of many random variables will have an approximately Gausssian distribution. In Figure 3 the derivative of g ( x ) = x 2 x 2 + ( 10 x ) 2 with Ω = 10 , g ( x ) = 5 ( 10 + x ) x ( 50 10 x + x 2 ) 2 is indicated and compared to Gaussian distribution N with μ = 5 and σ 2 = 1 . 9 2 .
The derivative of the sigmoid function g ( ω ) is less similar to the probability mass function of the the binomial distribution.
The l 2 sampling leads to natural sigmoid representation of probabilities that reflects the nature of Gaussian/Normal distribution. This leads to less uncertainty using l 2 compared to l 1 sampling represented by Shannon’s entropy.
H = p ( x ) · log 2 p ( x ) p ( ¬ x ) · log 2 p ( ¬ x )
as can be seen in Figure 4 for Ω = 100 .

Combination

Depending on l 1 or l 2 sampling we get different results when we combine probabilities. Multiplying two independent sampled events x and y results in the joint distribution
p l 1 ( x , y ) = x 10 · y 10
when l 1 sampled with Ω = 10 and
p l 2 ( x , y ) = x 2 x 2 + ( 10 x ) 2 · y 2 y 2 + ( 10 y ) 2
when l 2 sampled with Ω = 10 as indicated in the Figure 5. The same applies for weighted sum of information, corresponding to the Shannon’s entropy as used in the I D 3 algorithm [15,16,17] for symbolical machine learning to preform a greedy search for small decision trees. Computing
H l 1 = x 10 · log 2 x 10 10 x 10 · log 2 10 x 10 y 10 · log 2 y 10 10 y 10 · log 2 10 y 10
when l 1 sampled with Ω = 10 and
H l 2 = x 2 x 2 + ( 10 x ) 2 · log 2 x 2 x 2 + ( 10 x ) 2
1 x 2 x 2 + ( 10 x ) 2 · log 2 1 x 2 x 2 + ( 10 x ) 2
x 2 x 2 + ( 10 x ) 2 · log 2 x 2 x 2 + ( 10 x ) 2
1 x 2 x 2 + ( 10 x ) 2 · log 2 1 x 2 x 2 + ( 10 x ) 2
when l 2 sampled with Ω = 10 leads to different H 1 and H 2 values as indicated in the Figure 6.
Less uncertainty would lead to better results in machine learning. We preform empirical experiments with a simple Naïve Bayes classifier to fortify this hypothesis.

5. Naïve Bayes Classifier

For a target function f : X h , where each instance x described by attributes a 1 , a 2 a n Most probable value of f ( x ) is:
h m a p = arg max h i p ( h i | a 1 , a 2 , a 3 , . . , a n ) = p ( a 1 , a 2 , a 3 , . . . , a n | h i ) · P ( h i ) P ( a 1 , a 2 , a 3 , . . . , a n )
The likelihood of the conditional probability p ( a 1 , a 2 , a 3 , . . . , a n | h i ) is described by 2 n possible combinations. For n possible variables, the exponential growth of combinations being true or false becomes an intractable problem for large n since all 2 n 1 possible combinations must be known. The decomposition of large probabilistic domains into weakly connected subsets via conditional independence,
p ( a 1 , a 2 , a 3 , . . . , a n | h i ) = j = 1 n p ( a j | h i ) .
is known as the Naïve Bayes assumption and is one of the most important developments in the recent history of Artificial Intelligence [16]. It assumes that a single cause directly influences a number of events, all of which are conditionally independent. The Naïve Bayes classifier is defined as
h m a p = arg max h i j = 1 n P ( a j | h i ) · P ( h i ) .

Titanic Dataset

When the Titanic sank it killed 1502 out of 2224 passengers and crew. We are using a file titanic.cvs processed file from https://gist.github.com/michhar/ accessed on 1 July 2021 that corresponds to the file train.cvs from kaggle https://www.kaggle.com/c/titanic/data?select=train.csv accessed on 1 July 2021 that contains data for 891 of the real Titanic passengers. Each row represents one person. The columns describe different attributes about the person including their P a s s e n g e r I d , whether they S u r v i v e d , heir passenger-class P c l a s s , their N a m e , their S e x , their A g e , and other six attributes, see Figure 7. In our experiment we will only use the attributes S u r v i v e d , P c l a s s , S e x and A g e .
The binary attribute S u r v i v e d (=s) will be our target h resulting in the prior p ( s ) and p ( ¬ s ) . Out of the 891 passengers 342 survived and 549 did not survive. The attribute P c l a s s has three values 1 for upper class, 2 for middle class and 3 for lower class. It will be binarized into the attribute c indicating if the person had a lower class cabin by
c = 1 i f P c l a s s = = 3 e l s e c = 0 ,
in the case P c l a s s is not defined the default value is 0. The attribute S e x has two binary values represented by the binary attribute m
m = 1 i f S e x = = m a l e e l s e m = 0 ,
in the case S e x is not defined the default value is 0. The attribute A g e will be binarized into the attribute g indicating if the person is a child or grown up by
g = 0 i f A g e < 18 e l s e g = 1 ,
in the case A g e is not defined the default value is 1.
Using the l 1 sampling over the whole data set we get the values for the priors
p ( s ) = 0.3838 , p ( ¬ s ) = 0.6162 ,
and for the likelihoods,
p ( c | s ) = 0.3480 , p ( ¬ c | s ) = 0.6520 , p ( c | ¬ s ) = 0.6776 , p ( ¬ c | ¬ s ) = 0.3224 ,
p ( m | s ) = 0.3187 , p ( ¬ m | s ) = 0.6813 , p ( m | ¬ s ) = 0.8525 , p ( ¬ m | ¬ s ) = 0.14755 ,
p ( g | s ) = 0.8216 , p ( ¬ g | s ) = 0.178363 , p ( g | ¬ s ) = 0.9053 , p ( ¬ g | ¬ s ) = 0.0947 .
Using the l 2 sampling over the whole data set we get the values for the priors
p ( s ) = 0.27957406 , p ( ¬ s ) = 0.72042594
and for the likelihoods,
p ( c | s ) = 0.2217 , p ( ¬ c | s ) = 0.7783 , p ( c | ¬ s ) = 0.8154 , p ( ¬ c | ¬ s ) = 0.1846 ,
p ( m | s ) = 0.1796 , p ( ¬ m | s ) = 0.8524 , p ( m | ¬ s ) = 0.9709 , p ( ¬ m | ¬ s ) = 0.0291 ,
p ( g | s ) = 0.95499 , p ( ¬ g | s ) = 0.04501 , p ( g | ¬ s ) = 0.9892 , p ( ¬ g | ¬ s ) = 0.0108 .
The sampled probability values are quite different.
We measure the accuracy of our algorithm by dividing the number of entries it correctly classified by the total number of entries. For l 1 sampling there 190 are entries wrong classified resulting in the accuracy of 78.66 % , for the l 2 sampling 179 entries are wrong classified resulting in the accuracy of 79.91 % .
In the next step we separate our data set in a training set and a validation set resulting in 295 elements in the validation set We use sklearn.model_selection accessed on 1 July 2021 import train_test_split accessed on 1 July 2021 with the parameters train_test_split(XH,yy,test_size=0.33,random_state=42) accessed on 1 July 2021. The later is used to validate how well our algorithm is doing. For l 1 sampling 60 entries are wrong classified resulting in the accuracy of 79.66 % , for the l 2 sampling 55 entries are wrong classified resulting in the accuracy of 81.36 % .
In the next we sample the 100 times a random training set and validation set We use sklearn.model_selection accessed on 1 July 2021 import train_test_split accessed on 1 July 2021 with the parameters train_test_split(XH,yy,test_size=0.33) accessed on 1 July 2021. For l 1 sampling there were in the mean 62.42 entries wrong classified resulting in the accuracy of 78.84 % , for the l 2 sampling there were 60.59 entries wrong classified resulting in the accuracy of 79.46 % .
The trend from this simple evaluation indicates that l 2 sampling leads to better results compared to l 1 sampling in the empirical experiments.
So far we looked at the performance of classification models in terms of accuracy. Specifically, we measured error based on the fraction of mistakes. However, in some tasks, there are some types of mistakes that are worse than others. However in our simple task the correctly classified results when l 1 sampled are also correctly classified when l 2 sampled.

6. Conclusions

We introduced quantum-like sampling also called the l 2 sampling. The l 2 sampling leads to an interpretation of the amplitudes as a normalized frequency of occurrence of an event multiplied by the phase
a ( x , θ ) = card ( x ) card ( Φ ) · e i · θ ,
Φ is dependent on the distribution of all values card ( x i ) . When developing empirical experiments that are explained by quantum cognition models l 2 sampling should be used rather than l 1 sampling.
The quantum inspired l 2 sampling maps the probability values to a natural continuous sigmoid function, its derivative is is a bell-shaped function that is similar to the Gaussian distribution of events. The l 2 sampling improves the classification accuracy in machine learning models that are based on sampling as indicated by empirical experiments with Naïve Bayes classifier.

Funding

This work was supported by national funds through FCT, Fundação para a Ciência e a Tecnologia, under project UIDB/50021/2020.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets used in this paper are provided within the main body of the manuscript.

Acknowledgments

We would like to thank the anonymous reviewers for their valuable feedback.

Conflicts of Interest

Compliance with Ethical Standards: The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript. The author declare no conflict of interest. This article does not contain any studies with human participants or animals performed by any of the authors.

References

  1. Kolmogorov, A. Grundbegriffe der Wahrscheinlichkeitsrechnung; Springer: Berlin, Germany, 1933. [Google Scholar]
  2. Binney, J.; Skinner, D. The Physics of Quantum Mechanics; Oxford University Press: Oxford, UK, 2014. [Google Scholar]
  3. Aaronson, S. Is quantum mechanics an island in theoryspace? In Proceedings of the Växjö Conference Quantum Theory: Reconsideration of Foundations; Khrennikov, A., Ed.; University Press: Vaxjo, Sweden, 2004; Volume 10, pp. 15–28. [Google Scholar]
  4. Busemeyer, J.; Wang, Z.; Trueblood, J. Hierarchical bayesian estimation of quantum decision model parameters. In Proceedings of the 6th International Symposium on Quantum Interactions, Paris, France, 27–29 June 2012; pp. 80–89. [Google Scholar]
  5. Busemeyer, J.R.; Trueblood, J. Comparison of quantum and bayesian inference models. In Quantum Interaction; Lecture Notes in Computer, Science; Bruza, P., Sofge, D., Lawless, W., van Rijsbergen, K., Klusch, M., Eds.; Springer: Berlin/Heidelberg, Germany, 2009; Volume 5494, pp. 29–43. [Google Scholar]
  6. Busemeyer, J.R.; Wang, Z.; Lambert-Mogiliansky, A. Empirical comparison of markov and quantum models of decision making. J. Math. Psychol. 2009, 53, 423–433. [Google Scholar] [CrossRef]
  7. Busemeyer, J.R.; Wang, Z.; Townsend, J.T. Quantum dynamics of human decision-making. J. Math. Psychol. 2006, 50, 220–241. [Google Scholar] [CrossRef]
  8. Khrennikov, A. Quantum-like model of cognitive decision making and information processing. J. Biosyst. 2009, 95, 179–187. [Google Scholar] [CrossRef] [PubMed]
  9. Busemeyer, J.; Pothos, E.; Franco, R.; Trueblood, J. A quantum theoretical explanation for probability judgment errors. J. Psychol. Rev. 2011, 118, 193–218. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  10. Busemeyer, J.; Wang, Z. Quantum cognition: Key issues and discussion. Top. Cogn. Sci. 2014, 6, 43–46. [Google Scholar] [CrossRef] [PubMed]
  11. Wichert, A. Principles of Quantum Artificial Intelligence: Quantum Problem Solving and Machine Learning, 2nd ed.; World Scientific: Singapore, 2020. [Google Scholar]
  12. Wichert, A.; Moreira, C. Balanced quantum-like model for decision making. In Proceedings of the 11th International Conference on Quantum Interaction, Nice, France, 3–5 September 2018; pp. 79–90. [Google Scholar]
  13. Wichert, A.; Moreira, C.; Bruza, P. Quantum-like bayesian networks. Entropy 2020, 22, 170. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  14. Yukalov, V.; Sornette, D. Decision theory with prospect interference and entanglement. Theory Decis. 2011, 70, 283–328. [Google Scholar] [CrossRef] [Green Version]
  15. Luger, G.F.; Stubblefield, W.A. Artificial Intelligence, Structures and Strategies for Complex Problem Solving, 3rd ed.; Addison-Wesley: Boston, MA, USA, 1998. [Google Scholar]
  16. Mitchell, T. Machine Learning; McGraw-Hill: New York, NY, USA, 1997. [Google Scholar]
  17. Winston, P.H. Artificial Intelligence, 3rd ed.; Addison-Wesley: Boston, MA, USA, 1992. [Google Scholar]
Figure 1. f ( ω ) = ω Ω compared to g ( ω ) = ω 2 ω 2 + ( Ω ω ) 2 for Ω , (a) Ω = 4 , (b) Ω = 10 , (c) Ω = 100 and (d) Ω = 1000 . With growing size of Ω the function g ( x ) converge to a sigmoid function.
Figure 1. f ( ω ) = ω Ω compared to g ( ω ) = ω 2 ω 2 + ( Ω ω ) 2 for Ω , (a) Ω = 4 , (b) Ω = 10 , (c) Ω = 100 and (d) Ω = 1000 . With growing size of Ω the function g ( x ) converge to a sigmoid function.
Mathematics 09 02036 g001
Figure 2. Sigmoid functions: g ( x ) = x 2 x 2 + ( 10 x ) 2 versus logistic function σ ( x ) = 1 1 + e 0.9523 · ( x 5 ) .
Figure 2. Sigmoid functions: g ( x ) = x 2 x 2 + ( 10 x ) 2 versus logistic function σ ( x ) = 1 1 + e 0.9523 · ( x 5 ) .
Mathematics 09 02036 g002
Figure 3. (a) g ( x ) = x 2 x 2 + ( 10 x ) 2 and g ( x ) = 5 ( 10 + x ) x ( 50 10 x + x 2 ) 2 ; (b) g ( x ) = 5 ( 10 + x ) x ( 50 10 x + x 2 ) 2 versus N ( x | 5 , 1 . 9 2 ) = 1 1.9 2 π · e 1 2 · x 5 1.9 2 .
Figure 3. (a) g ( x ) = x 2 x 2 + ( 10 x ) 2 and g ( x ) = 5 ( 10 + x ) x ( 50 10 x + x 2 ) 2 ; (b) g ( x ) = 5 ( 10 + x ) x ( 50 10 x + x 2 ) 2 versus N ( x | 5 , 1 . 9 2 ) = 1 1.9 2 π · e 1 2 · x 5 1.9 2 .
Mathematics 09 02036 g003
Figure 4. Shannon’s entropy H with Ω = 100 , blue discrete plot for l 1 sampling and yellow discrete plot l 2 sampling.
Figure 4. Shannon’s entropy H with Ω = 100 , blue discrete plot for l 1 sampling and yellow discrete plot l 2 sampling.
Mathematics 09 02036 g004
Figure 5. Multiplying two independent sampled events x and y with Ω = 10 leads to different results: (a) p l 1 ( x , y ) , (b) p l 2 ( x , y ) , (c) counterplot of p l 1 ( x , y ) , (d) counterplot of p l 2 ( x , y ) .
Figure 5. Multiplying two independent sampled events x and y with Ω = 10 leads to different results: (a) p l 1 ( x , y ) , (b) p l 2 ( x , y ) , (c) counterplot of p l 1 ( x , y ) , (d) counterplot of p l 2 ( x , y ) .
Mathematics 09 02036 g005
Figure 6. Different H 1 and H 2 values for sampled events x and y with Ω = 10 : (a) H 1 ( x , y ) , (b) H 2 ( x , y ) , (c) counterplot of H 1 ( x , y ) , (d) counterplot of H 2 ( x , y ) .
Figure 6. Different H 1 and H 2 values for sampled events x and y with Ω = 10 : (a) H 1 ( x , y ) , (b) H 2 ( x , y ) , (c) counterplot of H 1 ( x , y ) , (d) counterplot of H 2 ( x , y ) .
Mathematics 09 02036 g006
Figure 7. The Titanic data set is represented in an Excel table that contains data for 891 of the real Titanic passengers, some entries are not defined.
Figure 7. The Titanic data set is represented in an Excel table that contains data for 891 of the real Titanic passengers, some entries are not defined.
Mathematics 09 02036 g007
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Wichert, A. Quantum-Like Sampling. Mathematics 2021, 9, 2036. https://doi.org/10.3390/math9172036

AMA Style

Wichert A. Quantum-Like Sampling. Mathematics. 2021; 9(17):2036. https://doi.org/10.3390/math9172036

Chicago/Turabian Style

Wichert, Andreas. 2021. "Quantum-Like Sampling" Mathematics 9, no. 17: 2036. https://doi.org/10.3390/math9172036

APA Style

Wichert, A. (2021). Quantum-Like Sampling. Mathematics, 9(17), 2036. https://doi.org/10.3390/math9172036

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop