Next Article in Journal
3D CNN-Based Speech Emotion Recognition Using K-Means Clustering and Spectrograms
Next Article in Special Issue
Conditional Rényi Divergence Saddlepoint and the Maximization of α-Mutual Information
Previous Article in Journal
Fluctuation Theorem of Information Exchange within an Ensemble of Paths Conditioned on Correlated-Microstates
Previous Article in Special Issue
Guessing with Distributed Encoders
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Distributed Hypothesis Testing with Privacy Constraints

by
Atefeh Gilani
1,
Selma Belhadj Amor
2,
Sadaf Salehkalaibar
1,* and
Vincent Y. F. Tan
2
1
Department of Electrical and Computer Engineering, College of Engineering, University of Tehran, Tehran 14171614418, Iran
2
Department of Electrical and Computer Engineering, National University of Singapore, Singapore 117583, Singapore
*
Author to whom correspondence should be addressed.
Entropy 2019, 21(5), 478; https://doi.org/10.3390/e21050478
Submission received: 6 February 2019 / Revised: 8 March 2019 / Accepted: 20 March 2019 / Published: 7 May 2019

Abstract

:
We revisit the distributed hypothesis testing (or hypothesis testing with communication constraints) problem from the viewpoint of privacy. Instead of observing the raw data directly, the transmitter observes a sanitized or randomized version of it. We impose an upper bound on the mutual information between the raw and randomized data. Under this scenario, the receiver, which is also provided with side information, is required to make a decision on whether the null or alternative hypothesis is in effect. We first provide a general lower bound on the type-II exponent for an arbitrary pair of hypotheses. Next, we show that if the distribution under the alternative hypothesis is the product of the marginals of the distribution under the null (i.e., testing against independence), then the exponent is known exactly. Moreover, we show that the strong converse property holds. Using ideas from Euclidean information theory, we also provide an approximate expression for the exponent when the communication rate is low and the privacy level is high. Finally, we illustrate our results with a binary and a Gaussian example.

1. Introduction

In the distributed hypothesis testing (or hypothesis testing with communication constraints) problem, some observations from the environment are collected by the sensors in a network. They describe these observations over the network which are finally received by the decision center. The goal is to guess the joint distribution governing the observations at terminals. In particular, there are two possible hypotheses H = 0 or H = 1 , where the joint distribution of the observations is specified under each of them. The performance of this system is characterized by two criteria: the type-I and the type-II error probabilities. The probability of deciding on H = 1 (respectively H = 0 ) when the original hypothesis is H = 0 (respectively H = 1 ) is referred to as the type-I error (type-II error) probability. There are several approaches for defining the performance of a hypothesis test. First, we can maximize the exponent (exponential rate of decay) of the Bayesian error probability. Second, we can impose that the type-II error probability decays exponentially fast and we can then maximize the exponent of the type-I error probability; this is known as the Hoeffding regime. The approach in this work is the Chernoff-Stein regime in which we upper bound the type-II error probability by a non-vanishing constant and we maximize the exponent of the type-II error probability.
A special case of interest is testing against independence where the joint distribution under H = 1 is the product of the marginals under H = 0 . The optimal exponent of type-II error probability for testing against independence is determined by Ahlswede and Csiszár in [1]. Several extensions of this basic problem are studied for a multi-observer setup [2,3,4,5,6], a multi-decision center setup [7,8] and a setup with security constraints [9]. The main idea of the achievable scheme in these works is typicality testing [10,11]. The sensor finds a jointly typical codeword with its observation and sends the corresponding bin index to the decision center. The final decision is declared based on typicality check of the received codeword with the observation at the center. We note that the coding scheme employed here is reminiscent of those used for source coding with side information [12] and for different variants of the information bottleneck problem [13,14,15,16].

1.1. Injecting Privacy Considerations into Our System

We revisit the distributed hypothesis testing problem from a privacy perspective. In many applications such as healthcare systems, there is a need to randomize the data before publishing it. For example, hospitals often have large amounts of medical records of their patients. These records are useful for performing various statistical inference tasks, such as learning about causes of a certain ailment. However, due to privacy considerations of the patients, the data cannot be published as is. The data needs to be sanitized, quantized, perturbed and then be fed to a management center before statistical inference, such as hypothesis testing, is being done.
In the proposed setup, we use a privacy mechanism to sanitize the observation at the terminal before it is compressed; see Figure 1. The compression is performed at a separate terminal called transmitter, which communicates the randomized data over a noiseless link of rate R to a receiver. The hypothesis testing is performed using the received data (the compression index and additional side information) to determine the correct hypothesis governing the original observations. The privacy criterion is defined by the mutual information [17,18,19,20] of the published and original data.
There is a long history of research to provide appropriate metrics to measure privacy. To quantify the information leakage an observation X ^ can induce on a latent variable X, Shannon’s mutual information I ( X ; X ^ ) is considered in [17,18,19,20]. Smith [18] proposed to use Arimoto’s mutual information of order , I ( X ; X ^ ) . Barthe and Köpf [21,22,23] proposed the maximal information leakage max P X I ( X ; X ^ ) . We refer the reader to [24] for a survey on the existing information leakage measures. A different line of works, in statistics, computer science, and other related fields, concerns differential privacy, initially proposed in [25]. Furthermore, a generalized notion— ( ϵ , δ ) -differential privacy [26]—provides a unified mathematical framework for data privacy. The reader is referred to the survey by Dwork [27] and the statistical framework studied by Wasserman and Zhou [28] and the references therein.
The privacy mechanism can be either memoryless or non-memoryless. In the former, the distribution of the randomized data at each time instant depends on the original sequence at the same time and not on the previous history of the data.

1.2. Description of Our System Model

We propose a coding scheme for the proposed setup. The idea is that the sensor, upon observing the source sequence, performs a typicality test and obtains its belief of the hypothesis. If the belief is H = 0 , it publishes the randomized data based on a specific memoryless mechanism. However, if its belief is H = 1 , it sends an all-zero sequence to let the transmitter know about its decision. The transmitter communicates the received data, which is a sanitized version of the original data or an all-zero sequence, over the noiseless link to the receiver. In this scheme, the whole privacy mechanism is non-memoryless since the typicality check of the source sequence which uses the history of the observation, determines the published data. It is shown that the achievable error exponent recovers previous results on hypothesis testing with zero and positive communication rates in [10]. Our work is related to a recent work [29] where a general hypothesis testing setup is considered from a privacy perspective. However, in [29], the problem at hand is different from ours. The authors consider equivocation and average distortion as possible measures of privacy whereas we constrain the mutual information between the original and released (published) data.
A difference of the proposed scheme with some previous works is highlighted as follows. The privacy mechanism even if it is memoryless, cannot be viewed as a noiseless link of a rate equivalent to the privacy criterion. Particularly, the proposed model is different from cascade hypothesis testing problem of [8] or similar works [3,4] which consider consecutive noiseless links for data compression and distributed hypothesis testing. The difference comes from the fact that in these works, a codeword is chosen jointly typical with the observed sequence at the terminal and its corresponding index is sent over the noiseless link. However, in our model, the randomized sequence is not necessarily jointly typical with the original sequence. Thus, there is a need for an achievable scheme which lets the transmitter know whether the original data is typical or not.
The problem of hypothesis testing against independence with a memoryless privacy mechanism is also considered. A coding scheme is proposed where the sensor outputs the randomized data based on the memoryless privacy mechanism. The optimality of the achievable type-II error exponent is shown by providing a strong converse. Specializing the optimal error exponent to a binary example shows that an increase in the privacy criterion (a less stringent privacy mechanism) results in a larger type-II error exponent. Thus, there exists a trade-off between privacy and hypothesis testing criteria. The optimal type-II error exponent is further studied for the case of restricted privacy mechanism and zero-rate communication. The Euclidean approach of [30,31,32,33] is used to approximate the error exponent for this regime. The result confirms the trade-off between the privacy criterion and type-II error exponent. Finally, a Gaussian setup is proposed and its optimal error exponent is established.

1.3. Main Contributions

The contributions of the paper are listed in the following:
  • An achievable type-II error exponent is proposed using a non-memoryless privacy mechanism (Theorem 1 in Section 3);
  • The optimal error exponent of testing against independence with a memoryless privacy mechanism is determined. In addition, a strong converse is also proved (Theorem 2 in Section 4.1);
  • A binary example is proposed to show the trade-off between the privacy and error exponent (Section 4.3);
  • An Euclidean approximation [30] of the error exponent is provided (Section 4.4);
  • A Gaussian setup is proposed and its optimal error exponent is derived (Proposition 2 in Section 4.5).

1.4. Notation

The notation mostly follows [34]. Random variables are denoted by capital letters, e.g., X, Y, and their realizations by lower case letters, e.g., x, y. The alphabet of the random variable X is denoted as X . Sequences of random variables and their realizations are denoted by ( X i , , X j ) and ( x i , , x j ) and are abbreviated as X i j and x i j . We use the alternative notation X j when i = 1 . Vectors and matrices are denoted by boldface letters, e.g., k , W . The 2 -norm of k is denoted as k . The notation k T denotes the transpose of k .
The probability mass function (pmf) of a discrete random variable X is denoted as P X , the conditional pmf of X given Y is denoted as P X | Y . The notation D ( P X Q X ) denotes the Kullback-Leibler (KL) divergence between two pmfs P X and Q X . The total variation distance between two pmfs P X and Q X is denoted by | P X Q X | = 1 2 x | P X ( x ) Q X ( x ) | . We use tp ( x n , y n ) to denote the joint type of ( x n , y n ) .
For a given P X Y and a positive number μ , we denote by T μ n ( P X Y ) , the set of jointly μ -typical sequences [34], i.e., the set of all ( x n , y n ) whose joint type is within μ of P X Y (in the sense of total-variation distance). The notation T n ( P X ) denotes for the type class of the type P X .
The notation h b ( · ) denotes the binary entropy function, h b 1 ( · ) its inverse over 0 , 1 2 , and a b a ( 1 b ) + ( 1 a ) b for 0 a , b 1 . The differential entropy of a continuous random variable X is h ( X ) . All logarithms log ( · ) are taken with respect to base 2.

1.5. Organization

The remainder of the paper is organized as follows. Section 2 describes a mathematical setup for our proposed problem. Section 3 discusses hypothesis testing with general distributions. The results for hypothesis testing against independence with a memoryless privacy mechanism are provided in Section 4. The paper is concluded in Section 5.

2. System Model

Let X , Y , and X ^ be arbitrary finite alphabets and let n be a positive integer. Consider the hypothesis testing problem with communication and privacy constraints depicted in Figure 1. The first terminal in the system, the Randomizer, receives the sequence X n = ( X 1 , , X n ) X n and outputs the sequence X ^ n = ( X ^ 1 , , X ^ n ) X ^ n , which is a noisy version of X n under a privacy mechanism determined by the conditional probability distribution P X ^ n | X n ; the second terminal, the Transmitter, receives the sequence X ^ n ; the third terminal, the Receiver, observes the side-information sequence Y n = ( Y 1 , , Y n ) Y n . Under the null hypothesis
H = 0 : ( X n , Y n ) i . i . d . P X Y ,
whereas under the alternative hypothesis
H = 1 : ( X n , Y n ) i . i . d . Q X Y ,
for two given pmfs P X Y and Q X Y .
The privacy mechanism is described by the conditional pmf P X ^ n | X n which maps each sequence X n X n to a sequence X ^ n X ^ n . For any ( x ^ n , x n , y n ) X ^ n × X n × Y n , the joint distributions considering the privacy mechanism are given by
P X ^ X Y n ( x ^ n , x n , y n ) P X ^ n | X n ( x ^ n | x n ) · i = 1 n P X Y ( x i , y i ) ,
Q X ^ X Y n ( x ^ n , x n , y n ) P X ^ n | X n ( x ^ n | x n ) · i = 1 n Q X Y ( x i , y i ) .
A memoryless/local privacy mechanism is defined by a conditional pmf P X ^ | X which stochastically and independently maps each entry X i X of X n to a released X ^ i X ^ to construct X ^ n . Consequently, for the memoryless privacy mechanism, the conditional pmf P X ^ n | X n ( x ^ n | x n ) factorizes as follows:
P X ^ n | X n ( x ^ n | x n ) = i = 1 n P X ^ | X ( x ^ i | x i ) = P X ^ | X n ( x ^ n | x n ) , ( x ^ n , x n ) X ^ n × X n .
There is a noise-free bit pipe of rate R from the transmitter to the receiver. Upon observing X ^ n , the transmitter computes the message M = ϕ ( n ) ( X ^ n ) using a possibly stochastic encoding function ϕ ( n ) : X ^ n { 0 , , 2 n R } and sends it over the bit pipe to the receiver.
The goal of the receiver is to produce a guess of H using a decoding function g ( n ) : Y n × { 0 , . . . , 2 n R } { 0 , 1 } based on the observation Y n and the received message M. Thus the estimate of the hypothesis is H ^ = g ( n ) ( Y n , M ) .
This induces a partition of the sample space X ^ n × X n × Y n into an acceptance region A n defined as follows:
A n ( x ^ n , x n , y n ) : g ( n ) ( y n , ϕ ( n ) ( x ^ n ) ) = 0 ,
and a rejection region denoted by A n c .
Definition 1.
For any ϵ [ 0 , 1 ) and for a given rate-privacy pair ( R , L ) R + 2 , we say that a type-II exponent θ R + is ( ϵ , R , L ) -achievable if there exists a sequence of functions and conditional pmfs ( ϕ ( n ) , g ( n ) , P X ^ n | X n ) , such that the corresponding sequences of type-I and type-II error probabilities at the receiver are defined as
α n P X ^ X Y n ( A n c ) a n d β n Q X ^ X Y n ( A n ) ,
respectively, and they satisfy
lim sup n α n ϵ and lim inf n 1 n log 1 β n θ .
Furthermore, the privacy measure
T n 1 n I ( X n ; X ^ n ) ,
satisfies
lim sup n T n L .
The optimal exponent θ ϵ * ( R , L ) is the supremum of all ( ϵ , R , L ) -achievable θ R + .

3. General Hypothesis Testing

3.1. Achievable Error Exponent

The following presents an achievable error exponent for the proposed setup.
Theorem 1.
For a given ϵ [ 0 , 1 ) and a rate-privacy pair ( R , L ) R + 2 , the optimal type-II error exponent θ ϵ * ( R , L ) for the multiterminal hypothesis testing setup under the privacy constraint L and the rate constraint R satisfies
r C l θ ϵ * ( R , L ) max P U | X ^ , P X ^ | X : R I ( U ; X ^ ) L I ( X ; X ^ ) min P ˜ U X ^ X Y P U X ^ X Y D ( P ˜ U X ^ X Y P U | X ^ P X ^ | X Q X Y ) ,
where the set P U X ^ X Y is defined as
P U X ^ X Y P ˜ U X ^ X Y P ˜ X = P X , P ˜ U Y = P U Y , P ˜ U X ^ = P U X ^ .
Given P U | X ^ and P X ^ | X , the mutual informations in (11) are calculated according to the following joint distribution:
P U X ^ | X P U | X ^ · P X ^ | X .
Proof. 
The coding scheme is given in the following section. For the analysis, see Appendix A. □

3.2. Coding Scheme

In this section, we propose a coding scheme for Theorem 1, under fixed rate and privacy constraints ( R , L ) R + 2 . Fix the joint distribution P U X ^ X Y as in (13). Let P U ( u ) be the marginal distribution of U U defined as
P U ( u ) x ^ X ^ P U | X ^ ( u | x ^ ) x X P X ^ X ( x ^ , x ) .
Fix positive μ > 0 and ζ > 0 , an arbitrary blocklength n and two conditional pmfs P X ^ | X and P U | X ^ over finite auxiliary alphabets X ^ and U . Fix also the rate and privacy leakage level as
R = I ( U ; X ^ ) + μ , and L = I ( X ^ ; X ) + ζ .
Codebook Generation: Randomly and independently generate a codebook
C U U n ( m ) : m { 0 , , 2 n R } ,
by drawing U n ( m ) in an i.i.d. manner according to P U . The codebook is revealed to all terminals.
Randomizer: Upon observing x n , it checks whether x n T μ / 4 n ( P X ) . If successful, it outputs the sequence x ^ n where its i-th component x ^ i is generated based on x i , according to P X ^ | X ( x ^ i | x i ) . If the typicality check is not successful, the randomizer then outputs 0 n which is an all-zero sequence of length n, where x ^ n = 0 n .
Transmitter: Upon observing x ^ n , if x ^ n 0 n , the transmitter finds an index m such that u n ( m ) , x ^ n T μ / 2 n ( P U X ^ ) . If successful, it sends the index m over the noiseless link to the receiver. Otherwise, if the typicality check is not successful or x ^ n = 0 n , it sends m = 0 .
Receiver: Upon observing y n and receiving the index m, if m = 0 , the receiver declares H ^ = 1 . If m 0 , it checks whether u n ( m ) , y n T μ n ( P U Y ) . If the test is successful, the receiver declares H ^ = 0 ; otherwise, it sets H ^ = 1 .
Remark 1.
In the above scheme, the sequence X ^ n is chosen to be an n-length zero-sequence when the randomizer finds that X n is not typical according to P X . Thus, the privacy mechanism is not memoryless and the sequence X ^ n is not identically and independently distributed (i.i.d.). A detailed analysis in Appendix A shows that the privacy criterion is not larger than L as the blocklength n .

3.3. Discussion

In the following, we discuss some special cases. First, suppose that R = 0 . The following corollary shows that Theorem 1 recovers Han’s result [1] for distributed hypothesis testing with zero-rate communication.
Corollary 1
(Theorem 5 in [10]). Suppose that Q X Y > 0 . For all ϵ [ 0 , 1 ) , the optimal error exponent of the zero-rate communication for any privacy mechanism (including non-memoryless mechanisms) is given by the following:
θ ϵ * ( 0 , L ) = min P ˜ X Y : P ˜ X = P X P ˜ Y = P Y D ( P ˜ X Y Q X Y ) .
Proof. 
The proof of achievability follows by Theorem 1, in which X ^ is arbitrary and the auxiliary U = due to the zero-rate constraint. The proof of the strong converse follows along the same lines as [35]. □
Remark 2.
Consider the case of R > 0 and L = 0 where X ^ is independent of X. Using Theorem 1, the optimal error exponent is lower bounded as follows:
θ ϵ * ( R , 0 ) min P ˜ X Y : P ˜ X = P X P ˜ Y = P Y D ( P ˜ X Y Q X Y ) .
However, there is no known converse result in this case where the communication rate is positive. Comparing this special case with the one in Corollary 1 shows that the proposed model does not, in general, admit symmetry between the rate and privacy constraints. However, we will see from some specific examples in the following that the roles of R and L are symmetric.
Now, suppose that L is so large such that L > H ( X ) . The following corollary shows that Theorem 1 recovers Han’s result in [10] for distributed hypothesis testing over a rate-R communication link.
Corollary 2
(Theorem 2 in [10]). Assuming L > H ( X ) , the optimal error exponent is lower bounded as the following:
θ ϵ * ( R , L ) max P U | X : R I ( U ; X ) min P ˜ U X Y : P ˜ U X = P U X P ˜ U Y = P U Y D ( P ˜ U X Y P U | X Q X Y ) .
Proof. 
The proof follows from Theorem 1 by specializing to X ^ = X . □
The above two special cases reveal a trade-off between the privacy criterion and the achievable error exponent when the communication rate is positive, i.e., R > 0 . An increase in L results in a larger achievable error exponent. This observation is further illustrated by an example in Section 4.3 to follow.

4. Hypothesis Testing against Independence with a Memoryless Privacy Mechanism

In this section, we consider testing against independence where the joint pmf under H = 1 factorizes as follows:
Q X Y = P X · P Y .
The privacy mechanism is assumed to be memoryless here.

4.1. Optimal Error Exponent

The following theorem, which includes a strong converse, states the optimal error exponent for this special case.
Theorem 2.
For any ( R , L ) R + 2 , define
θ ϵ * ( R , L ) = max P U | X ^ , P X ^ | X : R I ( U ; X ^ ) L I ( X ; X ^ ) I ( U ; Y ) .
Then, for any ϵ [ 0 , 1 ) and any ( R , L ) R + 2 , the optimal error exponent for testing against independence when using a memoryless privacy mechanism is given by (21), where it suffices to choose | U | | X ^ | + 1 and | X ^ | | X | according to Caratheodory’s theorem [36] (Theorem 15.3.5).
Proof. 
The coding scheme is given in the following section. For the rest of proof, see Appendix B. □

4.2. Coding Scheme

In this section, we propose a coding scheme for Theorem 2. Fix the joint distribution as in (13), and the rate and privacy constraints as in (15). Generate the codebook C U as in (16).
Randomizer: Upon observing x n , it outputs the sequence x ^ n in which the i-th component x ^ i is generated based on x i , according to P X ^ | X ( x ^ i | x i ) .
Transmitter: It finds an index m such that u n ( m ) , x ^ n T μ / 2 n ( P U X ^ ) . If successful, it sends the index m over the noiseless link to the receiver. Otherwise, it sends m = 0 .
Receiver: Upon observing y n and receiving the index m, if m = 0 , the receiver declares H ^ = 1 . If m 0 , it checks whether u n ( m ) , y n T μ n ( P U Y ) . If the test is successful, the receiver declares H ^ = 0 ; otherwise, it sets H ^ = 1 .
Remark 3.
In the above scheme, the sequence X ^ n is i.i.d. since it is generated based on the memoryless mechanism P X ^ | X .
When the communication rate is positive, there exists a trade-off between the optimal error exponent and the privacy criterion. The following example elucidates this trade-off.

4.3. Binary Example

In this section, we study hypothesis testing against independence for a binary example. Suppose that under both hypotheses, we have X Bern ( 1 2 ) . Under the null hypothesis,
H = 0 : Y = X N , N Bern ( q )
for some 0 q 1 , where N is independent of X. Under the alternative hypothesis
H = 1 : Y Bern 1 2 ,
where Y is independent of X. The cardinality constraint shows that it suffices to choose | X ^ | = 2 . Among all possible privacy mechanisms, the choice of P X ^ | X ( 1 | 0 ) = P X ^ | X ( 1 | 1 ) and P X ^ | X ( 0 | 0 ) = P X ^ | X ( 0 | 1 ) minimizes the mutual information I ( X ; X ^ ) . Thus, we restrict to this choice which also results in X ^ Bern 1 2 .
The cardinality bound on the auxiliary random variable U is | U | 3 . The following proposition states that it is also optimal to choose P U | X ^ to be a BSC.
Proposition 1.
The optimal error exponent of the proposed binary setup is given by the following:
θ ϵ * ( R , L ) = 1 h b q h b 1 ( 1 L ) h b 1 ( 1 R ) .
Proof. 
For the proof of achievability, choose the following auxiliary random variables:
X ^ = X Z ^ , Z ^ Bern ( p 1 )
U = X ^ Z , Z Bern ( p 2 ) ,
for some 0 p 1 , p 2 1 where Z ^ and Z are independent of X and ( X , X ^ ) , respectively. The optimal error exponent of Theorem 2 reduces to the following:
θ ϵ * ( R , L ) = max 0 p 1 , p 2 1 : R 1 h b ( p 2 ) L 1 h b ( p 1 ) 1 h b ( q p 1 p 2 ) ,
which can be simplified to (24). For the proof of the converse, see Appendix C. □
Figure 2 illustrates the error exponent versus the privacy parameter L for a fixed rate R. There is clearly a trade-off between θ ϵ * ( R , L ) and L. For a less stringent privacy requirement (large L), the error exponent θ ϵ * ( R , L ) increases.

4.4. Euclidean Approximation

In this section, we propose Euclidean approximations [30,31] for the optimal error exponent of testing against independence scenario (Theorem 2) when R 0 and L 0 . Consider the optimal error exponent as follows:
θ ϵ * ( R , L ) = max P U | X ^ , P X ^ | X : R I ( U ; X ^ ) L I ( X ; X ^ ) I ( U ; Y ) .
Let W of dimension | Y | × | X | , denote the transition matrix P Y | X , which is itself induced by P X and the joint distribution P X Y . Now, consider the rate constraint as follows:
I ( U ; X ^ ) = u U P U ( u ) D P X ^ | U ( · | u ) P X ^ R .
Assuming R 0 , we let P X ^ | U ( · | u ) be a local perturbation from P X ^ ( · ) , where we have
P X ^ | U ( · | u ) = P X ^ ( · ) + ψ u ( · ) ,
for a perturbation ψ u ( · ) satisfying
x ^ X ^ ψ u ( x ^ ) = 0 ,
in order to preserve the row stochasticity of P X ^ | U . Using a χ 2 -approximation [30], we can write:
D P X ^ | U ( · | u ) P X ^ 1 2 · log e · v u 2 ,
where v u denotes the length- | X ^ | column vector of weighted perturbations whose x ^ -th component is defined as:
v u ( x ^ ) 1 P X ^ ( x ^ ) · ψ u ( x ^ ) , x ^ X ^ .
Using the above definition, the rate constraint in (29) can be written as:
u U P U ( u ) v u 2 2 R log e .
Similarly, consider the privacy constraint as the following:
I ( X ; X ^ ) = x ^ X ^ P X ^ ( x ^ ) D P X | X ^ ( · | x ^ ) P X L .
Assuming L 0 , we let P X | X ^ ( · | x ^ ) be a local perturbation from P X ( · ) where
P X | X ^ ( · | x ^ ) = P X ( · ) + ϕ x ^ ( · ) ,
for a perturbation ϕ x ^ ( · ) that satifies:
x X ϕ x ^ ( x ) = 0 .
Again, using a χ 2 -approximation, we obtain the following:
D P X | X ^ ( · | x ^ ) P X 1 2 log e v x ^ 2 ,
where v x ^ is a length- | X | column vector and its x-th component is defined as follows:
v x ^ ( x ) 1 P X ( x ) · ϕ x ^ ( x ) , x X .
Thus, the privacy constraint in (35) can be written as:
x ^ X ^ P X ^ ( x ^ ) v x ^ 2 2 L log e .
For any x X and u U , we define the following:
(41) Λ u ( x ) x ^ X ^ ψ u ( x ^ ) ϕ x ^ ( x ) (42) = P X ( x ) x ^ X ^ P X ^ ( x ^ ) v u ( x ^ ) v x ^ ( x ) ,
and the corresponding length- | X | column vector Λ u defined as follows:
Λ u = P X V X ^ P X ^ v u ,
where P X denotes a diagonal | X | × | X | -matrix, so that its ( x , x ) -th element ( x X ) is P X ( x ) , and P X ^ is defined similarly. Moreover, V X ^ refers to the | X | × | X ^ | -matrix defined as follows:
V X ^ v 1 v 2 v x ^ v | X ^ | .
Let P Y 1 be the inverse of diagonal | Y | × | Y | -matrix P Y . As shown in Appendix D, the optimization problem in (28) can be written as follows:
max { v u } u U , V X ^ : P X ^ ( x ^ ) v u ( x ^ ) 1 P X ^ ( x ^ ) P X ^ ( x ^ ) P X ( x ) v x ^ ( x ) 1 P X ( x ) P X ( x ) 1 2 log e u U P U ( u ) · P Y 1 W P X V X ^ P X ^ V u 2
subject to : u U P U ( u ) v u 2 2 R log e ,
x ^ X ^ P X ^ ( x ^ ) v x ^ 2 2 L log e .
The following example specializes the above approximation to the binary case.
Example 1.
Consider the binary setup of Example 4.3 and the choice of auxiliary random variables in (26). Since the privacy mechanism is assumed to be a BSC, we have
P X = 1 2 1 2 T , P X ^ = 1 2 1 2 T ,
Now, we consider the vectors v u = 0 and v u = 1 defined as
v u = 0 = 2 ξ 1 2 ξ 1 T ,
v u = 1 = 2 ξ 1 2 ξ 1 T .
for some positive ξ 1 . This yields the following:
P X ^ | U = 0 = P X ^ + ξ 1 ξ 1 T ,
P X ^ | U = 1 = P X ^ + ξ 1 ξ 1 T
We also choose the vectors v x ^ = 0 and v x ^ = 1 as follows:
v x ^ = 0 = 2 ξ 2 2 ξ 2 T ,
v x ^ = 1 = 2 ξ 2 2 ξ 2 T ,
which results in
P X | X ^ = 0 = P X + ξ 2 ξ 2 T ,
P X | X ^ = 1 = P X + ξ 2 ξ 2 T .
Notice that the matrix W is given by
W = 1 q q q 1 q .
Thus, the optimization problem in (45) and (47) reduces to the following:
max 1 2 ξ 1 , ξ 2 1 2 8 log e ( 1 2 q ) 2 | ξ 1 | 2 | ξ 2 | 2
s u b j e c t t o : 4 | ξ 1 | 2 2 R log e a n d 4 | ξ 2 | 2 2 L log e .
Solving the above optimization yields
θ ϵ * ( R 0 , L 0 ) 2 log e ( 1 2 q ) 2 R L .
For some values of parameters, the approximation in (60) is compared to the error exponent of (24) in Figure 3. We observe that when R = L 0 , the approximation turns out to be excellent.
Remark 4.
The trade-off between the optimal error exponent and the privacy can again be verified from (60) in the case of L 0 and R 0 . As L becomes larger (which corresponds to a less stringent privacy requirement), the error exponent also increases. For a fixed error exponent, a trade-off between R and L exists. An increase in R results in a decrease of L.

4.5. Gaussian Setup

In this section, we consider hypothesis testing against independence over a Gaussian example. Suppose that X N ( 0 , 1 ) and under the null hypothesis H = 0 , the sources X and Y are jointly Gaussian random variables distributed as N ( 0 , G X Y ) , where G X Y is defined as the following:
G X Y = Δ 1 ρ ρ 1 ,
for some 0 ρ 1 .
Under the alternative hypothesis H = 1 , we assume that X and Y are independent Gaussian random variables, each distributed as N ( 0 , 1 ) . Consider the privacy constraint as follows:
L I ( X ; X ^ ) = h ( X ) h ( X | X ^ ) .
For a Gaussian source X, the conditional entropy h ( X | X ^ ) is maximized for a jointly Gaussian ( X , X ^ ) . This choice minimizes the RHS of (62). Thus, without loss of optimality, we choose
X = X ^ + Z , Z N 0 , 2 2 L ,
where Z is independent of X ^ . The following proposition states that it is optimal to choose U jointly Gaussian with ( X , X ^ , Y ) .
Proposition 2.
The optimal error exponent of the proposed Gaussian setup is given by
θ ϵ * ( R , L ) = 1 2 log 1 1 ρ 2 · ( 1 2 2 R ) · ( 1 2 2 L ) .
Proof. 
For the proof of achievability, we choose X ^ as in (63). Also, let
X ^ = U + Z ^ , Z ^ N ( 0 , β 2 ) ,
for some β 2 0 , where Z ^ is independent of U. It can be shown that Theorem 2 remains valid when it is extended to the continuous alphabet [5]. For the details of the simplification and also the proof of converse, see Appendix E. □
Remark 5.
If L = , the above proposition recovers the optimal error exponent of Rahman and Wagner [5] (Corollary 7) for testing against independence of Gaussian sources over a noiseless link of rate R.

5. Summary and Discussion

In this paper, distributed hypothesis testing with privacy constraints is considered. A coding scheme is proposed where the sensor decides on one of hypotheses and generates the randomized data based on its decision. The transmitter describes the randomized data over a noiseless link to the receiver. The privacy mechanism in this scheme is non-memoryless. The special case of testing against independence with a memoryless privacy mechanism is studied in detail. The optimal type-II error exponent of this case is established, together with a strong converse. A binary example is proposed where the trade-off between the privacy criterion and the error exponent is reported. Euclidean approximations are provided for the case in which the privacy level is high and and the communication rate is vanishingly small. The optimal type-II error exponent of a Gaussian setup is also established.
A future line of research is to study the second-order asymptotics of our model. The second-order analysis of distributed hypothesis testing without privacy constraints and with zero-rate communication was studied in [37]. In our proposed model, the trade-off between the privacy and type-II error exponent is observed, i.e., a less stringent privacy requirement yields a larger error exponent. The next step is to see whether the trade-off between privacy and error exponent affects the second-order term.
Another potential line for future research is to consider other metrics of privacy instead of the mutual information. A possible candidate is to use the maximal leakage [21,22,23] and to analyze the performance in tandem with distributed hypothesis testing problem.

Author Contributions

Investigation, A.G.; Supervision, S.S. and V.Y.F.T.; Writing—original draft, S.B.A.

Funding

This research was partially funded by grants R-263-000-C83-112 and R-263-000-C54-114.

Acknowledgments

The authors would like to thank Lin Zhou (National University of Singapore) and Daming Cao (Southeast University) for helpful discussions during the preparation of the manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Proof of Theorem 1

The analysis is based on the scheme described in Section 3.2.
Error Probability Analysis: We analyze the type-I and type-II error probabilities averaged over all random codebooks. By standard arguments [36] (p. 204), it can be shown that there exists at least one codebook that satisfies constraints on the error probabilities.
For the considered μ > 0 and the considered blocklength n, let P μ n be the set of all joint types π U X ^ X Y over U n × X ^ n × X n × Y n which satisfy the following constraints:
| π X P X | μ / 4 ,
| π U X ^ P U X ^ | μ / 2 ,
| π U Y P U Y | μ .
First, we analyze the type-I error probability. For the case of M 0 , we define the following event:
E ( U n ( M ) , Y n ) T μ n ( P U Y ) .
Thus, type-I error probability can be upper bounded as follows:
(A5) α n Pr X ^ n = 0 n or M = 0 or E | H = 0 (A6) Pr X ^ n = 0 n | H = 0 + Pr M = 0 | X ^ n 0 n , H = 0 + Pr E | M 0 , X ^ n 0 n , H = 0 (A7) ϵ / 3 + Pr M = 0 | X ^ n 0 n , H = 0 + Pr E | M 0 , X ^ n 0 n , H = 0 (A8) ϵ / 3 + ϵ / 3 + Pr E | M 0 , X ^ n 0 n , H = 0 (A9) ϵ / 3 + ϵ / 3 + ϵ / 3 (A10) = ϵ ,
where (A7) follows from AEP [36] (Theorem 3.1.1); (A8) follows from the covering lemma [34] (Lemma 3.3) and the rate constraint (15), (A10) follows from Markov lemma [34] (Lemma 12.1). In all justifications, n is taken to be sufficiently large.
Next, we analyze the type-II error probability. The acceptance region at the receiver is
A n Rx = m x ^ n , x n , y n : x ^ n 0 n , u n ( m ) , x ^ n , x n , y n T μ n ( P U X ^ X Y ) .
The set A n Rx is contained within the following acceptance region A ¯ n :
A ¯ n = m x ^ n , x n , y n : x ^ n 0 n , u n ( m ) , x ^ n , x n , y n π P μ n T n ( π ) .
Let F m { U n ( m ) , X ^ n , X n , Y n P μ n } . Therefore, the average of type-II error probability over all codebooks is upper bounded as follows:
(A13) E C [ β n ] Q X ^ X Y n A ¯ n (A14) m Pr X ^ n 0 n , F m | H = 1 (A15) m Pr F m | X ^ n 0 n , H = 1 (A16) 2 n R · ( n + 1 ) | U | · | X ^ | · | X | · | Y | · max π U X ^ X Y P μ n 2 n D ( π U X ^ X Y P U P X ^ | X Q X Y ) (A17) = ( n + 1 ) | U | · | X ^ | · | X | · | Y | · 2 n θ ˜ μ ,
where
θ ˜ μ min π U X ^ X Y P μ n D ( π U X ^ X Y P U P X ^ | X Q X Y ) R ,
and (A16) follows from the upper bound of Sanov’s theorem [36] (Theorem 11.4.1). Hence,
(A19) θ ˜ μ = min π U X ^ X Y P μ n D ( π U X ^ X Y P U P X ^ | X Q X Y ) R (A20) = min π U X ^ X Y P μ n D ( π U X ^ X Y P U P X ^ | X Q X Y ) I ( U ; X ^ ) μ (A21) = min π U X ^ X Y P μ n D ( π U X ^ X Y P U | X ^ P X ^ | X Q X Y ) + δ ( μ ) ,
where δ ( μ ) 0 as μ 0 . Equality (A20) follows from the rate constraint in (15) and (A21) holds because | π U X ^ P U X ^ | < μ / 2 .
Privacy Analysis: We first analyze the privacy under H = 0 . Notice that X ^ n is not necessarily i.i.d. because according to the scheme in Section 3.2, X ^ n is forced to be an all-zero sequence if the Randomizer decides that X n is not typical. However, conditioned on the event that X n T μ n ( P X ) , the sequence X ^ n is i.i.d. according to the conditional pmf P X ^ | X . The privacy measure T n satisfies
n T n = I ( X n ; X ^ n ) = H ( X ^ n ) H ( X ^ n | X n ) .
We now provide a lower bound on H ( X ^ n | X n ) as follows
H ( X ^ n | X n ) x n T μ n ( P X ) P X n ( x n ) H ( X ^ n | X n = x n )
For any x n T μ n ( P X ) and for μ > μ , it holds that
(A24) H ( X ^ n | X n = x n ) = x ^ n X ^ n P X ^ | X n ( x ^ n | x n ) log P X ^ | X n ( x ^ n | x n ) (A25) x ^ n T μ n ( P X ^ | X ( · | x n ) ) P X ^ | X n ( x ^ n | x n ) log P X ^ | X n ( x ^ n | x n ) (A26) x ^ n T μ n ( P X ^ | X ( · | x n ) ) P X ^ | X n ( x ^ n | x n ) log 2 n ( 1 μ ) H ( X ^ | X ) (A27) n ( 1 μ ) 2 H ( X ^ | X )
where (A26) is true because for any x ^ n T μ n ( P X ^ | X ( · | x n ) ) , it holds that P X ^ | X n ( x ^ n | x n ) 2 n ( 1 μ ) H ( X ^ | X ) , and (A27) follows because the conditional typicality lemma [34] (Chapter 2) implies that P X ^ | X n ( T μ n ( P X ^ | X ( · | x n ) | x n ) 1 μ for n sufficiently large.
Combining (A23) and (A27), we obtain
(A28) H ( X ^ n | X n ) n ( 1 μ ) 2 H ( X ^ | X ) x n T μ n ( P X ) P X n ( x n ) (A29) n ( 1 μ ) 2 ( 1 μ ) H ( X ^ | X ) ,
where (A29) follows because the AEP [36] (Theorem 3.1.1) implies that P X n ( T μ n ( P X ) ) 1 μ for n sufficiently large.
Hence, we have
(A30) I ( X n ; X ^ n ) = H ( X ^ n ) H ( X ^ n | X n ) (A31) n H ( X ^ ) H ( X ^ n | X n ) (A32) n H ( X ^ ) n ( 1 μ ) H ( X ^ | X ) (A33) = n I ( X ; X ^ ) + n μ H ( X ^ | X ) (A34) n L + n μ · log | X ^ | (A35) = n L + n ζ ,
where μ 1 ( 1 μ ) 2 ( 1 μ ) 0 , and ζ μ · log | X ^ | .
Next, consider the privacy analysis under H = 1 . Please note that when P X = Q X , the analysis is similar to that of H = 0 . Thus, we assume that P X Q X in the following. From (A22), the privacy measure T n satisfies:
n T n = I ( X n ; X ^ n ) H ( X ^ n ) .
To upper bound H ( X ^ n ) , we calculate the probability P X ^ n ( x ^ n ) for x ^ n = 0 n as follows:
(A37) P X ^ n ( 0 n ) = x n T μ n ( P X ) P X ^ | X n ( 0 n | x n ) · Q X n ( x n ) + x n T μ n ( P X ) P X ^ | X n ( 0 n | x n ) · Q X n ( x n ) (A38) = x n T μ n ( P X ) P X ^ | X n ( 0 n | x n ) · Q X n ( x n ) (A39) = x n T μ n ( P X ) Q X n ( x n ) (A40) = 1 Q X n T μ n ( P X ) (A41) 1 2 n ( D ( P X Q X ) + δ ( μ ) ) 1 γ n ,
where γ n 0 exponentially fast as n . Here, (A38) follows because if x n T μ n ( P X ) , then P X ^ | X n ( 0 n | x n ) = 0 , (A39) follows because when x n T μ n ( P X ) , then P X ^ | X n ( 0 n | x n ) = 1 and (A41) follows from Sanov’s theorem and the continuity of the relative entropy in its first argument [38] (Lemma 1.2.7).
Write H ( X ^ n ) as H ( P X ^ n ) and let P 0 n be the distribution on X ^ n that places all its probability mass on 0 n X ^ n . Since H ( P 0 n ) = 0 , by the uniform continuity of entropy [38] (Lemma 1.2.7),
H ( P X ^ n ) 2 | P X ^ n P 0 n | · log | X ^ | n 2 | P X ^ n P 0 n | .
Since γ n 0 exponentially fast, the same holds true for | P X ^ n P 0 n | and so by (A42), H ( P X ^ n ) = H ( X ^ n ) 0 . Therefore, under H = 1 , we have T n 0 as n .
Letting n and then letting μ , μ , γ 0 , we obtain θ ˜ μ θ and lim sup n n T n L , with θ given by the RHS of (11). This establishes the proof of Theorem 1.

Appendix B. Proof of Theorem 2

Achievability: The analysis is based on the scheme of Section 4.2. It follows similar steps as in [1]. Recall the definition of the event E in (A4). Consider the type-I error probability as follows:
(A43) α n Pr M = 0 or E | H = 0 (A44) Pr M = 0 | H = 0 + Pr E | M 0 , H = 0 (A45) ϵ / 2 + ϵ / 2 (A46) = ϵ ,
where (A46) follows from covering lemma [34] (Lemma 3.3) and the rate constraint in (15), and also the Markov lemma [34] (Lemma 12.1). Now, consider the type-II error probability as follows:
(A47) β n = Pr [ H ^ = 0 | H = 1 ] (A48) = Pr [ H ^ = 0 , M 0 | H = 1 ] (A49) Pr [ H ^ = 0 | H = 1 , M 0 ] (A50) = Pr [ H ^ = 0 | H = 1 , M = 1 ] ,
where the last equality follows from the symmetry of the code construction. Now, the average of type-II error probability over all codebooks satisfies:
E C β n 2 n [ I ( U ; Y ) δ ( μ ) ] ,
where δ ( μ ) is a function that tends to zero as μ 0 . The privacy analysis is straightforward since the privacy mechanism is memoryless whence we have
1 n I ( X n ; X ^ n ) = I ( X ; X ^ ) = L + ζ ,
where the last equality follows from the privacy constraint in (15). This concludes the proof of achievability.
Converse: Now, we prove the strong converse. It involves an extension of the η -image characterization technique [4,38]. The proof steps are given as follows. First, we find a truncated distribution P X ^ ̲ n which is arbitrarily close to P X ^ n in terms of entropy. Then, we analyze the type-II error probability under a constrained type-I error probability. Finally, a single-letter characterization of the rate and privacy constraints is given.
(1)
Construction of a Truncated Distribution:
Since the privacy machanism is memoryless, we conclude that ( X n , X ^ n , Y n ) is i.i.d. according to P X X ^ Y P X ^ | X P X Y . For a given P X ^ Y , define V n ( y n | x ^ n ) P Y | X ^ n ( y n | x ^ n ) for all x ^ n X ^ n and y n Y n . A set B Y n is an η -image of the set A X ^ n over the channel V n if
V n B | x ^ n η , x ^ n A .
The privacy mechanism is the same under both hypotheses, thus, we can define the acceptance region based on ( x ^ n , y n ) as follows:
A n ( x ^ n , y n ) : g ( n ) ( y n , ϕ ( n ) ( x ^ n ) ) = 0 .
For any encoding function ϕ ( n ) and an acceptance region A n X ^ n × Y n , let τ n denote the cardinality of codebook and define the following sets:
C i = Δ x ^ n X ^ n : ϕ ( n ) ( x ^ n ) = i ,
D i = Δ y n Y n : g ( n ) ( y n , i ) = 0 , 1 i τ n .
The acceptance region can be written as follows:
A n = i = 1 τ n C i × D i ,
where C i C j = ϕ for all i j . Define the set B n ( η ) as follows:
B n ( η ) x ^ n : V n D ϕ ( n ) ( x ^ n ) | x ^ n η .
Fix ϵ [ 0 , 1 ) and notice that the type-I error probability is upper-bounded as
α n = P X ^ Y n A n c ϵ ,
which we can write equivalently as
(A60) 1 ϵ P X ^ Y n A n (A61) = x ^ n B n ( η ) P X ^ n ( x ^ n ) V n D ϕ ( n ) ( x ^ n ) | x ^ n + x ^ n B n c ( η ) P X ^ n ( x ^ n ) V n D ϕ ( n ) ( x ^ n ) | x ^ n (A62) P X ^ n B n ( η ) + η 1 P X ^ n ( B n ( η ) ) ,
where the first term is because V n D ϕ ( n ) ( x ^ n ) | x ^ n 1 ; and the second term is because for any x ^ n B n c ( η ) , we have V n D ϕ ( n ) ( x ^ n ) | x ^ n < η .
In what follows, let η = 1 ϵ 2 . Inequality (A62) implies
P X ^ n ( B n ( η ) ) 1 ϵ 1 + ϵ .
Let μ n = n 1 / 3 . For the typical set T μ n n ( P X ^ ) , we have
P X ^ n T μ n n ( P X ^ ) 1 | X ^ | 2 μ n n .
Hence,
(A65) P X ^ n T μ n n ( P X ^ ) B n ( η ) P X ^ n T μ n n ( P X ^ ) + P X ^ n B n ( η ) 1 (A66) 1 ϵ 1 + ϵ | X ^ | 2 μ n n .
For any 0 < δ < 1 ϵ 1 + ϵ and for sufficiently large n,
P X ^ n T μ n n ( P X ^ ) B n ( η ) δ .
We can also write T μ n n ( P X ^ ) as
T μ n n ( P X ^ ) = P ^ X ^ : | P ^ X ^ P X ^ | μ n T n ( P ^ X ^ ) .
Combining the above equations, we get
P ^ X ^ : | P ^ X ^ P X ^ | μ n P X ^ n T n ( P ^ X ^ ) B n ( η ) δ .
Let P ˜ X ^ denote the type which maximizes the P X ^ n -probability of the type class among all such types. As there exist at most ( n + 1 ) | X ^ | possible types, it holds that
P X ^ n T n ( P ˜ X ^ ) B n ( η ) δ ( n + 1 ) | X ^ | .
Define the set Ψ n ( η ) T n ( P ˜ X ^ ) B n ( η ) . We can write the probability in (A70) as
(A71) P X ^ n T n ( P ˜ X ^ ) B n ( η ) = x ^ n Ψ n ( η ) P X ^ n ( x ^ n ) (A72) = x ^ n Ψ n ( η ) 2 n D ( P ˜ X ^ P X ^ ) + H P ˜ X ^ ( X ^ ) (A73) x ^ n Ψ n ( η ) 2 n [ H ( X ^ ) δ 1 ]
where δ 1 0 as n due to the fact that D ( P ˜ X ^ P X ^ ) 0 and | P ˜ X ^ P X ^ | μ n so the entropies are also arbitrarily close. It then follows from (A70) and (A73) that
1 n log | Ψ n ( η ) | H ( X ^ ) δ 2 ,
where δ 2 0 as μ n 0 .
The encoding function ϕ ( n ) partitions the set Ψ n ( η ) into τ n non-intersecting subsets { S i } i = 1 τ n such that ϕ ( n ) ( x ^ n ) = i for any x ^ n S i . Define the following distribution:
P X ^ ̲ n ( x ^ n ) P X ^ n ( x ^ n ) · 1 x ^ n Ψ n ( η ) P X ^ n ( Ψ n ( η ) ) .
Please note that this distribution, henceforth denoted as P ( n ) , corresponds to a uniform distribution over Ψ n ( η ) because all sequences in Ψ n ( η ) have the same type P ˜ X ^ , and the probability is uniform on a type class under any i.i.d. measure.
Finally, define the following truncated joint distribution:
P M ̲ X ̲ n X ^ ̲ n Y ̲ n ( m , x n , x ^ n , y n ) 1 ϕ ( n ) ( x n ) = m P X | X ^ n ( x n | x ^ n ) P X ^ ̲ n ( x ^ n ) P Y | X n ( y n | x n ) .
(2)
Analysis of Type-II Error Exponent:
The proof of the upper bound on the error exponent relies on the following Lemma A1. For a set A X ^ n , let B ( A , η ) denote the collection of all η -images of A, define Q X X ^ Y P X ^ | X Q X Y and
κ V n ( A , Q X ^ Y , η ) min B B ( A , η ) Q X ^ Y n ( A × B ) P X ^ n ( A ) .
This quantity is a generalization of the minimum cardinality of the η -images in [38] and is closely related to the minimum type-II error probability associated with the set A.
For the testing against independence setup, Q X ^ Y = P X ^ · P Y , and thus
Q X ^ Y n ( A × B ) P X ^ n ( A ) = P X ^ n ( A ) P Y n ( B ) P X ^ n ( A ) = P Y n ( B ) ,
and κ V n ( A , Q X ^ Y , η ) is simply written as κ V n ( A , η ) and is given by
κ V n ( A , η ) min B B ( A , η ) P Y n ( B ) .
Lemma A1
(Lemma 3 in [4]). For any set A X ^ n , consider a distribution P A ( n ) over A and let P A ( n ) V n be its corresponding output distribution induced by the channel V n , i.e.,
P A ( n ) V n ( y n ) x ^ n A P A ( n ) ( x ^ n ) V n y n | x ^ n .
Then, for every δ > 0 , 0 < η < 1 , we have
κ V n ( A , η ) 2 D ( P A ( n ) V n P Y n ) n δ
for sufficiently large n.
Let P i ( n ) V n be the distribution of the random variable Y ̲ n given M ̲ = i . The type-II error probability can be lower-bounded as:
(A82) β n x ^ n Ψ n ( η ) P X ^ n x ^ n · P Y n D ϕ ( n ) ( x ^ n ) (A83) = i = 1 τ n P X ^ n ( S i ) · P Y n ( D i ) (A84) i = 1 τ n P X ^ n ( S i ) · κ V n ( S i , η ) (A85) = P X ^ n Ψ n ( η ) · i = 1 τ n P ( n ) ( S i ) · κ V n ( S i , η ) (A86) 2 n δ · P X ^ n Ψ n ( η ) · i = 1 τ n P ( n ) ( S i ) · 2 D P i ( n ) V n P Y n (A87) 2 n δ · P X ^ n Ψ n ( η ) · 2 i = 1 τ n P ( n ) ( S i ) · D P i ( n ) V n P Y n (A88) 2 n δ δ ( n + 1 ) | X ^ | · 2 i = 1 τ n P ( n ) ( S i ) · D P i ( n ) V n P Y n ,
where (A84) follows from the definition of κ V n ( S i , η ) , (A86) follows because Lemma A1 implies that for any distribution P i ( n ) over the set S i it holds that κ V n ( S i , η ) 2 D P i ( n ) V n P Y n n δ , (A87) follows because of the convexity of the function t 2 t , and (A88) follows by (A70) and the fact that Pr ( A ) Pr ( A B ) . Hence,
1 n log β n δ 1 n i = 1 τ n P ( n ) ( S i ) · D P i ( n ) V n P Y n ,
where δ δ 1 n log δ ( n + 1 ) | X ^ | .
(3)
Single-letterization Steps and Analyses of Rate and Privacy Constraints:
In the following, we proceed to provide a single-letter characterization of the upper bound in (A89). Considering the fact that P ( n ) ( S i ) = P M ̲ ( i ) , the right-hand-side of (A89) can be upper-bounded as follows:
(A90) 1 n i = 1 τ n P ( n ) ( S i ) · D ( P i ( n ) V n P Y n ) = 1 n i = 1 τ n y n Y n P M ̲ Y ̲ n ( i , y n ) log P Y ̲ n | M ̲ ( y n | i ) P Y n ( y n ) (A91) = 1 n H ( Y ̲ n | M ̲ ) 1 n y n Y n P Y ̲ n ( y n ) log P Y n ( y n ) (A92) = 1 n H ( Y ̲ n | M ̲ ) 1 n y n Y n P Y ̲ n ( y n ) t = 1 n log P Y ( y t ) (A93) = 1 n H ( Y ̲ n | M ̲ ) 1 n t = 1 n y n Y n P Y ̲ n ( y n ) log P Y ( y t ) (A94) = 1 n H ( Y ̲ n | M ̲ ) 1 n t = 1 n y t Y P Y ̲ t ( y t ) log P Y ( y t ) (A95) = 1 n H ( Y ̲ n | M ̲ ) + 1 n t = 1 n H ( Y ̲ t ) + D ( P Y ̲ t P Y ) (A96) = 1 n t = 1 n H ( Y ̲ t ) H ( Y ̲ t | M ̲ , Y ̲ t 1 ) + D ( P Y ̲ t P Y ) (A97) 1 n t = 1 n I ( M ̲ , X ̲ t 1 , X ^ ̲ t 1 ; Y ̲ t ) + 1 n t = 1 n D ( P Y ̲ t P Y ) (A98) = 1 n t = 1 n I ( U ̲ t ; Y ̲ t ) + 1 n t = 1 n D ( P Y ̲ t P Y ) (A99) = I ( U ̲ ; Y ̲ ) + D ( P Y ̲ P Y ) .
Here, (A96)–(A99) are justified in the following:
  • (A96) follows by the chain rule;
  • (A97) follows from the Markov chain Y ̲ t 1 ( M ̲ , X ̲ t 1 , X ^ ̲ t 1 ) Y ̲ t ;
  • (A98) follows from the definition
    U ̲ t ( M ̲ , X ̲ t 1 , X ^ ̲ t 1 ) ;
  • (A99) follows by defining a time-sharing random variable T over { 1 , , n } and the following
    U ̲ = Δ ( U ̲ T , T ) , Y ̲ = Δ Y ̲ T .
This leads to the following upper-bound on the type-II error exponent:
1 n log β n I ( U ̲ ; Y ̲ ) + D ( P Y ̲ P Y ) + δ .
Next, the rate constraint satisfies the following:
(A103) n R H ( M ̲ ) (A104) I ( M ̲ ; X ^ ̲ n ) (A105) = H ( X ^ ̲ n ) H ( X ^ ̲ n | M ̲ ) (A106) = log | Ψ n ( η ) | H ( X ^ ̲ n | M ̲ ) (A107) n ( H ( X ^ ) δ 2 ) H ( X ^ ̲ n | M ̲ ) (A108) = n H ( X ^ ) t = 1 n H ( X ^ ̲ t | X ^ ̲ t 1 , M ̲ ) n δ 2 (A109) = n H ( X ^ ) t = 1 n H ( X ^ ̲ t | U ̲ t ) n δ 2 (A110) = n H ( X ^ ) n H ( X ^ ̲ | U ̲ ) n δ 2
where (A106) follows because the distribution P X ^ ̲ n is uniform over the set Ψ n ( η ) ; (A107) follows from (A74); (A109) follows from the definition in (A100); (A110) follows by defining X ^ ̲ X ^ ̲ T .
Finally, the privacy measure satisfies the following:
(A111) n L I ( X n ; X ^ n ) (A112) I ( X ̲ n ; X ^ ̲ n ) (A113) = H ( X ^ ̲ n ) H ( X ^ ̲ n | X ̲ n ) (A114) = log | Ψ n ( η ) | H ( X ^ ̲ n | X ̲ n ) (A115) = n H ( X ^ ) δ 2 H ( X ^ ̲ n | X ̲ n ) (A116) = n H ( X ^ ) δ 2 t = 1 n H ( X ^ ̲ t | X ^ ̲ t 1 , X ̲ n ) (A117) n H ( X ^ ) δ 2 t = 1 n H ( X ^ ̲ t | X ̲ t ) (A118) = n H ( X ^ ) n H ( X ^ ̲ | X ̲ ) n δ 2 ,
where (A112) follows because ( X ̲ n , X ^ ̲ n ) are functions of ( X n , X ^ n ) and from data processing inequality, (A114) follows because P X ^ ̲ n is uniform over the set Ψ n ( η ) (see definition in (A75)), (A115) follows from (A74), and (A118) follows by defining X ̲ ( X ̲ T , T ) , X ^ ̲ ( X ^ ̲ T , T ) and choosing T uniformly over { 1 , , n } .
Since Ψ n ( η ) T n ( P ˜ X ^ ) , for any x ^ X ^ ,
(A119) P X ^ ̲ ( x ^ ) = 1 n t = 1 n P X ^ ̲ t ( x ^ ) (A120) = x ^ n Ψ n ( η ) N x ^ | x ^ n n · | Ψ n ( η ) | (A121) = P ˜ X ^ ( x ^ ) .
Recall that | P ˜ X ^ P X ^ | μ n with μ n = n 1 / 3 . Hence, from (A121), it holds that | P X ^ ̲ P X ^ | μ n . By the definitions of X ^ ̲ , X ̲ and Y ̲ , we have P X ^ | X = P X ^ ̲ | X ̲ and P Y | X = P Y ̲ | X ̲ = V . The random variable U is chosen over the same alphabet as U ̲ and such that P U | X ^ = P U ̲ | X ^ ̲ .
Since P Y ( y ) > 0 for all y Y , letting n and μ n 0 and the uniform continuity of the involved information-theoretic quantities yields the following upper bound on the optimal error exponent:
θ ϵ * ( R , L ) I ( U ; Y ) ,
subject to the rate constraint:
R I ( U ; X ^ ) ,
and the privacy constraint:
L I ( X ; X ^ ) .
This concludes the proof of converse.

Appendix C. Proof of Converse of Proposition 1

We simplify Theorem 2 for the proposed binary setup. As discussed in Section 4.3, from the fact that | X ^ | = 2 and the symmetry of the source X on its alphabet, without loss of optimality, we can choose P X ^ | X to be a BSC. First, consider the rate constraint:
(A125) R I ( U ; X ^ ) (A126) = H ( X ^ ) H ( X ^ | U ) (A127) = 1 H ( X ^ | U ) ,
which can be equivalently written as the following:
H ( X ^ | U ) 1 R .
Also, the privacy criterion can be simplified as follows:
(A129) L I ( X ^ ; X ) (A130) = H ( X ^ ) H ( X ^ | X ) (A131) = 1 H ( X ^ | X ) (A132) = 1 H ( Z ^ ) ,
which can be equivalently written as
H ( Z ^ ) 1 L .
Now, consider the error exponent θ as follows:
(A134) θ I ( U ; Y ) (A135) = H ( Y ) H ( Y | U ) (A136) = H ( Y ) H ( X N | U ) (A137) = H ( Y ) H ( X ^ Z ^ N | U ) (A138) H ( Y ) h b h b 1 ( H ( X ^ | U ) ) h b 1 ( 1 L ) q (A139) H ( Y ) h b h b 1 ( 1 R ) h b 1 ( 1 L ) q ,
where (A138) follows from Mrs. Gerber’s lemma [39] (Theorem 1) and the fact that ( Z ^ , N ) is independent of U and also from (A133); (A139) follows from (A128). This concludes the proof of the proposition.

Appendix D. Euclidean Approximation of Testing agianst Independence

We analyze the Euclidean approximation with the parameters, W , ψ u ( x ^ ) , ϕ x ^ ( x ) and Λ u ( x ) defined in Section 4.4. Notice that since U X ^ X Y forms a Markov chain, it holds that, for any u U ,
P Y | U = u = W P X | U = u .
Now, consider the following chain of equalities for any x X :
(A141) P X | U ( x | u ) = x ^ X ^ P X X ^ | U ( x , x ^ | u ) (A142) = x ^ X ^ P X ^ | U ( x ^ | u ) P X | X ^ , U ( x | x ^ , u ) (A143) = x ^ X ^ P X ^ | U ( x ^ | u ) P X | X ^ ( x | x ^ ) (A144) = x ^ X ^ P X ^ ( x ^ ) + ψ u ( x ^ ) P X ( x ) + ϕ x ^ ( x ) (A145) = P X ( x ) + x ^ X ^ ψ u ( x ^ ) ϕ x ^ ( x ) + x ^ X ^ P X ^ ( x ^ ) ϕ x ^ ( x ) + P X ( x ) x ^ X ^ ψ u ( x ^ ) (A146) = P X ( x ) + x ^ X ^ ψ u ( x ^ ) ϕ x ^ ( x ) ,
where (A143)—(A146) are justified in the following:
  • (A143) follows from the Markov chain U X ^ X where given X ^ , U and X are independent;
  • (A144) follows from (30) and (36);
  • (A146) follows from (31) and also from (36) which yields the following:
    x ^ X ^ P X ^ ( x ^ ) · ϕ x ^ ( x ) = 0 .
With the definition of Λ u ( x ) in (42), we can write
P X | U ( x | u ) = P X ( x ) + Λ u ( x ) , x X , u U .
Thus, we get
(A149) P Y | U = u = W P X + W Λ u (A150) = P Y + W Λ u .
Applying the χ 2 -approximation and using (A150), we can rewrite I ( U ; Y ) as follows:
I ( U ; Y ) 1 2 log e u U P U ( u ) P Y 1 W Λ u 2
The above approximation with the definition of the vector Λ u in (43) yields the optimization problem in (45).

Appendix E. Proof of Proposition 2

Achievability: We specialize the achievable scheme of Theorem 2 to the proposed Gaussian setup. We choose the auxiliary random variables as in (63) and (65). Notice that from the Markov chain U X ^ X Y and also the Gaussian choice of X ^ in (63) which was discussed in Section 4.5, we can write Y = ρ X ^ + F where F N 0 , 1 ρ 2 · ( 1 2 2 L ) is independent of X ^ . These choices of auxiliary random variables lead to the following rate constraint:
R 1 2 log 1 2 2 L β 2 ,
which can be equivalently written as:
2 2 R · 1 2 2 L β 2 .
The optimal error exponent is also lower bounded as follows
θ ϵ * ( R , L ) 1 2 log 1 1 ρ 2 · 1 2 2 L β 2 .
Combining (A153) and (A154) gives the lower bound on the error exponent in (64).
Converse: Consider the following upper bound on the optimal error exponent in Theorem 2:
(A155) θ ϵ * ( R , L ) I ( U ; Y ) (A156) = h ( Y ) h ( Y | U ) (A157) = 1 2 log 2 π e h ( Y | U ) (A158) = 1 2 log 2 π e h ρ X ^ + F | U (A159) 1 2 log 2 π e 1 2 log 2 2 h ρ X ^ | U + 2 π e 1 ρ 2 · ( 1 2 2 L ) (A160) 1 2 log 2 π e 1 2 log ρ 2 2 2 h X ^ | U + 2 π e 1 ρ 2 · ( 1 2 2 L ) ,
where (A159) follows from the entropy power inequality (EPI) [34] (Chapter 2). Now, consider the rate constraint as follows:
(A161) R I ( X ^ ; U ) (A162) = h ( X ^ ) h ( X ^ | U ) (A163) = 1 2 log 2 π e 1 2 2 L h ( X ^ | U ) ,
which is equivalent to
2 2 h ( X ^ | U ) 2 π e · 2 2 R · 1 2 2 L .
Considering (A160) with (A164) yields the following upper bound on the error exponent:
(A165) θ ϵ * ( R , L ) 1 2 log 2 π e 1 2 log 2 π e ρ 2 2 2 R 1 2 2 L + 2 π e 1 ρ 2 ( 1 2 2 L ) (A166) = 1 2 log 1 1 ρ 2 ( 1 2 2 R ) ( 1 2 2 L ) .
This concludes the proof of the proposition.

References

  1. Ahlswede, R.; Csiszàr, I. Hypothesis testing with communication constraints. IEEE Trans. Inf. Theory 1986, 32, 533–542. [Google Scholar] [CrossRef]
  2. Zhao, W.; Lai, L. Distributed testing against independence with multiple terminals. In Proceedings of the 2014 52nd Annual Allerton Conference on Communication, Control, and Computing (Allerton), Monticello, IL, USA, 30 September–3 October 2014; pp. 1246–1251. [Google Scholar]
  3. Xiang, Y.; Kim, Y.H. Interactive hypothesis testing against independence. In Proceedings of the 2013 IEEE International Symposium on Information Theory, Istanbul, Turkey, 7–12 July 2013; pp. 2840–2844. [Google Scholar]
  4. Tian, C.; Chen, J. Successive refinement for hypothesis testing and lossless one-helper problem. IEEE Trans. Inf. Theory 2008, 54, 4666–4681. [Google Scholar] [CrossRef]
  5. Rahman, M.S.; Wagner, A.B. On the optimality of binning for distributed hypothesis testing. IEEE Trans. Inf. Theory 2012, 58, 6282–6303. [Google Scholar] [CrossRef]
  6. Sreekuma, S.; Gündüz, D. Distributed Hypothesis Testing Over Noisy Channels. In Proceedings of the 2017 IEEE International Symposium on Information Theory (ISIT), Aachen, Germany, 25–30 June 2017. [Google Scholar]
  7. Salehkalaibar, S.; Wigger, M.; Timo, R. On hypothesis testing against independence with multiple decision centers. arXiv 2017, arXiv:1708.03941. [Google Scholar] [CrossRef]
  8. Salehkalaibar, S.; Wigger, M.; Wang, L. Hypothesis Testing In Multi-Hop Networks. arXiv 2017, arXiv:1708.05198. [Google Scholar]
  9. Mhanna, M.; Piantanida, P. On secure distributed hypothesis testing. In Proceedings of the 2015 IEEE International Symposium on Information Theory (ISIT), Hong Kong, China, 14–19 June 2015; pp. 1605–1609. [Google Scholar]
  10. Han, T.S. Hypothesis testing with multiterminal data compression. IEEE Trans. Inf. Theory 1987, 33, 759–772. [Google Scholar] [CrossRef]
  11. Shimokawa, H.; Han, T.; Amari, S.I. Error bound for hypothesis testing with data compression. IEEE Trans. Inf. Theory 1994, 32, 533–542. [Google Scholar]
  12. Ugur, Y.; Aguerri, I.E.; Zaidi, A. Vector Gaussian CEO Problem Under Logarithmic Loss and Applications. arXiv 2018, arXiv:1811.03933. [Google Scholar]
  13. Zaidi, A.; Aguerri, I.E.; Caire, G.; Shamai, S. Uplink oblivious cloud radio access networks: An information theoretic overview. In Proceedings of the 2018 Information Theory and Applications Workshop (ITA), San Diego, CA, USA, 11–16 February 2018. [Google Scholar]
  14. Aguerri, I.E.; Zaidi, A.; Caire, G.; Shamai, S. On the capacity of cloud radio access networks with oblivious relaying. In Proceedings of the 2017 IEEE International Symposium on Information Theory (ISIT), Aachen, Germany, 25–30 June 2017. [Google Scholar]
  15. Aguerri, I.E.; Zaidi, A. Distributed information bottleneck method for discrete and Gaussian sources. In Proceedings of the 2018 International Zurich Seminar on Information and Communication (IZS), Zurich, Switzerland, 21–23 February 2018. [Google Scholar]
  16. Aguerri, I.E.; Zaidi, A. Distributed variational representation learning. arXiv 2018, arXiv:1807.04193. [Google Scholar]
  17. Evfimievski, A.V.; Gehrke, J.; Srikant, R. Limiting privacy breaches in privacy preserving data mining. In Proceedings of the Twenty-Second Symposium on Principles of Database Systems, San Diego, CA, USA, 9–11 June 2003; pp. 211–222. [Google Scholar]
  18. Smith, G. On the Foundations of Quantitative Information Flow. In Proceedings of the 12th International Conference on Foundations of Software Science and Computational Structures: Held as Part of the Joint European Conferences on Theory and Practice of Software, ETAPS 2009; Springer: Berlin/Heidelberg, Germany, 2009; pp. 288–302. [Google Scholar]
  19. Sankar, L.; Rajagopalan, S.R.; Poor, H.V. Utility-Privacy Tradeoffs in Databases: An Information-Theoretic Approach. IEEE Trans. Inf. Forensics Secur. 2013, 8, 838–852. [Google Scholar] [CrossRef]
  20. Liao, J.; Sankar, L.; Tan, V.Y.F.; Calmon, F. Hypothesis Testing Under Mutual Information Privacy Constraints in the High Privacy Regime. IEEE Trans. Inf. Forensics Secur. 2018, 13, 1058–1071. [Google Scholar] [CrossRef]
  21. Barthe, G.; Köpf, B. Information-theoretic bounds for differentially private mechanisms. In Proceedings of the 2011 IEEE 24th Computer Security Foundations Symposium, Cernay-la-Ville, France, 27–29 June 2011; pp. 191–204. [Google Scholar]
  22. Issa, I.; Wagner, A.B. Operational definitions for some common information leakage metrics. In Proceedings of the 2017 IEEE International Symposium on Information Theory (ISIT), Aachen, Germany, 25–30 June 2017; pp. 769–773. [Google Scholar]
  23. Liao, J.; Sankar, L.; Calmon, F.; Tan, V.Y.F. Hypothesis testing under maximal leakage privacy constraints. In Proceedings of the 2017 IEEE International Symposium on Information Theory (ISIT), Aachen, Germany, 25–30 June 2017; pp. 779–783. [Google Scholar]
  24. Wagner, I.; Eckhoff, D. Technical Privacy Metrics: A Systematic Survey. ACM Comput. Surv. (CSUR) 2018, 51. to appear. [Google Scholar] [CrossRef]
  25. Dwork, C. Differential Privacy. In Proceedings of the 33rd International Colloquium on Automata, Languages and Programming, Part II (ICALP 2006); Springer: Venice, Italy, 2006; Volume 4052, pp. 1–12. [Google Scholar]
  26. Dwork, C.; Kenthapadi, K.; McSherry, F.; Mironov, I.; Naor, M. Our Data, Ourselves: Privacy Via Distributed Noise Generation. In Advances in Cryptology (EUROCRYPT 2006); Springer: Saint Petersburg, Russia, 2006; Volume 4004, pp. 486–503. [Google Scholar]
  27. Dwork, C. Differential Privacy: A Survey of Results; Chapter Theory and Applications of Models of Computation; TAMC 2008; Lecture Notes in Computer Science; Springer: Heidelberg, Germany, 2008; Volume 4978. [Google Scholar]
  28. Wasserman, L.; Zhou, S. A statistical framework for differential privacy. J. Am. Stat. Assoc. 2010, 105, 375–389. [Google Scholar] [CrossRef]
  29. Sreekumar, A.C.; Gunduz, D. Distributed hypothesis testing with a privacy constraint. arXiv 2018, arXiv:1806.02015. [Google Scholar]
  30. Borade, S.; Zheng, L. Euclidean Information Theory. In Proceedings of the 2008 IEEE International Zurich Seminar on Communications, Zurich, Switzerland, 12–14 March 2008; pp. 14–17. [Google Scholar]
  31. Huang, S.; Suh, C.; Zheng, L. Euclidean information theory of networks. IEEE Trans. Inf. Theory 2015, 61, 6795–6814. [Google Scholar] [CrossRef]
  32. Viterbi, A.J.; Omura, J.K. Principles of Digital Communication and Coding; McGraw-Hill: New York, NY, USA, 1979. [Google Scholar]
  33. Weinberger, N.; Merhav, N. Optimum tradeoffs between the error exponent and the excess-rate exponent of variable rate Slepian-Wolf coding. IEEE Trans. Inf. Theory 2015, 61, 2165–2190. [Google Scholar] [CrossRef]
  34. El Gamal, A.; Kim, Y.H. Network Information Theory; Cambridge University Press: Cambridge, UK, 2011. [Google Scholar]
  35. Shalaby, H.M.H.; Papamarcou, A. Multiterminal detection with zero-rate data compression. IEEE Trans. Inf. Theory 1992, 38, 254–267. [Google Scholar] [CrossRef]
  36. Cover, T.M.; Thomas, J.A. Elements of Information Theory, 2nd ed.; Wiley: Hoboken, NJ, USA, 2006. [Google Scholar]
  37. Watanabe, S. Neyman-Pearson Test for Zero-Rate Multiterminal Hypothesis Testing. IEEE Trans. Inf. Theory 2018, 64, 4923–4939. [Google Scholar] [CrossRef]
  38. Csiszàr, I.; Körner, J. Information theory: Coding Theorems for Discrete Memoryless Systems; Academic Press: New York, NY, USA, 1982. [Google Scholar]
  39. Wyner, A.D.; Ziv, J. A theorem on the entropy of certain binary sequences and applications (Part I). IEEE Trans. Inf. Theory 1973, 19, 769–772. [Google Scholar] [CrossRef]
Figure 1. Hypothesis testing with communication and privacy constraints.
Figure 1. Hypothesis testing with communication and privacy constraints.
Entropy 21 00478 g001
Figure 2. θ ϵ * ( R , L ) versus L for q = 0 . 1 and various values of R.
Figure 2. θ ϵ * ( R , L ) versus L for q = 0 . 1 and various values of R.
Entropy 21 00478 g002
Figure 3. θ ϵ * ( R 0 , L 0 ) versus L for q = 0.1 and R = L .
Figure 3. θ ϵ * ( R 0 , L 0 ) versus L for q = 0.1 and R = L .
Entropy 21 00478 g003

Share and Cite

MDPI and ACS Style

Gilani, A.; Belhadj Amor, S.; Salehkalaibar, S.; Tan, V.Y.F. Distributed Hypothesis Testing with Privacy Constraints. Entropy 2019, 21, 478. https://doi.org/10.3390/e21050478

AMA Style

Gilani A, Belhadj Amor S, Salehkalaibar S, Tan VYF. Distributed Hypothesis Testing with Privacy Constraints. Entropy. 2019; 21(5):478. https://doi.org/10.3390/e21050478

Chicago/Turabian Style

Gilani, Atefeh, Selma Belhadj Amor, Sadaf Salehkalaibar, and Vincent Y. F. Tan. 2019. "Distributed Hypothesis Testing with Privacy Constraints" Entropy 21, no. 5: 478. https://doi.org/10.3390/e21050478

APA Style

Gilani, A., Belhadj Amor, S., Salehkalaibar, S., & Tan, V. Y. F. (2019). Distributed Hypothesis Testing with Privacy Constraints. Entropy, 21(5), 478. https://doi.org/10.3390/e21050478

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop