Next Article in Journal
Identify the Rotating Stall in Centrifugal Compressors by Fractal Dimension in Reconstructed Phase Space
Previous Article in Journal
Distributed Vector Quantization Based on Kullback-Leibler Divergence
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Feeding Back the Output or Sharing the State: Which Is Better for the State-Dependent Wiretap Channel?

1
School of Information Science and Technology, Southwest JiaoTong University, Northbound Section Second Ring Road 111, Chengdu 610031, China
2
The National Mobile Communications Research Laboratory, Southeast University, Nanjing 210096, China
3
School of Economics and Management, Chengdu Textile College, Chengdu 611731, China
*
Author to whom correspondence should be addressed.
Entropy 2015, 17(12), 7900-7925; https://doi.org/10.3390/e17127852
Submission received: 29 July 2015 / Revised: 2 October 2015 / Accepted: 23 November 2015 / Published: 30 November 2015
(This article belongs to the Section Information Theory, Probability and Statistics)

Abstract

:
In this paper, the general wiretap channel with channel state information (CSI) at the transmitter and noiseless feedback is investigated, where the feedback is from the legitimate receiver to the transmitter, and the CSI is available at the transmitter in a causal or noncausal manner. The capacity-equivocation regions are determined for this model in both causal and noncausal cases, and the results are further explained via Gaussian and binary examples. For the Gaussian model, we find that in some particular cases, the noiseless feedback performs better than Chia and El Gamal’s CSI sharing scheme, i.e., the secrecy capacity of this feedback scheme is larger than that of the CSI sharing scheme. For the degraded binary model, we find that the noiseless feedback performs no better than Chia and El Gamal’s CSI sharing scheme. However, if the cross-over probability of the wiretap channel is large enough, we show that the two schemes perform the same.

1. Introduction

It is known to all that the capacity of a point-to-point discrete memoryless channel (DMC) cannot be increased by using noiseless feedback. However, does the feedback (from the legitimate receiver to the transmitter) enhance the security of the wiretap channel? Ahlswede and Cai [1] and Dai et al. [2] studied this problem. Specifically, Ahlswede and Cai [1] showed that the secrecy capacity C s f of the degraded wiretap channel with noiseless feedback is given by:
C s f = max p ( x ) min { I ( X ; Y ) , I ( X ; Y ) - I ( X ; Z ) + H ( Y | X , Z ) } ,
where X, Y and Z are for the transmitter, legitimate receiver and wiretapper, respectively, and X Y Z forms a Markov chain. Recall that the secrecy capacity C s of the degraded wiretap channel is determined by Wyner [3], and it is given by:
C s = max p ( x ) min { I ( X ; Y ) , I ( X ; Y ) - I ( X ; Z ) } .
From (1) and (2), it is easy to see that the noiseless feedback increases the secrecy capacity of the wiretap channel. Based on the work of [1], Dai et al. [2] studied a special wiretap channel with feedback ( Y X Z ) and showed that the secrecy capacity of this model is larger than that of the model without feedback, i.e., the noiseless feedback helps to enhance the security of the special wiretap channel Y X Z . Here, note that in [1] and [2], the legitimate receiver just sends back the previous received symbols to the transmitter, and it is natural to ask: is it better for the legitimate receiver to send back purely random secret keys to the transmitter? Ardestanizadeh et al. [4] answered this question by considering the model of the wiretap channel with secure rate-limited feedback. Ardestanizadeh et al. [4] showed that if the limits (capacity) of the feedback channel are denoted by R f , the secrecy capacity of the physically-degraded wiretap channel ( X Y Z ) with secure rate-limited feedback is given by:
C s f = max p ( x ) min { I ( X ; Y ) , I ( X ; Y ) - I ( X ; Z ) + R f } .
Compared to (1), it is easy to see that if R f H ( Y | X , Z ) , sending purely random secret keys is no better than sending Y i - 1 back. If R f > H ( Y | X , Z ) , I ( X ; Y ) - I ( X ; Z ) + R f > H ( Y | Z ) , sending purely random secret keys is better than sending Y i - 1 back. Besides these works on the wiretap channel with feedback, Lai et al. [5] studied the wiretap channel with noisy feedback; He et al. [6] studied the Gaussian two-way wiretap channel and the Gaussian half-duplex two-way relay channel with an un-trusted relay; and Bassi et al. [7] studied the wiretap channel with generalized feedback. Bounds on the secrecy capacities of these feedback models are obtained in [5,6,7].
Recently, the wiretap channel with channel state information (CSI) has received much attention. The Gaussian wiretap channel with noncausal CSI at the transmitter was studied in [8,9], and an achievable rate-equivocation region was provided for this Gaussian model. Based on the work of [8], Chen et al. [10] studied the discrete memoryless wiretap channel with noncausal CSI at the transmitter and also provided an achievable rate-equivocation region for this model. The encoding-decoding scheme of [10] is a combination of the binning technique of Gel’fand and Pinsker’s channel [11] and the random binning technique of Wyner’s wiretap channel [3]. After that, Dai et al. [12] studied the outer bound on the capacity-equivocation region of [10] and also investigated the capacity results of the discrete memoryless wiretap channel with causal or memoryless CSI at the transmitter. Besides these works on the wiretap channel with CSI only available at the transmitter, Chia and El Gamal [13] investigated the wiretap channel with CSI causally or non-causally at both the transmitter and the legitimate receiver and provided an achievable secrecy rate, which was larger than that of [10]. In [13], since both the transmitter and the legitimate receiver have access to the CSI, the CSI serves as a secret key shared by them. Therefore, the encoding-decoding scheme of [13] is similar to that of the wiretap channel with rate-limited feedback [4]. Besides these works on the wiretap channel with CSI, Liu et al. [14] studied the block Rayleigh fading MIMO wiretap channel with no CSI available at the legitimate receiver, the wiretapper and the transmitter, and they showed that if the legitimate receiver had more antennas than the wiretapper, non-zero secure degrees of freedom (s.d.o.f) could also be achieved.
In this paper, we study the general wiretap channel with CSI (causally or non-causally at the transmitter) and noiseless feedback; see Figure 1. In Figure 1, the transition probability of the channel depends on a CSI sequence V N , which is available at the channel encoder in a noncausal or causal manner. The inputs of the channel are X N and V N , while the outputs of the channel are Y N and Z N . Moreover, there exists a noiseless feedback from Y N to the channel encoder. The motivation of this work is to find whether the noiseless feedback helps to enhance the secrecy rate of the wiretap channel with noncausal or causal CSI at the transmitter [10,12] and whether the noiseless feedback does better than the shared CSI between the legitimate receiver and the transmitter [13] in enhancing the secrecy rate of the state-dependent wiretap channel.
Figure 1. General wiretap channel with noncausal or causal channel state information (CSI) and noiseless feedback.
Figure 1. General wiretap channel with noncausal or causal channel state information (CSI) and noiseless feedback.
Entropy 17 07852 g001
The capacity-equivocation region of the model of Figure 1 is determined for both the noncausal and causal cases, and the results are further explained via degraded binary and Gaussian examples. For the Gaussian example, we find that both the feedback scheme and the CSI sharing scheme [13] help to enhance the security of the wiretap channel with noncausal CSI at the transmitter [10,12], and moreover, we find that in some particular cases, the noiseless feedback performs even better than the shared CSI [13], i.e., the secrecy capacity of the degraded Gaussian case of the model of Figure 1 is larger than that of the degraded Gaussian case of [13]. For the binary example, we also find that both the feedback scheme and the CSI sharing scheme [13] help to enhance the security of the wiretap channel with causal CSI at the transmitter. Unlike the Gaussian case, we find that the noiseless feedback performs no better than the shared CSI [13], i.e., the secrecy capacity of the degraded binary case of the model of Figure 1 is not more than that of the degraded binary case of [13]. However, if the cross-over probability of the wiretap channel is large enough, we find that the two schemes perform the same.
The remainder of this paper is organized as follows. The capacity-equivocation region of the model of Figure 1 is provided in Section 2. Gaussian and binary examples of the model of Figure 1 are shown in Section 3. Section 4 is for the final conclusion.

2. Capacity-Equivocation Region of the Model of Figure 1

In this paper, random variables, sample values and alphabets are denoted by capital letters, lower case letters and calligraphic letters, respectively. A similar convention is applied to the random vectors and their sample values. For example, U N denotes a random N-vector ( U 1 , . . . , U N ) , and u N = ( u 1 , . . . , u N ) is a specific vector value in U N that is the N-th Cartesian power of U . U i N denotes a random N - i + 1 -vector ( U i , . . . , U N ) , and u i N = ( u i , . . . , u N ) is a specific vector value in U i N . Let P V ( v ) denote the probability mass function P r { V = v } . Throughout the paper, the logarithmic function is to the base two.

2.1. Definitions of the Model of Figure 1

Let W, uniformly distributed over the alphabet W , be the message sent by the transmitter. The components of the channel state sequence V N are independent and identically distributed. The probability of each component is P V ( v ) . V N is independent of W. Let Y i - 1 ( 2 i N ) be the i-th time feedback from the legitimate receiver to the transmitter. For the noncausal case, the i-th time channel encoder f i is a (stochastic) mapping:
f i : W × Y i - 1 × V N X i ,
where f i ( w , y i - 1 , v N ) = x i X , w W , y i - 1 Y i - 1 and v N V N . For the causal case, the i-th time channel encoder f i is a (stochastic) mapping:
f i : W × Y i - 1 × V i X i ,
where f i ( w , y i - 1 , v i ) = x i X , w W , y i - 1 Y i - 1 and v i V i . Here, note that for the causal case, V i is independent of ( Y i - 1 , W , V i + 1 N , Z i - 1 ) .
The channel is discrete memoryless, and its transition probability is given by:
P Z N , Y N | X N , V N ( z N , y N | x N , v N ) = i = 1 N P Z , Y | X , V ( z i , y i | x i , v i ) ,
where x i X , v i V , y i Y and z i Z .
The wiretapper’s equivocation about the message W is denoted by:
Δ = 1 N H ( W | Z N ) .
The decoder f D is a function that maps a received sequence of N channel outputs to the messages set:
f D : Y N W .
We denote the probability of error P e by P r { W W ^ } .
Given a pair ( R , R e ) ( R , R e > 0 ), it is said to be achievable if, for arbitrary small positive ϵ, there exists an encoding-decoding scheme, such that:
lim N log W N = R , lim N Δ R e , P e ϵ .
The set R ( n f ) , which is composed of all achievable ( R , R e ) pairs, is called the capacity-equivocation region of the model of Figure 1 with noncausal CSI at the transmitter. An achievable rare C s ( n f ) , which is denoted by:
C s ( n f ) = max ( R , R e = R ) R ( n f ) R ,
is called the secrecy capacity of the model of Figure 1 with noncausal CSI at the transmitter.
Analogously, let R ( c f ) be the capacity-equivocation region of the model of Figure 1 with causal CSI at the transmitter and C s ( c f ) , which is denoted by:
C s ( c f ) = max ( R , R e = R ) R ( c f ) R ,
be the secrecy capacity of the model of Figure 1 with causal CSI at the transmitter.

2.2. Main Result of the Model of Figure 1

The following Theorem 1 characterizes the capacity-equivocation region R ( n f ) of the model of Figure 1 with noncausal CSI at the transmitter; see the following.
Theorem 1. A single-letter characterization of the region R ( n f ) is as follows,
R ( n f ) = { ( R , R e ) : 0 R e R , 0 R I ( K ; Y ) - I ( K ; V ) , R e H ( Y | Z ) } ,
for some distribution:
P K V X Y Z ( k , v , x , y , z ) = P Z Y | X V ( z , y | x , v ) P X | K V ( x | k , v ) P K V ( k , v ) ,
which implies the Markov chain K ( X , V ) ( Y , Z ) .
Proof. See Section A and Section B.  ☐
Remark 1.
  • The range of the random variable K satisfies K X V + 1 . The proof is standard and easily obtained by using the support lemma (see [15]), and thus, we omit the proof here.
  • Corollary 1. The secrecy capacity C s ( n f ) satisfies:
    C s ( n f ) = max P X | K V P K V min { I ( K ; Y ) - I ( K ; V ) , H ( Y | Z ) } .
    Proof. Substituting R e = R into the region R ( n f ) in Theorem 1, we have:
    R I ( K ; Y ) - I ( K ; V ) ,
    R H ( Y | Z ) ,
    By using (10), (13) and (14), Formula (12) is achieved; thus, the proof is completed. ☐
  • Here, note that if Z N is a degraded version of Y N (which implies the existence of the Markov chain K ( X , V ) Y Z ), the capacity-equivocation region R ( n f ) still holds. The proof of this degraded case is along the lines of the proof of Theorem 1, and thus, we omit the proof here. In [10,12], an achievable rate-equivocation region R i n is provided for the wiretap channel with noncausal CSI, and it is given by:
    R i n = { ( R , R e ) : R e R , R I ( K ; Y ) - I ( K ; V ) , R e I ( K ; Y ) - I ( K ; Z ) } ,
    where the joint probability distribution P K V X Y Z ( k , v , x , y , z ) of R i n satisfies:
    P K V X Y Z ( k , v , x , y , z ) = P Z | Y ( z | y ) P Y | X V ( y | x , v ) P X | K V ( x | k , v ) P K V ( k , v ) .
    Here, note that:
    I ( K ; Y ) - I ( K ; Z ) = H ( K | Z ) - H ( K | Y ) = ( a ) H ( K | Z ) - H ( K | Y , Z ) = I ( K ; Y | Z ) H ( Y | Z ) ,
    where (a) is from K Y Z . Therefore, it is easy to see that the achievable rate-equivocation region R i n of [10] and [12] is enhanced by using this noiseless feedback.
The following Theorem 2 characterizes the capacity-equivocation region R ( c f ) of the model of Figure 1 with causal CSI at the transmitter; see the following.
Theorem 2. A single-letter characterization of the region R ( c f ) is as follows,
R ( c f ) = { ( R , R e ) : 0 R e R , 0 R I ( K ; Y ) , R e H ( Y | Z ) } ,
for some distribution:
P K V X Y Z ( k , v , x , y , z ) = P Y Z | X V ( y , z | x , v ) P X | K V ( x | k , v ) P K ( k ) P V ( v ) .
which implies the Markov chain K ( X , V ) ( Y , Z ) and the fact that V is independent of K.
Proof.
  • Proof of the converse: Using the fact that V i is independent of Y i - 1 and Z i - 1 , the converse proof of Theorem 2 is along the lines of that of Theorem 1 (see Section A), and thus, we omit the proof here.
  • Proof of the achievability: The achievability proof of Theorem 2 is along the lines of the achievability proof of Theorem 1 (see Section B), and the only difference is that for the causal case, there is no need to use the binning technique. Thus, we omit the proof here.
The proof of Theorem 2 is completed.  ☐
Remark 2.
  • The range of the auxiliary random variable K satisfies K X V . The proof is standard and easily obtained by using the support lemma (see p. 310 of [16]), and thus, we omit the proof here.
  • Corollary 2. The secrecy capacity C s ( c f ) satisfies:
    C s ( c f ) = max P X | K V P K min { I ( K ; Y ) , H ( Y | Z ) } .
    Proof. Proof of (16): Substituting R e = R into the region R ( c f ) , we have:
    R I ( K ; Y ) ,
    R H ( Y | Z ) , .
    By using (11), (17) and (18), Formula (16) is achieved; thus, the proof is completed. ☐
  • Here, note that if Z N is a degraded version of Y N , the capacity-equivocation region R ( c f ) still holds. The proof of this degraded case is along the lines of the proof of Theorem 2, and thus, we omit the proof here. In [12], an achievable rate-equivocation region R i c is provided for the wiretap channel with causal CSI, and it is given by:
    R i c = { ( R , R e ) : R e R ,
    R I ( K ; Y ) , R e I ( K ; Y ) - I ( K ; Z ) } ,
    where the joint probability distribution P K V X Y Z ( k , v , x , y , z ) of R i c satisfies:
    P K V X Y Z ( k , v , x , y , z ) = P Z | Y ( z | y ) P Y | X V ( y | x , v ) P X | K V ( x | k , v ) P K ( k ) P V ( v ) .
    By using (15), it is easy to see that the achievable rate-equivocation region R i c is enhanced by using this noiseless feedback.

3. Examples of the Model of Figure 1

3.1. Gaussian Case of the Model of Figure 1 with Noncausal CSI at the Transmitter

For the Gaussian case of the model of Figure 1 with noncausal CSI at the transmitter, the i-th time ( 1 i N ) channel inputs and outputs are given by:
Y i = X i + V i + Z 1 , i , Z i = X i + V i + Z 2 , i ,
where V i N ( 0 , Q ) , Z 1 , i N ( 0 , N 1 ) and Z 2 , i N ( 0 , N 2 ) . Here, note that V i , Z 1 , i and Z 2 , i are independent random variables, X i is independent of Z 1 , i and Z 2 , i and 1 N i = 1 N E ( X i 2 ) P . The noise V i is non-causally known by the transmitter. The following Theorem 3 shows the secrecy capacity of the Gaussian case of the model of Figure 1 with noncausal CSI at the transmitter.
Theorem 3. For the Gaussian case of the model of Figure 1 with noncausal CSI at the transmitter, the secrecy capacity C s g f is characterized in the following two cases.
Case 1: If N 1 N 2 , the secrecy capacity C s g f is given by:
C s g f = max α min 1 2 ln ( P + Q + N 1 ) ( P + α 2 Q ) P Q ( 1 - α ) 2 + N ( P + α 2 Q ) - 1 2 ln P + α 2 Q P , 1 2 ln 2 π e ( P + Q + N 1 ) ( N 2 - N 1 ) P + Q + N 2 = min { 1 2 ln ( 1 + P N 1 ) , 1 2 ln 2 π e ( P + Q + N 1 ) ( N 2 - N 1 ) P + Q + N 2 } ,
where the maximum is achieved when α = P P + N 1 .
Case 2: If N 1 > N 2 , the secrecy capacity C s g f is given by:
C s g f = min { 1 2 ln ( 1 + P N 1 ) , 1 2 ln 2 π e ( N 1 - N 2 ) } .
Remark 3.
If N 1 N 2 , the relationship of the channel inputs and outputs defined in (20) can be equivalently characterized by:
Y i = X i + V i + Z 1 , i , Z i = X i + V i + Z 1 , i + Z 2 , i * ,
where Z 2 , i * N ( 0 , N 2 - N 1 ) , and it is independent of Z 1 , i . Similar to the determination of the capacity region of the Gaussian broadcast channel (pp. 117–118 of [17]), the relationship (23) implies that there exists a Markov chain ( X i , V i ) Y i Z i , i.e., the Gaussian case of the model of Figure 1 reduces to a degraded model of Figure 1.
Analogously, if N 1 > N 2 , the relationship of the channel inputs and outputs defined in (20) can be equivalently characterized by:
Y i = X i + V i + Z 1 , i * + Z 2 , i , Z i = X i + V i + Z 2 , i ,
where Z 1 , i * N ( 0 , N 1 - N 2 ) , and it is independent of Z 2 , i , X i and V i . The relationship (24) implies that there exists a Markov chain ( X i , V i ) Z i Y i in the Gaussian case of the model of Figure 1.
Proof. For the direct part of Theorem 3, like [18] and [10], the achievability of C s g f is proven by substituting K = X + α V , X N ( 0 , P ) , V N ( 0 , Q ) and the fact that X is independent of V in Theorem 1; the details of the proof are omitted in this paper. Here, note that the calculation of I ( K ; Y ) - I ( K ; V ) is exactly the same as that of the dirty paper channel (page 440 of [18]), and it is easy to see that the maximum of I ( K ; Y ) - I ( K ; V ) is achieved when α = P P + N 1 .
For the converse part of Theorem 3, note that the transmitter-receiver channel is Costa’s dirty paper channel [18]; thus, the secrecy capacity is upper bounded by the capacity of the dirty paper channel, i.e., C s g f 1 2 ln ( 1 + P N 1 ) . Now, it remains to show C s g f 1 2 ln 2 π e ( P + Q + N 1 ) ( N 2 - N 1 ) P + Q + N 2 for N 1 N 2 and C s g f 1 2 ln 2 π e ( N 1 - N 2 ) for N 1 > N 2 ; see the following.
Proof of C s g f 1 2 ln 2 π e ( P + Q + N 1 ) ( N 2 - N 1 ) P + Q + N 2 for N 1 N 2 :
First, note that:
1 N H ( W | Z N ) ( a ) 1 N ( I ( W ; Y N | Z N ) + δ ( P e ) ) 1 N i = 1 N h ( Y i | Z i ) + δ ( P e ) N ,
where (a) is from Fano’s inequality. The conditional differential entropy h ( Y i | Z i ) in (25) is bounded by:
h ( Y i | Z i ) = h ( Y i , Z i ) - h ( Z i ) = h ( Z i | Y i ) + h ( Y i ) - h ( Z i ) = ( 1 ) h ( X i + V i + Z 1 , i + Z 2 , i * | X i + V i + Z 1 , i ) + h ( X i + V i + Z 1 , i ) - h ( X i + V i + Z 1 , i + Z 2 , i * ) = ( 2 ) h ( Z 2 , i * ) + h ( X i + V i + Z 1 , i ) - h ( X i + V i + Z 1 , i + Z 2 , i * ) ( 3 ) h ( Z 2 , i * ) + h ( X i + V i + Z 1 , i ) - 1 2 ln ( e 2 h ( X i + V i + Z 1 , i ) + e 2 h ( Z 2 , i * ) ) = h ( Z 2 , i * ) + 1 2 ln ( e 2 h ( X i + V i + Z 1 , i ) ) - 1 2 ln ( e 2 h ( X i + V i + Z 1 , i ) + e 2 h ( Z 2 , i * ) ) = ( 4 ) 1 2 ln ( 2 π e ( N 2 - N 1 ) ) + 1 2 ln e 2 h ( X i + V i + Z 1 , i ) e 2 h ( X i + V i + Z 1 , i ) + 2 π e ( N 2 - N 1 ) ( 5 ) 1 2 ln ( 2 π e ( N 2 - N 1 ) ) + 1 2 ln 2 π e ( P + Q + N 1 ) 2 π e ( P + Q + N 1 ) + 2 π e ( N 2 - N 1 ) = 1 2 ln 2 π e ( N 2 - N 1 ) ( P + Q + N 1 ) P + Q + N 2 ,
where (1) is from Definition (23), (2) is from the fact that Z 2 , i * is independent of X i , V i and Z 1 , i , (3) is from the entropy power inequality e 2 h ( X i + V i + Z 1 , i + Z 2 , i * ) e 2 h ( X i + V i + Z 1 , i ) + e 2 h ( Z 2 , i * ) (see [19]), (4) is from the fact that the differential entropy of a Gaussian distributed random variable X is h ( X ) = 1 2 ln ( 2 π e D ( X ) ) (here, D ( X ) is the variance of the Gaussian random variable X) and (5) is from 1 2 ln e 2 h ( X i + V i + Z 1 , i ) e 2 h ( X i + V i + Z 1 , i ) + 2 π e ( N 2 - N 1 ) increasing while h ( X i + V i + Z 1 , i ) is increasing and the fact that h ( X i + V i + Z 1 , i ) 1 2 ln ( 2 π e ( P + Q + N 1 ) ) (here, note that “=” is achieved if X i N ( 0 , P ) ).
Substituting (26) into (25), we have:
1 N H ( W | Z N ) 1 N i = 1 N 1 2 ln 2 π e ( N 2 - N 1 ) ( P + Q + N 1 ) P + Q + N 2 + δ ( P e ) N = 1 2 ln 2 π e ( N 2 - N 1 ) ( P + Q + N 1 ) P + Q + N 2 + δ ( P e ) N .
Substituting P e ϵ into (27) and letting N , it is easy to see that C s g f 1 2 ln 2 π e ( P + Q + N 1 ) ( N 2 - N 1 ) P + Q + N 2 for N 1 N 2 .
Proof of C s g f 1 2 ln 2 π e ( N 1 - N 2 ) for N 1 > N 2 :
For the case N 1 > N 2 , the conditional differential entropy h ( Y i | Z i ) in (25) can be bounded by:
h ( Y i | Z i ) = ( a ) h ( X i + V i + Z 1 , i * + Z 2 , i | X i + V i + Z 2 , i ) = ( b ) h ( Z 1 , i * ) = ( c ) 1 2 ln 2 π e ( N 1 - N 2 ) ,
where (a) is from (24), (b) is from the fact that Z 1 , i * is independent of Z 2 , i , X i and V i and (c) is from the fact that the differential entropy of a Gaussian distributed random variable X is h ( X ) = 1 2 ln ( 2 π e D ( X ) ) (here, D ( X ) is the variance of the Gaussian random variable X). Substituting (28) and P e ϵ into (25) and letting N , it is easy to see that C s g f 1 2 ln 2 π e ( N 1 - N 2 ) for N 1 > N 2 . Thus, the converse part of Theorem 3 is proven. The proof of Theorem 3 is completed.  ☐
In [13] (p.2841, Theorem 3), Chia and El Gamal showed that if Y is less noisy than Z ( I ( X ; Y | V ) I ( X ; Z | V ) for every P X | V ( x | v ) ), the secrecy capacity of the wiretap channel with CSI non-causally known by both the transmitter and the legitimate receiver is given by:
C s - b o t h = max p ( x | v ) min { I ( X ; Y | V ) , I ( X ; Y | V ) - I ( X ; Z | V ) + H ( V | Z ) } .
Here, the I ( X ; Z | V ) - H ( V | Z ) in the above C s - b o t h can be rewritten as follows.
I ( X ; Z | V ) - H ( V | Z ) = H ( Z | V ) - H ( Z | X , V ) - H ( V | Z ) = H ( V , Z ) - H ( V ) - H ( Z | X , V ) - H ( V , Z ) + H ( Z ) = H ( Z ) - H ( V ) - H ( Z | X , V ) .
Substituting (29) into C s - b o t h , we have:
C s - b o t h = max p ( x | v ) min { I ( X ; Y | V ) , I ( X ; Y | V ) - H ( Z ) + H ( V ) + H ( Z | X , V ) } .
On the other hand, for Z less noisy than Y ( I ( X ; Z | V ) I ( X ; Y | V ) for every P X | V ( x | v ) ), Chia and El Gamal provided an achievable secrecy rate (lower bound on the secrecy capacity) for the wiretap channel with CSI non-causally known by both the transmitter and the legitimate receiver, and it is given by:
C s - b o t h i = max p ( x | v ) min { I ( X ; Y | V ) , H ( V | Z , X ) } .
The following Theorem 4 shows the results on the secrecy capacity of the Gaussian case of the wiretap channel with CSI non-causally known by both the transmitter and the legitimate receiver.
Theorem 4. For the Gaussian wiretap channel with part of the Gaussian noise non-causally known by both the transmitter and the legitimate receiver, the secrecy capacity C s - b o t h g is characterized by the following two cases.
Case 1: If N 1 N 2 , the secrecy capacity C s - b o t h g is given by:
C s - b o t h g = min 1 2 ln ( 1 + P N 1 ) , 1 2 ln ( 1 + P N 1 ) + 1 2 ln ( 2 π e Q ) - 1 2 ln ( P + Q + N 2 N 2 ) .
Case 2: If N 1 > N 2 , a lower bound C s - b o t h g i on the secrecy capacity C s - b o t h g is given by:
C s - b o t h g C s - b o t h g i = min { 1 2 ln 2 π e Q N 2 Q + N 2 , 1 2 ln ( 1 + P N 1 ) } .
Remark 4.
For the Gaussian case, the conditional mutual information I ( X ; Y | V ) is calculated by using the fact that when the CSI is known by both the legitimate receiver and the transmitter, it can be simply subtracted off, which in effect reduces the channel to a Gaussian channel with no CSI, i.e., I ( X ; Y | V ) = 1 2 ln ( 1 + P N 1 ) . Analogously, we have I ( X ; Z | V ) = 1 2 ln ( 1 + P N 2 ) . Then, it is easy to see that Y is less noisy than Z ( I ( X ; Z | V ) I ( X ; Y | V ) for every P X | V ( x | v ) ), which can be further expressed by N 1 N 2 , and Z is less noisy than Y ( I ( X ; Z | V ) I ( X ; Y | V ) for every P X | V ( x | v ) ), which can be further expressed by N 1 N 2 .
Proof. The achievability proof of (32) and (33) is easily obtained by substituting X N ( 0 , P ) , V N ( 0 , Q ) and (20) into (30) and (31), respectively. Now, it remains to prove the converse of (32); see the following.
The converse part of (32) is based on the converse proof of (30), (see p.2846, Proof of Theorem 2 of [13] and the left bottom and right top of page 2841 [13]). However, the converse proof of (30) is for the discrete memoryless case, and it needs to be further processed for the Gaussian case. Based on the converse proof of (30) [13] and the fact that N 1 N 2 , we have the following (34) and (35),
C s - b o t h g 1 N i = 1 N ( I ( X i ; Y i | V i ) - I ( X i ; Z i | V i ) + h ( V i | Z i ) ) = ( 1 ) 1 N i = 1 N ( I ( X i ; Y i | V i ) - h ( Z i ) + h ( V i ) + h ( Z i | X i , V i ) ) = ( 2 ) 1 N i = 1 N ( h ( X i + Z 1 , i | V i ) - h ( Z 1 , i ) - h ( Z i ) + h ( V i ) + h ( Z 1 , i + Z 2 , i * ) ) 1 N i = 1 N ( h ( X i + Z 1 , i ) - h ( Z 1 , i ) - h ( Z i ) + h ( V i ) + h ( Z 1 , i + Z 2 , i * ) ) = ( 3 ) 1 N i = 1 N ( h ( X i + Z 1 , i ) - 1 2 ln ( 2 π e N 1 ) - h ( X i + V i + Z 1 , i + Z 2 , i * ) + 1 2 ln ( 2 π e Q ) + 1 2 ln ( 2 π e N 2 ) ) ( 4 ) 1 N i = 1 N ( 1 2 ln ( e 2 h ( X i + Z 1 , i ) ) - 1 2 ln ( 2 π e N 1 ) - 1 2 ln ( e 2 h ( X i + Z 1 , i ) + e 2 h ( V i + Z 2 , i * ) ) + 1 2 ln ( 2 π e Q ) + 1 2 ln ( 2 π e N 2 ) ) = ( 5 ) 1 N i = 1 N ( 1 2 ln ( e 2 h ( X i + Z 1 , i ) ) - 1 2 ln ( 2 π e N 1 ) - 1 2 ln ( e 2 h ( X i + Z 1 , i ) + 2 π e ( Q + N 2 - N 1 ) ) + 1 2 ln ( 2 π e Q ) + 1 2 ln ( 2 π e N 2 ) ) = 1 N i = 1 N ( 1 2 ln e 2 h ( X i + Z 1 , i ) e 2 h ( X i + Z 1 , i ) + 2 π e ( Q + N 2 - N 1 ) - 1 2 ln ( 2 π e N 1 ) + 1 2 ln ( 2 π e Q ) + 1 2 ln ( 2 π e N 2 ) ) ( 6 ) 1 N i = 1 N ( 1 2 ln 2 π e ( P + N 1 ) 2 π e ( P + N 1 ) + 2 π e ( Q + N 2 - N 1 ) - 1 2 ln ( 2 π e N 1 ) + 1 2 ln ( 2 π e Q ) + 1 2 ln ( 2 π e N 2 ) ) = 1 2 ln ( 1 + P N 1 ) + 1 2 ln ( 2 π e Q ) - 1 2 ln ( P + Q + N 2 N 2 ) ,
and:
C s - b o t h g 1 N i = 1 N I ( X i ; Y i | V i ) = 1 N i = 1 N ( h ( Y i | V i ) - h ( Y i | V i , X i ) ) = ( 7 ) 1 N i = 1 N ( h ( X i + Z 1 , i | V i ) - h ( Z 1 , i ) ) 1 N i = 1 N ( h ( X i + Z 1 , i ) - h ( Z 1 , i ) ) ( 8 ) 1 N i = 1 N ( 1 2 ln ( 2 π e ( P + N 1 ) ) - 1 2 ln ( 2 π e N 1 ) ) = 1 2 ln ( 1 + P N 1 ) ,
where (1) is from (29), (2) is from Definition (23) and Z 2 , i * N ( 0 , N 2 - N 1 ) , (3) is from the fact that the differential entropy of a Gaussian distributed random variable X is h ( X ) = 1 2 ln ( 2 π e D ( X ) ) (here, D ( X ) is the variance of the Gaussian random variable X), (4) is from the entropy power inequality e 2 h ( X i + V i + Z 1 , i + Z 2 , i ) e 2 h ( X i + Z 1 , i ) + e 2 h ( V i + Z 2 , i ) (see [19]), (5) is from h ( V i + Z 2 , i ) = 1 2 ln ( 2 π e ( Q + N 2 ) ) , (6) is from 1 2 ln e 2 h ( X i + Z 1 , i ) e 2 h ( X i + Z 1 , i ) + 2 π e ( Q + N 2 ) increasing while h ( X i + Z 1 , i ) is increasing and the fact that h ( X i + Z 1 , i ) 1 2 ln ( 2 π e ( P + N 1 ) ) (here, note that “=” is achieved if X i N ( 0 , P ) ), (7) is from Definition (23) and (8) is from h ( X i + Z 1 , i ) 1 2 ln ( 2 π e ( P + N 1 ) ) . Thus, the converse part of (32) is proven. The proof of Theorem 4 is completed.  ☐
Recall that for the degraded Gaussian wiretap channel with noncausal CSI at the transmitter ( ( X , V ) Y Z ), an achievable secrecy rate (a lower bound on the secrecy capacity) is provided [10]; see the following Theorem 5.
Theorem 5. For the Gaussian non-feedback model of Figure 1 with the condition that N 1 N 2 , an achievable secrecy rate C s g i is denoted by:
C s g i = max 0 α 1 min 1 2 ln ( P + N 1 ) ( P + α 2 Q ) α 2 Q ( P + N 1 ) + N 1 P - 1 2 ln P + α 2 Q P , 1 2 ln ( P + N 1 ) ( P + α 2 Q ) α 2 Q ( P + N 1 ) + N 1 P - 1 2 ln ( P + N 2 ) ( P + α 2 Q ) α 2 Q P + N 2 ( P + α 2 Q ) .
Proof. The result is directly obtained from [10], and therefore, the proof is omitted here.  ☐
Remark 5.
  • For the case N 1 N 2 , the relationship (20) of the channel inputs and outputs can be equivalently characterized by (23), which implies the Markov chain ( X , V ) Y Z .
  • To the best of the authors’ knowledge, for the case N 1 > N 2 , the bounds on the secrecy capacity of the Gaussian wiretap channel with noncausal CSI at the transmitter are still unknown.
Finally, note that if the CSI is not available at the legitimate receiver, the wiretapper and the transmitter and there is no feedback link from the legitimate receiver to the transmitter, the Gaussian case of the model of Figure 1 (see (20)) reduces to the model of the Gaussian wiretap channel, where V i and Z 1 , i of (20) are the legitimate receiver’s channel noises and V i and Z 2 , i are the wiretapper’s channel noises. From [20], it is easy to see that the secrecy capacity C s * of the Gaussian wiretap channel is given by:
C s * = 1 2 ln P + Q + N 1 Q + N 1 - 1 2 ln P + Q + N 2 Q + N 2 .
Comparing Theorem 3 to Theorem 4, we can conclude that if N 1 N 2 , for given P, N 1 and N 2 , C s g f is larger than C s - b o t h g if and only if:
Q N 1 ( N 2 - N 1 ) ( P + N 1 ) N 1 2 + N 2 P .
For the case that N 1 > N 2 , we find that if N 1 2 < N 2 < N 1 , for given P, N 1 and N 2 , C s g f is larger than C s - b o t h g i if and only if:
Q N 2 ( N 1 - N 2 ) 2 N 2 - N 1 .
If N 2 = N 1 2 , C s g f is always larger than C s - b o t h g i .
If N 2 < N 1 2 , for given P, N 1 and N 2 , C s g f is larger than C s - b o t h g i if and only if:
Q N 2 ( N 1 - N 2 ) 2 N 2 - N 1 .
Figure 2. For N 1 N 2 , the relationships of P - C s g f , P - C s - b o t h g , P - C s g i and P - C s * for several values of N 1 , N 2 and Q.
Figure 2. For N 1 N 2 , the relationships of P - C s g f , P - C s - b o t h g , P - C s g i and P - C s * for several values of N 1 , N 2 and Q.
Entropy 17 07852 g002
For the case N 1 N 2 , Figure 2 plots the relationships of P - C s * , P - C s g i , P - C s g f and P - C s - b o t h g for several values of N 1 , N 2 and Q. It is easy to see that the noiseless feedback ( C s g f ), the CSI sharing scheme ( C s - b o t h g ) and the CSI only available at the transmitter ( C s g i ) help to enhance the secrecy capacity C s * of the Gaussian wiretap channel. Furthermore, we can see that both the noiseless feedback and the CSI sharing scheme perform better than the CSI only available at the transmitter. Moreover, when Q is small ( Q = 0 . 1 , 0 . 5 ), the noiseless feedback performs better than the CSI sharing scheme, and while Q is increasing ( Q = 1 ), the CSI sharing scheme is beginning to take advantage of the noiseless feedback.
For the case N 1 > N 2 , the following Figure 3 plots the relationships of P - C s g f and P - C s - b o t h g for several values of N 1 , N 2 and Q. Since C s * = 0 for the case that N 1 > N 2 , both the noiseless feedback ( C s g f ) and the CSI sharing scheme ( C s - b o t h g i ) enhance the secrecy capacity C s * of the Gaussian wiretap channel. Moreover, we can see that for fixed Q, if the gap between the legitimate receiver’s channel noise variance N 1 and the wiretapper’s channel noise variance N 2 is large, the noiseless feedback performs better than the CSI sharing scheme, and vice versa.
Figure 3. For N 1 > N 2 , the relationships of P - C s g f and P - C s - b o t h g i for several values of N 1 , N 2 and Q.
Figure 3. For N 1 > N 2 , the relationships of P - C s g f and P - C s - b o t h g i for several values of N 1 , N 2 and Q.
Entropy 17 07852 g003

3.2. Binary Case of the Model of Figure 1

In this subsection, we calculate the secrecy capacity of a degraded binary case of the model of Figure 1 with causal CSI at the transmitter, where “degraded” means that there exists a Markov chain ( X , V ) Y Z .
Suppose that the random variable V is uniformly distributed over { 0 , 1 } , i.e., p V ( 0 ) = p V ( 1 ) = 1 2 . Meanwhile, the random variables X, Y and Z take values in { 0 , 1 } , and the wiretap channel is a BSC (binary symmetric channel) with crossover probability q. The transition probability of the main channel is defined as follows:
When v = 0 ,
p Y | X , V ( y | x , v = 0 ) = 1 - p , if y = x , p , otherwise .
When v = 1 ,
p Y | X , V ( y | x , v = 1 ) = p , if y = x , 1 - p , otherwise .
From Remark 2, we know that the secrecy capacity for the model of Figure 1 with causal CSI at the transmitter is given by:
C s ( c f ) = max P K ( k ) P X | K , V ( x | k , v ) min { I ( K ; Y ) , H ( Y | Z ) }
and the maximum achievable secrecy rate C s ( c i ) of the wiretap channel with causal CSI [12] is given by:
C s ( c i ) = max P K ( k ) P X | K , V ( x | k , v ) ( I ( K ; Y ) - I ( K ; Z ) ) ,
where (43) is from (19).
In addition, from ([13], Theorem 3), we know that the secrecy capacity C s - b o t h of the wiretap channel with CSI causally or non-causally at both the transmitter and the legitimate receiver is given by:
C s - b o t h = max P X | V ( x | v ) min { I ( X ; Y | V ) - I ( X ; Z | V ) + H ( V | Z ) , I ( X ; Y | V ) } .
It remains to calculate C s ( c f ) , C s ( c i ) and C s - b o t h ; see the following.
The calculation of C s ( c f ) and C s ( c i ) :
Let K take values in { 0 , 1 } . The probability of K is defined as follows. p K ( 0 ) = α , and p K ( 1 ) = 1 - α . Define the conditional probability mass function p X | K , V as follows.
p X | K , V ( 0 | 0 , 0 ) = β 1 , p X | K , V ( 1 | 0 , 0 ) = 1 - β 1 , p X | K , V ( 0 | 0 , 1 ) = β 2 , p X | K , V ( 1 | 0 , 1 ) = 1 - β 2 ,
p X | K , V ( 0 | 1 , 0 ) = β 3 , p X | K , V ( 1 | 1 , 0 ) = 1 - β 3 , p X | K , V ( 0 | 1 , 1 ) = β 4 , p X | K , V ( 1 | 1 , 1 ) = 1 - β 4 .
The joint probability mass functions p K Y is calculated by:
p K Y ( k , y ) = x , v p K Y X V ( k , y , x , v ) = x , v p Y | X V ( y | x , v ) p X | K , V ( x | k , v ) p K ( k ) p V ( v ) .
Then, we have:
p K Y ( 0 , 0 ) = α 2 [ 1 - ( β 1 - β 2 ) ( 1 - 2 p ) ] ,
p K Y ( 0 , 1 ) = α 2 [ 1 + ( β 1 - β 2 ) ( 1 - 2 p ) ] ,
p K Y ( 1 , 0 ) = α 2 [ 1 - ( β 3 - β 4 ) ( 1 - 2 p ) ] ,
p K Y ( 1 , 1 ) = α 2 [ 1 + ( β 3 - β 4 ) ( 1 - 2 p ) ] .
By calculating, we have:
C s ( c f ) = min { 1 - h ( p ) , h ( q ) } ,
and:
C s ( c i ) = h ( p + q - 2 p q ) - h ( p ) ,
where h ( x ) = - x log x - ( 1 - x ) log ( 1 - x ) and 0 x 1 .
The calculation of C s - b o t h :
Define p X | V ( 0 | 0 ) = α , p X | V ( 1 | 0 ) = 1 - α , p X | V ( 0 | 1 ) = β , p X | V ( 1 | 1 ) = 1 - β .
By calculating, C s - b o t h is given by:
C s - b o t h = min { 1 - h ( p ) , 1 - h ( p ) + h ( p + q - 2 p q ) } = 1 - h ( p ) .
The following Figure 4, Figure 5 and Figure 6 show C s ( c f ) , C s ( c i ) and C s - b o t h for several values of q. Here, note that the noise of the wiretap channel is increasing while q is increasing. It is easy to see that when q < 0 . 5 , C s - b o t h and C s ( c f ) are always larger than C s ( c i ) , i.e., both the noiseless feedback (the model of this paper) and the shared CSI [13] help to enhance the security of the wiretap channel with causal CSI at the transmitter. When q = 0 . 5 , there is no wiretapper in the channel; thus, C s ( c f ) = C s ( c i ) = C s - b o t h = 1 - h ( p ) .
Moreover, from Figure 4, Figure 5 and Figure 6, we see that the noiseless feedback performs no better than the shared CSI. However, when q is large enough (satisfying h ( q ) 1 - h ( p ) ), the two ways perform the same.
Figure 4. The C s ( c f ) , C s ( c i ) and C s - b o t h for q = 0 . 1 .
Figure 4. The C s ( c f ) , C s ( c i ) and C s - b o t h for q = 0 . 1 .
Entropy 17 07852 g004
Figure 5. The C s ( c f ) , C s ( c i ) and C s - b o t h for q = 0 . 2 .
Figure 5. The C s ( c f ) , C s ( c i ) and C s - b o t h for q = 0 . 2 .
Entropy 17 07852 g005
Figure 6. The C s ( c f ) , C s ( c i ) and C s - b o t h for q = 0 . 5 .
Figure 6. The C s ( c f ) , C s ( c i ) and C s - b o t h for q = 0 . 5 .
Entropy 17 07852 g006

4. Conclusions

In this paper, we study the general wiretap channel with CSI and noiseless feedback, where the CSI is available at the transmitter in a noncausal or causal manner. Both the capacity-equivocation region and the secrecy capacity are determined for the noncausal and causal cases, and the results are further explained via Gaussian and binary examples. For the Gaussian example, we show that both the noiseless feedback and the CSI sharing scheme [13] help to enhance the security of the Gaussian wiretap channel. Moreover, we show that in some particular cases, the noiseless feedback performs even better than the CSI sharing scheme [13]. For the degraded binary example, we also find that the noiseless feedback enhances the security of the wiretap channel with causal CSI. Unlike the Gaussian example, we find that the noiseless feedback always performs no better than the CSI sharing scheme [13].

Acknowledgment

The authors would like to thank the anonymous reviewers for their valuable suggestions to improve this paper. This work was supported by a sub-project in the National Basic Research Program of China under Grant 2012CB316100 on Broadband Mobile Communications at High Speeds, the National Natural Science Foundation of China under Grant 61301121, the Fundamental Research Fund for the Central Universities under Grant 2682014CX099, the Key Grant Project of Chinese Ministry of Education (No. 311031 100), the Young Innovative Research Team of Sichuan Province (2011JTD0007) and the Open Research Fund of National Mobile Communications Research Laboratory, Southeast University (No. 2014D01).

Author Contributions

Bin Dai designed research; Bin Dai and Zheng Ma performed research; Linman Yu analyzed the data; Bin Dai wrote the paper. All authors have read and approved the final manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix

A. Converse Proof of Theorem 1

Given an achievable ( R , R e ) pair, we need to show that there exists a joint distribution of the form P Z | Y ( z | y ) P Y | X V ( y | x , v ) P X | K V ( x | k , v ) P K V ( k , v ) , such that,
0 R e R ,
0 R I ( K ; Y ) - I ( K ; V ) ,
R e H ( Y | Z ) .

A.1. Proof of (A.1)

R e lim N Δ lim N 1 N H ( W ) = lim N log W N = R .

A.2. Proof of (A.2)

1 N H ( W ) = 1 N ( I ( W ; Y N ) + H ( W | Y N ) ) ( a ) 1 N ( I ( W ; Y N ) + δ ( P e ) ) = ( b ) 1 N ( I ( W ; Y N ) - I ( W ; V N ) + δ ( P e ) ) = ( c ) 1 N i = 1 N ( I ( Y i ; W , V i + 1 N | Y i - 1 ) - I ( V i ; W , Y i - 1 | V i + 1 N ) + δ ( P e ) ) ( d ) 1 N i = 1 N ( H ( Y i ) - H ( Y i | Y i - 1 , W , V i + 1 N ) - H ( V i ) + H ( V i | V i + 1 N , W , Y i - 1 ) + δ ( P e ) ) = 1 N i = 1 N ( I ( Y i ; W , V i + 1 N , Y i - 1 ) - I ( V i ; W , Y i - 1 , V i + 1 N ) + δ ( P e ) ) = ( e ) 1 N i = 1 N ( I ( Y i ; W , V i + 1 N , Y i - 1 | J = i ) - I ( V i ; W , Y i - 1 , V i + 1 N | J = i ) + δ ( P e ) ) = ( f ) I ( Y J ; W , V J + 1 N , Y J - 1 | J ) - I ( V J ; W , Y J - 1 , V J + 1 N | J ) + δ ( P e ) N ( g ) I ( Y J ; W , V J + 1 N , Y J - 1 , J ) - I ( V J ; W , Y J - 1 , V J + 1 N , J ) + δ ( P e ) N = ( h ) I ( K ; Y ) - I ( K ; V ) + δ ( P e ) N ,
where (a) is from Fano’s inequality, (b) is from W is independent of V N , (c) is from Csisz a ´ r’s equality:
i = 1 N I ( Y i ; V i + 1 N | Y i - 1 , W ) = i = 1 N I ( V i ; Y i - 1 | V i + 1 N , W ) ,
(d) is from V i being independent of V i + 1 N , (e) and (f) are from J being a random variable (uniformly distributed over [ 1 , N ] ) and being independent of W, V N and Y N , (g) is from V J being independent of J and (h) is from the definitions that Y Y J , V V J and K ( W , Y J - 1 , V J + 1 N , J ) .
By using P e ϵ , ϵ 0 as N , lim N H ( W ) N = R and (A.4), it is easy to see that R I ( K ; Y ) - I ( K ; V ) .

A.3. Proof of (A.3)

1 N H ( W | Z N ) ( 1 ) 1 N ( I ( W ; Y N | Z N ) + δ ( P e ) ) 1 N i = 1 N H ( Y i | Z i ) + δ ( P e ) N = ( 2 ) 1 N i = 1 N H ( Y i | Z i , J = i ) + δ ( P e ) N ( 3 ) H ( Y J | Z J , J ) + δ ( P e ) N ( 4 ) H ( Y | Z ) + δ ( P e ) N ,
where (1) is from Fano’s inequality, (2) is from J being a random variable (uniformly distributed over { 1 , 2 , . . . , N } ) and being independent of Y N and Z N , (3) is from J being uniformly distributed over { 1 , 2 , . . . , N } and (4) is from the definitions that Y Y J , and Z Z J .
By using P e ϵ , ϵ 0 as N , lim N H ( W | Z N ) N R e and (A.6), it is easy to see that R e H ( Y | Z ) .
The converse proof of Theorem 1 is completed.

B. Direct Proof of Theorem 1

The direct part (achievability) of Theorem 1 is proven by considering the following two cases.
  • Case 1: If I ( K ; Y ) - I ( K ; V ) H ( Y | Z ) , we need to show that ( R = I ( K ; Y ) - I ( K ; V ) - ϵ , R e = H ( Y | Z ) ) is achievable, where ϵ 0 + .
  • Case 2: If I ( K ; Y ) - I ( K ; V ) H ( Y | Z ) , we need to show that ( R = I ( K ; Y ) - I ( K ; V ) - ϵ , R e = R = I ( K ; Y ) - I ( K ; V ) - ϵ ) is achievable.
The direct proof of Theorem 1 is organized as follows. The balanced coloring lemma introduced by Ahlswede and Cai is provided in Subsection B.1, and it will be used in the remainder of this section. The code-book generation is shown in Subsection B.2, and the equivocation analysis is given in Subsection B.3.

B.1. The Balanced Coloring Lemma

The balanced coloring lemma was first introduced by Ahlswede and Cai; see the following.
Lemma 1. Balanced coloring lemma: For all ϵ 1 , ϵ 2 , ϵ 3 , δ > 0 , sufficiently large N and all N-type P Y ( y ) , there exists a γ-coloring c : T Y N ( ϵ 1 ) { 1 , 2 , . . , γ } of T Y N ( ϵ 1 ) such that for all joint N-type P Y Z ( y , z ) with marginal distribution P Z ( z ) and | T Y | Z N ( z N ) | γ > 2 N ϵ 2 , z N T Z N ( ϵ 3 ) ,
| c - 1 ( k ) | | T Y | Z N ( z N ) | ( 1 + δ ) γ ,
for k = 1 , 2 , . . . , γ , where c - 1 is the inverse image of c.
Proof. Letting U = c o n s t , Lemma 1 is directly from p. 259 of [1], and thus, we omit it here.  ☐
Lemma 1 shows that if y N and z N are joint typical, for given z N , the number of y N T Y | Z N ( z N ) for a certain color k ( k = 1 , 2 , . . . , γ ), which is denoted as | c - 1 ( k ) | , is upper bounded by | T Y | Z N ( z N ) | ( 1 + δ ) γ . By using Lemma 1, it is easy to see that the typical set T Y | Z N ( z N ) maps into at least:
| T Y | Z N ( z N ) | | T Y | Z N ( z N ) | ( 1 + δ ) γ = γ 1 + δ
colors. On the other hand, the typical set T Y | Z N ( z N ) maps into at most γ colors.

B.2. Code-Book Generation

Fix the joint probability mass function P Z , Y | X , V ( z , y | x , v ) P X | K , V ( x | k , v ) P K V ( k , v ) . The message set W satisfies:
lim N log W N = R = I ( K ; Y ) - I ( K ; V ) - ϵ .
Let W = { 1 , 2 , . . . , 2 N R } .
The block Markov encoding scheme is used in the direct proof of Theorem 1. The random vectors K N , V N , X N , Y N and Z N consist of n blocks of length N. Let K ˜ i , V ˜ i , Y ˜ i and Z ˜ i ( 1 i n ) be the random vectors for block i. Define k ˜ n = ( k ˜ 1 , k ˜ 2 , . . . , k ˜ n ) , v ˜ n = ( v ˜ 1 , v ˜ 2 , . . . , v ˜ n ) , y ˜ n = ( y ˜ 1 , y ˜ 2 , . . . , y ˜ n ) and z ˜ n = ( z ˜ 1 , z ˜ 2 , . . . , z ˜ n ) to be the specific vectors for all blocks. The message W n for all n blocks is denoted by W n = ( W 1 , W 2 , . . . , W n ) , where W i ( 2 i n ) is uniformly distributed over the alphabet W , and W i is independent of W j ( 2 j n and j i ). Note that w 1 does not exist.
Construction of K N :
Gel’fand and Pinsker’s binning and block Markov coding scheme are used in the construction of K N .
  • Construction of K N for Case 1:
    For each block, generate 2 N ( I ( K ; Y ) - ϵ 2 , N ) ( ϵ 2 , N 0 ) i.i.d. sequences of k N , according to p K ( k ) . Partition these sequences at random into 2 N R = 2 N ( I ( K ; Y ) - I ( K ; V ) - γ 1 ) bins, such that each bin has 2 N ( I ( K ; V ) + γ 1 - ϵ 2 , N ) sequences. Index each bin by l { 1 , 2 , . . . , 2 N R } .
    Denote the message w i ( 2 i n ) by w i = ( w i 1 , w i 2 ) , where w i 1 W i 1 = { 1 , 2 , . . . , 2 N H ( Y | Z ) } and w i 2 W i 2 = { 1 , 2 , . . . , 2 N ( R - H ( Y | Z ) ) } . Here, note that W i 1 is independent of W i 2 .
    In the first block, for a given side information v ˜ 1 , try to find a k ˜ 1 , such that ( k ˜ 1 , v ˜ 1 ) T K V N ( ϵ ) . If multiple sequences exist, randomly choose one for transmission. If there is no such sequence, declare an encoding error.
    For the i-th block ( 2 i n ), the transmitter receives the output y ˜ i - 1 of the i - 1 -th block; he or she gives up if y ˜ i - 1 T Y N ( ϵ 2 ) ( ϵ 2 0 as N ). It is easy to see that the probability for giving up at the i - 1 -th block tends to zero as N . In the case y ˜ i - 1 T Y N ( ϵ 2 ) , generate a mapping g f : T Y N ( ϵ 2 ) { 1 , 2 , . . . , 2 N H ( Y | Z ) } . Define a random variable K i * by K i * = g f ( Y ˜ i - 1 ) ( 2 i n ), and it is uniformly distributed over the set W i 1 = { 1 , 2 , . . . , 2 N H ( Y | Z ) } . K i * is independent of W i . Reveal the mapping g f to the legitimate receiver, the wiretapper and the transmitter. Then, since the transmitter gets y ˜ i - 1 , he computes k i * = g f ( y ˜ i - 1 ) { 1 , 2 , . . . , 2 N H ( Y | Z ) } . For a given w i = ( w i 1 , w i 2 ) ( 2 i n ), the transmitter selects a sequence k ˜ i in the bin ( w i 1 k i * , w i 2 ) (where ⊕ is the modulo addition over W i 1 ), such that ( k ˜ i , v ˜ i ) T K V N ( ϵ ) . If multiple sequences in bin ( w i 1 k i * , w i 2 ) exist, choose the sequence with the smallest index in the bin. If there is no such sequence, declare an encoding error. Here, note that since K i * is independent of W i = ( W i 1 , W i 2 ) , W i 1 K i * is independent of W i and K i * . The proof is given as follows.
    Proof. Since:
    P r { K i * W i 1 = a } = k i * W i 1 P r { K i * W i 1 = a , K i * = k i * } = k i * W i 1 P r { W i 1 = a k i * , K i * = k i * } = k i * W i 1 P r { W i 1 = a k i * } P r { K i * = k i * } = k i * W i 1 1 W i 1 2 = 1 W i 1 ,
    and:
    P r { K i * W i 1 = a , K i * = k i * } = P r { W i 1 = a k i * , K i * = k i * } = P r { W i 1 = a k i * } P r { K i * = k i * } = 1 W i 1 2 ,
    it is easy to see that P r { K i * W i 1 = a , K i * = k i * } = P r { K i * W i 1 = a } · P r { K i * = k i * } , which implies that K i * W i 1 is independent of K i * .
    Analogously, we can prove that P r { K i * W i 1 = a , W i 1 = w i 1 , W i 2 = w i 2 } = P r { K i * W i 1 = a } · P r { W i 1 = w i 1 } · P r { W i 2 = w i 2 } , which implies that K i * W i 1 is independent of W i = ( W i 1 , W i 2 ) . Thus, the proof of W i 1 K i * is independent of W i , and K i * is completed.
     ☐
  • Construction of K N for Case 2: The construction of K N for Case 2 is similar to that of Case 1, except that there is no need to divide w i into two parts. The detail is as follows. For the i-th block ( 2 i n ), if y ˜ i - 1 T Y N ( ϵ 2 ) , generate a mapping g f : T Y N ( ϵ 2 ) W (note that | T Y N ( ϵ 2 ) | | W | ). Let K i * = g f ( Y ˜ i - 1 ) ( 2 i n ), and it is uniformly distributed over the set W . K i * is independent of W i . Reveal the mapping g f to the legitimate receiver, the wiretapper and the transmitter. When the transmitter receives the feedback y ˜ i - 1 of the i - 1 -th block, he or she computes k i * = g f ( y ˜ i - 1 ) W . For a given transmitted message w i ( 2 i n ), the transmitter selects a codeword k ˜ i in the bin w i k i * (where ⊕ is the modulo addition over W ), such that ( k ˜ i , v ˜ i ) T K V N ( ϵ ) . If multiple sequences in bin w i k i * exist, select the one with the smallest index in the bin. If there is no such sequence, declare an encoding error. Here, note that W i K i * is independent of W i and K i * , and the proof is similar to that of Case 1. Thus, we omit the proof here.
Construction of X N :
In each block, the channel input x N is generated by a pre-fixed discrete memoryless channel with transition probability P X | K , V ( x | k , v ) . The inputs of the channel are k N and v N , and the output is x N .
Here, note that for Case 1, the random vector K ˜ i of block i ( 2 i n ) is i.i.d. generated corresponding to the encrypted message ( W i 1 K i * , W i 2 ) and V ˜ i (here, V ˜ i is also i.i.d. generated according to the probability mass function P V ( v ) ). Since Y ˜ i and Z ˜ i are generated according to K ˜ i , V ˜ i and the discrete memoryless channel, the only connection between ( W i , K ˜ i , V ˜ i , Y ˜ i , Z ˜ i ) of the i-th block and ( W i - 1 , K ˜ i - 1 , V ˜ i - 1 , Y ˜ i - 1 , Z ˜ i - 1 ) of the i - 1 -th block is the secret key K i * , which is generated by Y ˜ i - 1 . As stated above, both the encrypted message ( W i 1 K i * , W i 2 ) and the real message W i = ( W i 1 , W i 2 ) are independent of K i * , and thus, ( W i , K ˜ i , V ˜ i , Y ˜ i , Z ˜ i ) of the i-th block are independent of ( W i - 1 , K ˜ i - 1 , V ˜ i - 1 , Y ˜ i - 1 , Z ˜ i - 1 ) of the i - 1 -th block. Since ( W i 1 K i * , W i 2 ) and W i are also independent of W j and K j * ( 2 i , j n and j i ), it is easy to see that ( W i , K ˜ i , V ˜ i , Y ˜ i , Z ˜ i ) are independent of ( W j , K ˜ j , V ˜ j , Y ˜ j , Z ˜ j ) . Finally, note that ( W i 1 K i * , W i 2 ) ( 2 i n ) is independent of K 2 * (generated by Y ˜ 1 ); thus, ( W i , K ˜ i , V ˜ i , Y ˜ i , Z ˜ i ) are independent of ( K ˜ 1 , V ˜ 1 , Y ˜ 1 , Z ˜ 1 ) .
Analogously, in Case 2, for 2 i , j n and j i , the fact that ( W i , K ˜ i , V ˜ i , Y ˜ i , Z ˜ i ) are independent of ( W j , K ˜ j , V ˜ j , Y ˜ j , Z ˜ j ) and ( K ˜ 1 , V ˜ 1 , Y ˜ 1 , Z ˜ 1 ) also holds.
Decoding: For block i ( 2 i n ), given a vector y ˜ i Y N , try to find a sequence k ˜ i ( w ^ i 1 k i * , w ^ i 2 , j ^ ) (Case 1) or k ˜ i ( w ^ i k i * , j ^ ) (Case 2), such that k ˜ i and y ˜ i are joint typical. If there exists a unique sequence, put out the corresponding index of the bin ( w ^ i 1 k i * , w ^ i 2 ) or w ^ i k i * . Otherwise, declare a decoding error. Since the legitimate receiver has k i * , put out the corresponding w ^ i from ( w ^ i 1 k i * , w ^ i 2 ) or w ^ i k i * .

B.3. Proof of Achievability

Here, note that the above encoding-decoding scheme for the achievability proof of Theorem 1 is exactly the same as that in [11], except that the transmitter transmits an “encrypted message” by using the secret key k i * . Since the legitimate receiver has k i * , the decoding scheme for the achievability proof of Theorem 1 is in fact the same as that in [11]. Hence, we omit the proof of P e ϵ here. It remains to prove that lim N Δ R e ; see the following.
  • For Case 1, part of the message w i is encrypted by k i * . In the analysis of the equivocation, we drop w i 2 from w i . Then, the equivocation about w i is equivalent to the equivocation about k i * . Since k i * = g f ( y ˜ i - 1 ) , the wiretapper tries to guess k i * from y ˜ i - 1 . Note that for a given z ˜ i - 1 and sufficiently large N, P r { y ˜ i - 1 T Y | Z N ( z ˜ i - 1 ) } 1 . Thus, the wiretapper can guess y ˜ i - 1 from the conditional typical set T Y | Z N ( z ˜ i - 1 ) . By using the above Lemma 1 and (B.2), the set T Y | Z N ( z ˜ i - 1 ) maps into at least 2 N H ( Y | Z ) 1 + δ (here, γ = 2 N H ( Y | Z ) ) k i * (colors). Thus, in the i-th block, the uncertainty about K i * is bounded by:
    1 N H ( K i * | Z ˜ i - 1 ) H ( Y | Z ) - log ( 1 + δ ) N ,
    Here, note that K i * is uniformly distributed.
  • For Case 2, the alphabet of the secret key k i * equals the alphabet W i = { 1 , 2 , . . . , 2 N R } , and the encrypted message is denoted by w i k i * . Then, by using the above Lemma 1 and (B.2), the set T Y | Z N ( z ˜ i - 1 ) maps into at least 2 N R 1 + δ (here, γ = 2 N R ) k i * (colors). Thus, in the i-th block, the uncertainty about K i * is bounded by:
    1 N H ( K i * | Z ˜ i - 1 ) R - log ( 1 + δ ) N .
Proof of lim N Δ R e for Case 1:
Δ = H ( W n | Z n ) n N = i = 2 n H ( W i | W i - 1 , Z n ) n N = ( a ) i = 2 n H ( W i | Z ˜ i , Z ˜ i - 1 ) n N i = 2 n H ( W i 1 | Z ˜ i , Z ˜ i - 1 ) n N i = 2 n H ( W i 1 | Z ˜ i , Z ˜ i - 1 , W i 1 K i * ) n N i = 2 n H ( W i 1 | W i 2 , Z ˜ i , Z ˜ i - 1 , W i 1 K i * ) n N = ( b ) i = 2 n H ( W i 1 | W i 2 , Z ˜ i - 1 , W i 1 K i * ) n N = ( c ) i = 2 n H ( W i 1 | Z ˜ i - 1 , W i 1 K i * ) n N = i = 2 n H ( K i * | Z ˜ i - 1 , W i 1 K i * ) n N = ( d ) i = 2 n H ( K i * | Z ˜ i - 1 ) n N ( e ) i = 2 n ( N H ( Y | Z ) - log ( 1 + δ ) ) n N = ( n - 1 ) ( N H ( Y | Z ) - log ( 1 + δ ) ) n N ,
where (a) is from W i ( Z ˜ i , Z ˜ i - 1 ) ( W i - 1 , Z ˜ i - 2 , Z ˜ i + 1 n ) (proven in the remainder of this section), (b) is from W i 1 ( W i 2 , W i 1 K i * , Z ˜ i - 1 ) Z ˜ i (proven in the remainder of this section), (c) is from W i 2 being independent of Z ˜ i - 1 , W i 1 K i * and W i 1 , (d) follows from the fact that W i 1 K i * is independent of K i * , W i 1 and Z ˜ i - 1 and (e) is from (B.6).
Letting N and n , it is easy to see that:
lim N Δ = lim N lim n H ( W n | Z n ) n N H ( Y | Z ) = R e .
The proof of lim N Δ R e for Case 1 is completed.
Proof of lim N Δ R e for Case 2:
Δ = H ( W n | Z n ) n N = ( a ) i = 2 n H ( W i | Z ˜ i , Z ˜ i - 1 ) n N i = 2 n H ( W i | Z ˜ i , Z ˜ i - 1 , W i K i * ) n N = ( b ) i = 2 n H ( W i | Z ˜ i - 1 , W i K i * ) n N = i = 2 n H ( K i * | Z ˜ i - 1 , W i K i * ) n N = ( c ) i = 2 n H ( K i * | Z ˜ i - 1 ) n N ( d ) i = 2 n ( N R - log ( 1 + δ ) ) n N = ( n - 1 ) ( N R - log ( 1 + δ ) ) n N ,
where (a) is from W i ( Z ˜ i , Z ˜ i - 1 ) ( W i - 1 , Z ˜ i - 2 , Z ˜ i + 1 n ) (proven in the remainder of this section), (b) is from W i ( W i K i * , Z ˜ i - 1 ) Z ˜ i (proven in the remainder of this section), (c) follows from the fact that W i K i * is independent of K i * and Z ˜ i - 1 and (d) is from (B.7).
Letting N and n , it is easy to see that:
lim N Δ = lim N lim n H ( W n | Z n ) n N R = R e .
The proof of lim N Δ R e for Case 2 is completed.
It remains to prove the Markov chains W i ( Z ˜ i , Z ˜ i - 1 ) ( W i - 1 , Z ˜ i - 2 , Z ˜ i + 1 n ) and W i 1 ( W i 2 , W i 1 K i * , Z ˜ i - 1 ) Z ˜ i of the proof of lim N Δ R e for Case 1 and W i ( Z ˜ i , Z ˜ i - 1 ) ( W i - 1 , Z ˜ i - 2 , Z ˜ i + 1 n ) , W i ( W i K i * , Z ˜ i - 1 ) Z ˜ i of the proof of lim N Δ R e for Case 2.
Proof. Proof of W i ( Z ˜ i , Z ˜ i - 1 ) ( W i - 1 , Z ˜ i - 2 , Z ˜ i + 1 n ) for Case 1:
For convenience, we denote the probability P r { V = v } by P r { v } .
By definition, W i ( Z ˜ i , Z ˜ i - 1 ) ( W i - 1 , Z ˜ i - 2 , Z ˜ i + 1 n ) holds if and only if:
P r { w i | z ˜ i , z ˜ i - 1 , w i - 1 , z ˜ i - 2 , z ˜ i + 1 n } = P r { w i | z ˜ i , z ˜ i - 1 } .
Equation (B.12) can be further expressed as:
P r { w i , z ˜ i , z ˜ i - 1 , w i - 1 , z ˜ i - 2 , z ˜ i + 1 n } P r { z ˜ i , z ˜ i - 1 , w i - 1 , z ˜ i - 2 , z ˜ i + 1 n } = P r { w i , z ˜ i , z ˜ i - 1 } P r { z ˜ i , z ˜ i - 1 } .
It remains to calculate the joint probabilities in (B.13); see the following.
   P r { w i , z ˜ i , z ˜ i - 1 , w i - 1 , z ˜ i - 2 , z ˜ i + 1 n } = P r { w i , z ˜ n } = v ˜ n y ˜ n k ˜ n P r { w i , z ˜ n , v ˜ n , y ˜ n , k ˜ n } = ( a ) v ˜ n y ˜ n k ˜ n P r { w i , z ˜ i , v ˜ i , y ˜ i , k ˜ i } · P r { z ˜ i + 1 n , v ˜ i + 1 n , y ˜ i + 1 n , k ˜ i + 1 n } = v ˜ i y ˜ i k ˜ i P r { w i , z ˜ i , v ˜ i , y ˜ i , k ˜ i } v ˜ i + 1 n y ˜ i + 1 n k ˜ i + 1 n P r { z ˜ i + 1 n , v ˜ i + 1 n , y ˜ i + 1 n , k ˜ i + 1 n } = v ˜ i y ˜ i k ˜ i P r { w i , z ˜ i , v ˜ i , y ˜ i , k ˜ i } P r { z ˜ i + 1 n } = ( b ) P r { z ˜ i + 1 } · · · · P r { z ˜ n } ( v ˜ i y ˜ i k ˜ i P r { w i , z ˜ i , v ˜ i , y ˜ i , k ˜ i } ) = P r { z ˜ i + 1 } · · · · P r { z ˜ n } ( v ˜ i y ˜ i k ˜ i P r { k ˜ i | w i , z ˜ i , v ˜ i , y ˜ i } P r { w i , z ˜ i , v ˜ i , y ˜ i } ) = ( c ) P r { z ˜ i + 1 } · · · · P r { z ˜ n } ( v ˜ i y ˜ i P r { w i , z ˜ i , v ˜ i , y ˜ i } ) = ( d ) P r { z ˜ i + 1 } · · · · P r { z ˜ n } ( v ˜ i y ˜ i P r { z ˜ 1 , v ˜ 1 , y ˜ 1 } j = 2 i P r { w j , z ˜ j , v ˜ j , y ˜ j } ) = P r { z ˜ i + 1 } · · · · P r { z ˜ n } ( v ˜ 1 y ˜ 1 P r { z ˜ 1 , v ˜ 1 , y ˜ 1 } ) ( v ˜ 2 y ˜ 2 P r { w 2 , z ˜ 2 , v ˜ 2 , y ˜ 2 } ) · · · ( v ˜ i y ˜ i P r { w i , z ˜ i , v ˜ i , y ˜ i } ) = P r { z ˜ i + 1 } · · · · P r { z ˜ n } P r { z ˜ 1 } P r { w 2 , z ˜ 2 } · · · P r { w i , z ˜ i } ,
where (a) is from the fact that ( z ˜ i + 1 n , v ˜ i + 1 n , y ˜ i + 1 n , k ˜ i + 1 n ) are independent of ( w i , z ˜ i , v ˜ i , y ˜ i , k ˜ i ) , (b) is from the fact that Z ˜ j is independent of Z ˜ l for all of the i + 1 j , l n and j l , (c) is from the fact that given w i , z ˜ i , v ˜ i and y ˜ i , k ˜ i is uniquely determined, and (d) follows from the fact that ( z ˜ 1 , v ˜ 1 , y ˜ 1 ) , ( w 2 , z ˜ 2 , v ˜ 2 , y ˜ 2 ) , ..., ( w i , z ˜ i , v ˜ i , y ˜ i ) are independent.
Replacing i by i - 1 , the joint probability P r { z ˜ i , z ˜ i - 1 , w i - 1 , z ˜ i - 2 , z ˜ i + 1 n } can be calculated by:
P r { z ˜ i , z ˜ i - 1 , w i - 1 , z ˜ i - 2 , z ˜ i + 1 n } = P r { w i - 1 , z ˜ n } = ( e ) P r { z ˜ i } · · · · P r { z ˜ n } P r { z ˜ 1 } P r { w 2 , z ˜ 2 } · · · P r { w i - 1 , z ˜ i - 1 } ,
where (e) follows from (B.14) (replacing i by i - 1 ).
Substituting (B.14) and (B.15) into the left-hand side of (B.13), we have:
P r { w i , z ˜ i , z ˜ i - 1 , w i - 1 , z ˜ i - 2 , z ˜ i + 1 n } P r { z ˜ i , z ˜ i - 1 , w i - 1 , z ˜ i - 2 , z ˜ i + 1 n } = P r { w i , z ˜ i } P r { z ˜ i } .
Next, we need to calculate the right-hand side of (B.13); see the following.
P r { w i , z ˜ i , z ˜ i - 1 } = ( 1 ) P r { w i , z ˜ i } P r { z ˜ i - 1 } ,
where (1) is from the fact that W i and Z ˜ i are independent of Z ˜ i - 1 .
The joint probability P r { z ˜ i , z ˜ i - 1 } is calculated by:
P r { z ˜ i , z ˜ i - 1 } = ( 2 ) P r { z ˜ i } · P r { z ˜ i - 1 } ,
where (1) is from the fact that Z ˜ i is independent of Z ˜ i - 1 .
Substituting (B.17) and (B.18) into the right-hand side of (B.13), we have:
P r { w i , z ˜ i , z ˜ i - 1 } P r { z ˜ i , z ˜ i - 1 } = P r { w i , z ˜ i } P r { z ˜ i } .
By checking (B.16) and (B.19), the Markov chain W i ( Z ˜ i , Z ˜ i - 1 ) ( W i - 1 , Z ˜ i - 2 , Z ˜ i + 1 n ) is proven.  ☐
Proof. Proof of W i 1 ( W i 2 , W i 1 K i * , Z ˜ i - 1 ) Z ˜ i for Case 1:
By definition, W i 1 ( W i 2 , W i 1 K i * , Z ˜ i - 1 ) Z ˜ i holds if and only if:
P r { w i 1 | w i 2 , w i 1 k i * , z ˜ i - 1 , z ˜ i } = P r { w i 1 | w i 2 , w i 1 k i * , z ˜ i - 1 } .
Equation (B.20) can be further expressed as:
P r { w i 1 , w i 2 , w i 1 k i * , z ˜ i - 1 , z ˜ i } P r { w i 2 , w i 1 k i * , z ˜ i - 1 , z ˜ i } = P r { w i 1 , w i 2 , w i 1 k i * , z ˜ i - 1 } P r { w i 2 , w i 1 k i * , z ˜ i - 1 } .
It remains to calculate the joint probabilities in (B.21); see the following.
P r { w i 1 , w i 2 , w i 1 k i * , z ˜ i - 1 , z ˜ i } = ( a ) P r { w i 1 } · P r { w i 2 , w i 1 k i * , z ˜ i } · P r { z ˜ i - 1 } ,
where (a) is from the fact that W i 1 is independent of W i 2 , W i 1 K i * , Z ˜ i and Z ˜ i - 1 and Z ˜ i - 1 is independent of W i 2 , W i 1 K i * , Z ˜ i .
Similarly, we have:
P r { w i 2 , w i 1 k i * , z ˜ i - 1 , z ˜ i } = ( b ) P r { w i 2 , w i 1 k i * , z ˜ i } · P r { z ˜ i - 1 } ,
where (b) is from the fact that Z ˜ i - 1 is independent of W i 2 , W i 1 K i * and Z ˜ i .
Substituting (B.22) and (B.23) into the left-hand side of (B.21), we have:
P r { w i 1 , w i 2 , w i 1 k i * , z ˜ i - 1 , z ˜ i } P r { w i 2 , w i 1 k i * , z ˜ i - 1 , z ˜ i } = P r { w i 1 } .
Next, we need to calculate the right-hand side of (B.21); see the following.
P r { w i 1 , w i 2 , w i 1 k i * , z ˜ i - 1 } = ( c ) P r { w i 1 } · P r { w i 2 } · P r { w i 1 k i * } · P r { z ˜ i - 1 } ,
where (c) is from the fact that W i 1 , W i 2 , W i 1 K i * and Z ˜ i - 1 are independent.
The joint probability P r { w i 2 , w i 1 k i * , z ˜ i - 1 } is calculated by:
P r { w i 2 , w i 1 k i * , z ˜ i - 1 } = ( d ) P r { w i 2 } · P r { w i 1 k i * } · P r { z ˜ i - 1 } ,
where (d) is from the fact that W i 2 , W i 1 K i * and Z ˜ i - 1 are independent.
Substituting (B.25) and (B.26) into the right-hand side of (B.21), we have:
P r { w i 1 , w i 2 , w i 1 k i * , z ˜ i - 1 } P r { w i 2 , w i 1 k i * , z ˜ i - 1 } = P r { w i 1 } .
By checking (B.24) and (B.27), the Markov chain W i 1 ( W i 2 , W i 1 K i * , Z ˜ i - 1 ) Z ˜ i is proven.  ☐
Proof. Proof of W i ( Z ˜ i , Z ˜ i - 1 ) ( W i - 1 , Z ˜ i - 2 , Z ˜ i + 1 n ) for Case 2:
Letting W i 2 = and W i 1 = W i for all 2 i n , the proof of W i ( Z ˜ i , Z ˜ i - 1 ) ( W i - 1 , Z ˜ i - 2 , Z ˜ i + 1 n ) for Case 2 is along the lines of that for Case 1, and therefore, we omit it here.  ☐
Proof. Proof of W i ( W i K i * , Z ˜ i - 1 ) Z ˜ i for Case 2:
Letting W i 2 = and W i 1 = W i for all 2 i n , the proof of W i ( W i K i * , Z ˜ i - 1 ) Z ˜ i for Case 2 is along the lines of that for Case 1, and therefore, we omit it here.  ☐
Thus, the direct proof of Theorem 1 is completed.

References

  1. Ahlswede, R.; Cai, N. Transmission, Identification and Common Randomness Capacities for Wire-Tap Channels with Secure Feedback from the Decoder. In General Theory of Information Transfer and Combinatorics; Springer-Verlag: Berlin/Heidelberg, Germany, 2006; Volume 4123, pp. 258–275. [Google Scholar]
  2. Dai, B.; Vinck, A.J.H.; Luo, Y.; Zhuang, Z. Capacity region of non-degraded wiretap channel with noiseless feedback. In Proceedings of 2012 IEEE International Symposium on Information Theory, Cambridge, MA, USA, 1–6 July 2012.
  3. Wyner, A.D. The wire-tap channel. Bell Syst. Tech. J. 1975, 54, 1355–1387. [Google Scholar] [CrossRef]
  4. Ardestanizadeh, E.; Franceschetti, M.; Javidi, T.; Kim, Y. Wiretap channel with secure rate-limited feedback. IEEE Trans. Inf. Theory 2009, 55, 5353–5361. [Google Scholar] [CrossRef]
  5. Lai, L.; El Gamal, H.; Poor, H.V. The wiretap channel with feedback: Encryption over the channel. IEEE Trans. Inf. Theory 2008, 54, 5059–5067. [Google Scholar] [CrossRef]
  6. He, X.; Yener, A. The role of feedback in two-way secure communication. IEEE Trans. Inf. Theory 2013, 59, 8115–8130. [Google Scholar] [CrossRef]
  7. Bassi, G.; Piantanida, P.; Shamai, S. On the capacity of the wiretap channel with generalized feedback. In Processdings of 2015 IEEE International Symposium on Information Theory, Hong Kong, China, 14–19 June 2015.
  8. Mitrpant, C.; Vinck, A.J.H.; Luo, Y. An achievable region for the gaussian wiretap channel with side information. IEEE Trans. Inf. Theory 2006, 52, 2181–2190. [Google Scholar] [CrossRef]
  9. El-Halabi, M.; Liu, T.; Georghiades, C.N.; Shamai, S. Secret writing on dirty paper: A deterministic view. IEEE Trans. Inf. Theory 2012, 58, 3419–3429. [Google Scholar] [CrossRef]
  10. Chen, Y.; Vinck, A.J.H. Wiretap channel with side information. IEEE Trans. Inf. Theory 2008, 54, 395–402. [Google Scholar] [CrossRef]
  11. Gel’fand, S.I.; Pinsker, M.S. Coding for channel with random parameters. Probl. Control Inf. Theory 1980, 9, 19–31. [Google Scholar]
  12. Dai, B.; Luo, Y. Some new results on wiretap channel with side information. Entropy 2012, 14, 1671–1702. [Google Scholar] [CrossRef]
  13. Chia, Y.K.; El Gamal, A. Wiretap channel with causal state information. IEEE Trans. Inf. Theory 2012, 58, 2838–2849. [Google Scholar] [CrossRef]
  14. Liu, T.; Mukherjee, P.; Ulukus, S.; Lin, S.; Hong, Y.W.P. Secure degrees of freedom of MIMO Rayleigh block fading wiretap channels with no CSI anywhere. IEEE Trans. Wireless Commun. 2015, 14, 2655–2669. [Google Scholar] [CrossRef]
  15. Csiszár, I.; Körner, J. Information Theory: Coding Theorems for Discrete Memoryless Systems; Academic Press: New York, NY, USA, 1981; pp. 123–124. [Google Scholar]
  16. Csiszár, I.; Körner, J. Broadcast channels with confidential messages. IEEE Trans. Inf. Theory 1978, 24, 339–348. [Google Scholar]
  17. Gamal, A.E.; Kim, Y.H. Network Information Theory; Cambridge University Press: Cambridge, UK, 2011. [Google Scholar]
  18. Costa, M.H.M. Writing on dirty paper. IEEE Trans. Inf. Theory 1983, 29, 439–441. [Google Scholar] [CrossRef]
  19. Shannon, C.E. A mathematical theory of communication. Bell Syst. Tech. J. 1948, 27, 379–423. [Google Scholar] [CrossRef]
  20. Leung-Yan-Cheong, S.; Hellman, M.E. The Gaussian wire-tap channel. IEEE Trans. Inf. Theory 1978, 24, 451–456. [Google Scholar] [CrossRef]

Share and Cite

MDPI and ACS Style

Dai, B.; Ma, Z.; Yu, L. Feeding Back the Output or Sharing the State: Which Is Better for the State-Dependent Wiretap Channel? Entropy 2015, 17, 7900-7925. https://doi.org/10.3390/e17127852

AMA Style

Dai B, Ma Z, Yu L. Feeding Back the Output or Sharing the State: Which Is Better for the State-Dependent Wiretap Channel? Entropy. 2015; 17(12):7900-7925. https://doi.org/10.3390/e17127852

Chicago/Turabian Style

Dai, Bin, Zheng Ma, and Linman Yu. 2015. "Feeding Back the Output or Sharing the State: Which Is Better for the State-Dependent Wiretap Channel?" Entropy 17, no. 12: 7900-7925. https://doi.org/10.3390/e17127852

APA Style

Dai, B., Ma, Z., & Yu, L. (2015). Feeding Back the Output or Sharing the State: Which Is Better for the State-Dependent Wiretap Channel? Entropy, 17(12), 7900-7925. https://doi.org/10.3390/e17127852

Article Metrics

Back to TopTop