Next Article in Journal
Study on the Stability of an Artificial Stock Option Market Based on Bidirectional Conduction
Previous Article in Journal
An Unified Approach to Limits on Power Generation and Power Consumption in Thermo-Electro-Chemical Systems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Capacity Region of a New Bus Communication Model

1
School of Information Science and Technology, Southwest JiaoTong University, Northbound Section Second Ring Road 111, Chengdu, 610031, China
2
Computer Science and Engineering Department, Shanghai Jiao Tong University, Dongchuan road 800, Shanghai, 200240, China
3
Institute for Experimental Mathematics, Duisburg-Essen University, Ellernstr.29, Essen, 45326, Germany
*
Author to whom correspondence should be addressed.
Entropy 2013, 15(2), 678-697; https://doi.org/10.3390/e15020678
Submission received: 11 November 2012 / Revised: 17 November 2012 / Accepted: 7 February 2013 / Published: 18 February 2013

Abstract

:
In this paper, we study a new bus communication model, where two transmitters wish to send their corresponding private messages and a common message to a destination, while they also wish to send the common message to another receiver connected to the same wire. From an information-theoretical point of view, we first study a general case of this new model (with discrete memoryless channels). The capacity region composed of all achievable ( R 0 , R 1 , R 2 ) triples is determined for this general model, where R 1 and R 2 are the transmission rates of the private messages and R 0 is the transmission rate of the common message. Then, the result is further explained via the Gaussian example. Finally, we give the capacity region for the new bus communication model with additive Gaussian noises and attenuation factors. This new bus communication model captures various communication scenarios, such as the bus systems in vehicles, and the bus type of communication channel in power line communication (PLC) networks.

1. Introduction

The bus communication model has been widely studied for many years. It captures various communication scenarios, such as the bus systems in vehicles, and the bus type of communication channel in power line communication (PLC) networks (see [1,2,3,4,5]).
Let us consider the bus communication model of Figure 1 from an information-theoretical point of view. Figure 1 can be equivalent to the model of the broadcast channel (see Figure 2). Note that Figure 2 implies that f 1 = f 2 = f .
Figure 1. The bus communication model (Gains f 1 and f 2 , the power of the sender σ x 2 , and the power of the noise σ n 2 ).
Figure 1. The bus communication model (Gains f 1 and f 2 , the power of the sender σ x 2 , and the power of the noise σ n 2 ).
Entropy 15 00678 g001
Figure 2. The broadcast presentation of the bus communication model.
Figure 2. The broadcast presentation of the bus communication model.
Entropy 15 00678 g002
The model of the boradcast channel was first investigated by Cover [6], and the capacity region of the general case (two private messages and one common message) is still not known. After the publication of Cover’s work, Körner and Marton [7] studied the broadcast channel with degraded message set (one private and one common message), and found its capacity region. For the degraded broadcast channel, the capacity region is totally determined, (see [8,9,10]). In addition, Gamal and Cover [11] showed that the Gaussian broadcast channel is a kind of degraded broadcast channel, and therefore, the capacity region for the Gaussian case can be directly obtained from the result of the degraded broadcast channel.
The following Theorem 1 shows the capacity region of the model of Figure 2, which is a kind of Gaussian broadcast channel.
Theorem 1 
The capacity region of the model of Figure 2 is the set of rate pairs ( R 1 , R 2 ) , such that
R 2 1 2 ln ( 1 + α σ x 2 f 2 σ n 2 ) R 1 1 2 ln ( 1 + ( 1 - α ) σ x 2 f 4 α σ x 2 f 4 + σ n 2 )
for some α [ 0 , 1 ] . Note that σ x 2 is the power constraint of the channel input, X, f and f 2 are channel gains ( 0 < f 1 ) and σ n 2 is the power of the noise.
Theorem 1 is directly obtained from the capacity region of the Gaussian broadcast channel [11], and therefore, the proof is omitted here.
In this paper, we study a two-sender bus communication model (see Figure 3). Two transmitters wish to send their corresponding private messages and a common message to receiver 1, while they also wish to send the common message to receiver 2.
Figure 3. A new bus communication model with two transmitters.
Figure 3. A new bus communication model with two transmitters.
Entropy 15 00678 g003
Figure 3 can be equivalent to the following Figure 4. Note that the capacity region of the Gaussian broadcast channel (BC) can be obtained from the capacity region of the discrete memoryless degraded broadcast channel. Therefore, first, we study the discrete memoryless case of the model of Figure 4, where two transmitters wish to send their private messages, W 1 and W 2 , and a common message, W 0 to receiver 1, and meanwhile, they also wish to send the common message, W 0 to receiver 2. Receiver 2 can receive a degraded version of the output of the multiple-access channel (MAC) via a discrete memoryless channel (DMC) (see Figure 5). This model can be viewed as a combination of multiple-access channel and degraded broadcast channel. For convenience, we call it MAC-DBC in this paper.
Figure 4. An equivalent model for the model of Figure 3.
Figure 4. An equivalent model for the model of Figure 3.
Entropy 15 00678 g004
Then, we study the model of Figure 4, which is a Gaussian example of the MAC-DBC in Figure 5. The capacity regions of the MAC-DBC and the model of Figure 4 are totally determined.
The study of MAC-DBC from an information-theoretical point of view is due to the fact that the network information theory has recently become an active research area. Both MAC and BC play an important role in the network information theory, and they have been extensively studied separately. However, the cascade of MAC and BC (MAC-BC) has seldom drawn people’s attention. To investigate the capacity region and the capacity-achieving coding scheme for the MAC-BC is the motivation of this work.
Figure 5. A combination of multiple-access channel and degraded broadcast channel (MAC-DBC).
Figure 5. A combination of multiple-access channel and degraded broadcast channel (MAC-DBC).
Entropy 15 00678 g005
In this paper, random variab1es, sample values and alphabets are denoted by capital letters, lower case letters and calligraphic letters, respectively. A similar convention is applied to the random vectors and their sample values. For example, U N denotes a random N-vector ( U 1 , . . . , U N ) , and u N = ( u 1 , . . . , u N ) is a specific vector value in U N that is the Nth Cartesian power of U . U i N denotes a random N - i + 1 -vector ( U i , . . . , U N ) , and u i N = ( u i , . . . , u N ) is a specific vector value in U i N . Let p V ( v ) denote the probability mass function P r { V = v } . Throughout the paper, the logarithmic function is to the base 2.
The remainder of this paper is organized as follows. In Section 2, we present the basic definitions and the main result on the capacity region of MAC-DBC. In Section 3, we provide the capacity region of the model of Figure 4. Final conclusions are presented in Section 4. The proofs are provided from Section A to Section E.

2. Notations, Definitions and the Main Results of MAC-DBC

In this section, a description of the MAC-DBC is given by Definition 1 to Definition 3. The capacity region, R ( A ) , composed of all achievable ( R 0 , R 1 , R 2 ) triples is given in Theorem 2, where the achievable ( R 0 , R 1 , R 2 ) triple is defined in Definition 4.
Definition 1 
(Encoders) The private messages, W 1 and W 2 , take values in W 1 and W 2 , respectively. The common message, W 0 , takes values in W 0 . W 0 , W 1 and W 2 are independent and uniformly distributed over their ranges. The channel encoders are two mappings:
f 1 N : W 0 × W 1 X 1 N
where f 1 N ( w 0 , w 1 ) = x 1 N X 1 N , w 1 W 1 , and w 0 W 0 .
f 2 N : W 0 × W 2 X 2 N
where f 2 N ( w 0 , w 2 ) = x 2 N X 2 N , w 2 W 2 , and w 0 W 0 . Note that W 1 and X 2 N are independent, and W 2 is independent of X 1 N .
The transmission rates of the private messages and the common message are log W 1 N , log W 2 N , and log W 0 N , respectively.
Definition 2 
(Channels) The MAC is a DMC with finite input alphabet X 1 × X 2 , finite output alphabet Y and transition probability Q 1 ( y | x 1 , x 2 ) , where x 1 X 1 , x 2 X 2 , y Y . Q 1 ( y N | x 1 N , x 2 N ) = n = 1 N Q 1 ( y n | x 1 , n , x 2 , n ) . The inputs of the MAC are X 1 N and X 2 N , while the output is Y N .
Receiver 2 has access to the output of the MAC via a discrete memoryless channel (DMC). The input of this DMC is Y N , and the output is Z N . The transition probability satisfies that
p Z N | Y N ( z N | y N ) = i = 1 N p Z | Y ( z i | y i )
where z i Z and y i Y .
Definition 3 
(Decoders) The decoder for receiver 1 is a mapping, f D 1 : Y N W 1 × W 2 × W 0 , with input Y N and outputs W ^ 1 , W ^ 2 , W ^ 0 . Let P e 1 be the error probability of the receiver 1 , and it is defined as P r { ( W 0 , W 1 , W 2 ) ( W ^ 0 , W ^ 1 , W ^ 2 ) } .
The decoder for receiver 2 is a mapping, f D 2 : Z N W 0 , with input Z N and output W ^ 0 . Let P e 2 be the error probability of the receiver 2 , and it is defined as P r { W 0 W ^ 0 } .
Definition 4 
(Achievable ( R 0 , R 1 , R 2 ) triple in the model of Figure 5) A triple ( R 0 , R 1 , R 2 ) (where R 0 , R 1 , R 2 > 0 ) is called achievable if, for any ϵ > 0 (where ϵ is an arbitrary small positive real number and ϵ 0 ), there exists channel encoders-decoders ( N , P e 1 , P e 2 ) , such that
log W 0 N R 0 - ϵ , log W 1 N R 1 - ϵ , log W 2 N R 2 - ϵ , P e 1 ϵ , P e 2 ϵ
Theorem 2 gives a single-letter characterization of the set R ( A ) , which is composed of all achievable ( R 0 , R 1 , R 2 ) triples in the model of Figure 5, and it is proved in Section A and Section B.
Theorem 2 
A single-letter characterization of the region R ( A ) is as follows,
R ( A ) = ( R 0 , R 1 , R 2 ) : 0 R 0 I ( U ; Z ) 0 R 1 I ( X 1 ; Y | X 2 , U ) 0 R 2 I ( X 2 ; Y | X 1 , U ) R 0 + R 1 I ( X 1 , X 2 ; Y | V 2 ) R 0 + R 2 I ( X 1 , X 2 ; Y | V 1 ) R 1 + R 2 I ( X 1 , X 2 ; Y | U ) R 0 + R 1 + R 2 I ( X 1 , X 2 ; Y ) .
where ( U , V 1 , V 2 ) ( X 1 , X 2 ) Y Z .
Remark 1 
There are some notes on Theorem 2; see the following.
  • The region R ( A ) is convex, and the proof is in Section C.
  • The ranges of the random variables U, V 1 and V 2 satisfy
    U X 1 X 2 + 2
    V 1 X 1 X 2
    V 2 X 1 X 2
    The proof is in Section D.
  • The auxiliary random variables U, V 1 and V 2 , in fact, are corresponding to W 0 , W 1 and W 2 , respectively.
  • If R 0 = 0 , receiver 2 is useless, and the model of Figure 5 reduces to the multiple-access channel (MAC).
    Let R 0 = 0 , and the corresponding U = c o n s t , the region, R ( A ) , reduces to
    ( R 1 , R 2 ) : 0 R 1 I ( X 1 ; Y | X 2 ) 0 R 2 I ( X 2 ; Y | X 1 ) R 1 I ( X 1 , X 2 ; Y | V 2 ) R 2 I ( X 1 , X 2 ; Y | V 1 ) R 1 + R 2 I ( X 1 , X 2 ; Y ) .
    Note that the Markov chains, V 1 X 1 Y and V 2 X 2 Y , then the inequalities R 1 I ( X 1 , X 2 ; Y | V 2 ) and R 2 I ( X 1 , X 2 ; Y | V 1 ) are included in R 1 I ( X 1 ; Y | X 2 ) and R 2 I ( X 2 ; Y | X 1 ) , respectively. Therefore, the above region (5) is simplified as
    ( R 1 , R 2 ) : 0 R 1 I ( X 1 ; Y | X 2 ) 0 R 2 I ( X 2 ; Y | X 1 ) R 1 + R 2 I ( X 1 , X 2 ; Y ) .
    , and this is coincident with the capacity region of the MAC [12,13].

3. A Gaussian Example of MAC-DBC and the Capacity Region of the Model of Figure 4

In this section, we first study a Gaussian example of Figure 5, where the channel input-output relationships at each time instant i ( 1 i N ) are given by
Y i = X 1 , i + X 2 , i + Z 1 , i
and
Z i = X 1 , i + X 2 , i + Z 1 , i + Z 2 , i
where Z 1 , i N ( 0 , σ n , 1 2 ) and Z 2 , i N ( 0 , σ n , 2 2 ) . The random vectors, Z N , Z 1 N and Z 2 N , are independent with i.i.d. components. The channel inputs, X 1 N and X 2 N , are subject to the average power constraints, σ x , 1 2 and σ x , 2 2 , respectively, i.e.,
1 N i = 1 N E [ X 1 , i 2 ] σ x , 1 2 , 1 N i = 1 N E [ X 2 , i 2 ] σ x , 2 2
Theorem 3 
For the Gaussian case of Figure 5, the capacity-equivocation region, R ( B ) , is given by
R ( B ) = 0 α 1 ( R 0 , R 1 , R 2 ) : 0 R 0 1 2 log ( 1 + α ( σ x , 1 2 + σ x , 2 2 ) ( 1 - α ) ( σ x , 1 2 + σ x , 2 2 ) + σ n , 1 2 + σ n , 2 2 ) 0 R 1 1 2 log ( 1 + ( 1 - α ) σ x , 1 2 σ n , 1 2 ) 0 R 2 1 2 log ( 1 + ( 1 - α ) σ x , 2 2 σ n , 1 2 ) R 1 + R 2 1 2 log ( 1 + ( 1 - α ) ( σ x , 1 2 + σ x , 2 2 ) σ n , 1 2 ) R 0 + R 1 1 2 log ( 1 + σ x , 1 2 + α σ x , 2 2 σ n , 1 2 ) R 0 + R 2 1 2 log ( 1 + σ x , 2 2 + α σ x , 1 2 σ n , 1 2 ) R 0 + R 1 + R 2 1 2 log ( 1 + σ x , 1 2 + σ x , 2 2 σ n , 1 2 ) .
The proof of Theorem 3 is in Section E.
Then, we will show that the capacity region of the model of Figure 4 can be obtained from the above Theorem 3. The channel input-output relationships of Figure 4 at each time instant i ( 1 i N ) are given by
Y i = f X 1 , i + f X 2 , i + Z 1 , i
and
Z i = f 2 X 1 , i + f 2 X 2 , i + Z 2 , i
, where Z 1 , i N ( 0 , σ n , 1 2 ) , Z 2 , i N ( 0 , σ n , 2 2 ) and σ n , 1 2 σ n , 2 2 . The channel inputs X 1 N and X 2 N , are subject to the average power constraints σ x , 1 2 and σ x , 2 2 , respectively.
Note that the additive Gaussian noise, Z 2 , i , can be viewed as a cascade of Z 1 , i and Z 2 , i , where Z 2 , i N ( 0 , σ n , 2 2 - σ n , 1 2 ) . Moreover, Equations (11) and (12) are equivalent to Equations (13) and (14), respectively, where
Y i = X 1 , i + X 2 , i + 1 f Z 1 , i
and
Z i = X 1 , i + X 2 , i + 1 f 2 Z 2 , i
Therefore, the model of Figure 4 is analogous to the above Gaussian example of MAC-DBC. The capacity region is as follows.
Theorem 4 
For the model of Figure 4, the capacity region R ( C ) is given by
R ( C ) = 0 α 1 ( R 0 , R 1 , R 2 ) : 0 R 0 1 2 log ( 1 + α f 4 ( σ x , 1 2 + σ x , 2 2 ) ( 1 - α ) ( σ x , 1 2 + σ x , 2 2 ) f 4 + σ n , 2 2 ) 0 R 1 1 2 log ( 1 + ( 1 - α ) σ x , 1 2 f 2 σ n , 1 2 ) 0 R 2 1 2 log ( 1 + ( 1 - α ) σ x , 2 2 f 2 σ n , 1 2 ) R 1 + R 2 1 2 log ( 1 + ( 1 - α ) ( σ x , 1 2 + σ x , 2 2 ) f 2 σ n , 1 2 ) R 0 + R 1 1 2 log ( 1 + ( σ x , 1 2 + α σ x , 2 2 ) f 2 σ n , 1 2 ) R 0 + R 2 1 2 log ( 1 + ( σ x , 2 2 + α σ x , 1 2 ) f 2 σ n , 1 2 ) R 0 + R 1 + R 2 1 2 log ( 1 + ( σ x , 1 2 + σ x , 2 2 ) f 2 σ n , 1 2 ) .
The proof is directly obtained from Theorem 3, and it is omitted here.
Figure 6, Figure 7, Figure 8, Figure 9 and Figure 10 plot the capacity region, R ( C ) , with different values of f, σ x , 1 2 , σ x , 2 2 , σ n , 1 2 and σ n , 2 2 . It is easy to see that R ( C ) reduces to the capacity region of the Gaussian MAC when R 0 = 0 . From these figures, we can see that R ( C ) enlarges as σ n , 2 1 and σ n , 2 2 decrease. Moreover, for fixed f, σ n , 1 2 and σ n , 2 2 , R ( C ) enlarges as σ x , 1 2 and σ x , 2 2 increase.
Figure 6. The capacity region of Figure 4 with σ x , 1 2 = σ x , 2 2 = 1 , σ n , 1 2 = 1 , σ n , 2 2 = 2 and f = 1 .
Figure 6. The capacity region of Figure 4 with σ x , 1 2 = σ x , 2 2 = 1 , σ n , 1 2 = 1 , σ n , 2 2 = 2 and f = 1 .
Entropy 15 00678 g006
Figure 7. The capacity region of Figure 4 with σ x , 1 2 = σ x , 2 2 = 1 , σ n , 1 2 = 1 , σ n , 2 2 = 100 and f = 1 .
Figure 7. The capacity region of Figure 4 with σ x , 1 2 = σ x , 2 2 = 1 , σ n , 1 2 = 1 , σ n , 2 2 = 100 and f = 1 .
Entropy 15 00678 g007
Figure 8. The capacity region of Figure 4 with σ x , 1 2 = σ x , 2 2 = 1 , σ n , 1 2 = 1 , σ n , 2 2 = 0 . 01 and f = 1 .
Figure 8. The capacity region of Figure 4 with σ x , 1 2 = σ x , 2 2 = 1 , σ n , 1 2 = 1 , σ n , 2 2 = 0 . 01 and f = 1 .
Entropy 15 00678 g008
Figure 9. The capacity region of Figure 4 with σ x , 1 2 = σ x , 2 2 = 2 , σ n , 1 2 = 0 . 1 , σ n , 2 2 = 0 . 01 and f = 1 .
Figure 9. The capacity region of Figure 4 with σ x , 1 2 = σ x , 2 2 = 2 , σ n , 1 2 = 0 . 1 , σ n , 2 2 = 0 . 01 and f = 1 .
Entropy 15 00678 g009
Figure 10. The capacity region of Figure 4 with σ x , 1 2 = σ x , 2 2 = 1 , σ n , 1 2 = 1 , σ n , 2 2 = 2 and f = 0 . 5 .
Figure 10. The capacity region of Figure 4 with σ x , 1 2 = σ x , 2 2 = 1 , σ n , 1 2 = 1 , σ n , 2 2 = 2 and f = 0 . 5 .
Entropy 15 00678 g010

4. Conclusions

In this paper, we first study the discrete memoryless MAC-DBC (the model of Figure 5). The capacity region is totally determined for this new model. Then, we study the model of Figure 4 and show that the capacity region of Figure 4 can be directly obtained from the Gaussian example of the MAC-DBC.

Acknowledgement

The authors would like to thank N. Cai for his help to improve this paper. This work was supported by a sub-project in National Basic Research Program of China under Grant 2012CB316100 on Broadband Mobile Communications at High Speeds, the National Natural Science Foundation of China under Grant 61271222, and the Research Fund for the Doctoral Program of Higher Education of China (No. 20100073110016).

References

  1. Lampe, L.; Schober, R.; Yiu, S. Distributed space-time block coding for multihop transmission in power line communication networks. IEEE J. Select. Areas Commun. 2006, 24, 1389–1400. [Google Scholar] [CrossRef]
  2. Bumiller, G.; Lampe, L.; Hrasnica, H. Power line communications for large-scale control and automation systems. IEEE Commun. Mag. 2010, 48, 106–113. [Google Scholar] [CrossRef]
  3. Lampe, L.; Han Vinck, A.J. On cooperative coding for narrow band PLC networks. Int. J. Electron. Commun. (AEÜ) 2011, 8, 681–687. [Google Scholar] [CrossRef]
  4. Balakirsky, V.; Vinck, A. Potential performance of PLC systems composed of several communication links. Intl. Symp. Power Line Commun. Appl. Vancouver 2005, 10, 12–16. [Google Scholar]
  5. Johansson, M.; Nielsen, L. Vehicle applications of controller area network. Tech. Rep. 2003, 7, 27–52. [Google Scholar]
  6. Cover, T.M. Broadcast channels. IEEE Trans. Inf. Theory 1972, 18, 2–14. [Google Scholar] [CrossRef]
  7. Körner, J.; Marton, K. General broadcast channels with degraded message sets. IEEE Trans. Inf. Theory 1977, 23, 60–64. [Google Scholar] [CrossRef]
  8. Gallager, R.G. Capacity and coding for degraded broadcast channels. Probl. Inf. Transm. 1974, 10, 3–14. [Google Scholar]
  9. Bergmans, P.P. A simple converse for broadcast channels with additive white Gaussian noise. IEEE Trans. Inf. Theory 1974, 20, 279–280. [Google Scholar] [CrossRef]
  10. Bergmans, P.P. Random coding theorem for broadcast channels with degraded components. IEEE Trans. Inf. Theory 1973, 19, 197–207. [Google Scholar] [CrossRef]
  11. El Gamal, A.; Cover, T.M. Multiple user information theory. Proc. IEEE 1980, 68, 1466–1483. [Google Scholar] [CrossRef]
  12. Ahlswede, R. Multiway Communication Channels. In Proceedings of the 2nd International Symposium Information Theory, Tsahkadsor, Armenia, /hlconference date, day and month 1971; pp. 23–52.
  13. Liao, H.H.J. Multiple access channels. Ph.D. Thesis, University of Hawaii, Honolulu, 1972. [Google Scholar]
  14. Csisza´r, I.; Körner, J. Information Theory. Coding Theorems for Discrete Memoryless Systems; Academic: London, UK, 1981. [Google Scholar]
  15. El Gamal, A.; Kim, Y. Lecture notes on network information theory. Available online: http://arxiv.org/abs/1001.3404 (accessed on 20 January 2010).
  16. Yeung, R.W. Information Theory and Network Coding; Springer: New York, NY, USA, 1981. [Google Scholar]
  17. Liang, Y.; Poor, H.V. Multiple-access channels with confidential messages. IEEE Trans. Inf. Theory 2008, 54, 976–1002. [Google Scholar] [CrossRef]

Appendix

A. Proof of the Converse Part of Theorem 2

In this section, we establish the converse part of Theorem 2: all the achievable ( R 0 , R 1 , R 2 ) triples are contained in the set R ( A ) , i.e., for any achievable triple, there exists random variables U, V 1 , V 2 , X 1 , X 2 , Y and Z such that the inequalities in Theorem 2 hold, and ( U , V 1 , V 2 ) ( X 1 , X 2 ) Y ( Y 1 , Y 2 ) forms a Markov chain. We will prove the inequalities of Theorem 2 in the remainder of this section.
(Proof of 0 R 0 I ( U ; Z ) ) The proof of this inequality is as follows:
1 N H ( W 0 ) ( a ) 1 N ( I ( W 0 ; Z N ) + δ ( P e 2 ) ) = 1 N ( H ( Z N ) - H ( Z N | W 0 ) + δ ( P e 2 ) ) = 1 N ( i = 1 N ( H ( Z i | Z i - 1 ) - H ( Z i | W 0 , Z i - 1 ) + δ ( P e 2 ) ) 1 N ( i = 1 N ( H ( Z i | Z i - 1 ) - H ( Z i | W 0 , Z i - 1 , Y i - 1 ) + δ ( P e 2 ) ) = ( b ) 1 N ( i = 1 N ( H ( Z i | Z i - 1 ) - H ( Z i | W 0 , Y i - 1 ) + δ ( P e 2 ) ) 1 N ( i = 1 N ( H ( Z i ) - H ( Z i | W 0 , Y i - 1 ) + δ ( P e 2 ) ) = ( c ) 1 N ( i = 1 N ( H ( Z i ) - H ( Z i | U i ) + δ ( P e 2 ) ) = ( d ) 1 N ( i = 1 N ( H ( Z i | J = i ) - H ( Z i | U i , J = i ) + δ ( P e 2 ) ) = ( e ) H ( Z J | J ) - H ( Z J | U J , J ) + 1 N δ ( P e 2 ) H ( Z J ) - H ( Z J | U J , J ) + 1 N δ ( P e 2 ) = ( f ) H ( Z ) - H ( Z | U ) + 1 N δ ( P e 2 ) = I ( U ; Z ) + δ ( P e 2 ) N
where (a) is from the Fano’s inequality, (b) is from Z i - 1 ( Y i - 1 , W 0 ) Z i , (c) is from the definition that U i ( W 0 , Y i - 1 ) , (d) is from J as a random variable (uniformly distributed over { 1 , 2 , . . . , N } ), and J is independent of Z i and U i , (e) is from J as uniformly distributed over { 1 , 2 , . . . , N } , and (f) is from the definitions that Z Z J and U ( U J , J ) .
Letting ϵ 0 and note that P e 2 ϵ , H ( W 0 ) N R 0 - ϵ , it is easy to see that 0 R 0 I ( U ; Z ) .
(Proof of 0 R 1 I ( X 1 ; Y | X 2 , U ) ) The proof of this inequality is as follows:
1 N H ( W 1 ) ( 1 ) 1 N ( H ( W 1 | X 2 N , W 0 ) + δ ( P e 1 ) - H ( W 1 | Y N , X 2 N , W 0 ) ) = 1 N ( I ( W 1 ; Y N | X 2 N , W 0 ) + δ ( P e 1 ) ) = 1 N ( H ( Y N | X 2 N , W 0 ) - H ( Y N | X 2 N , W 0 , W 1 ) + δ ( P e 1 ) ) 1 N ( H ( Y N | X 2 N , W 0 ) - H ( Y N | X 2 N , W 0 , W 1 , X 1 N ) + δ ( P e 1 ) ) = ( 2 ) 1 N ( H ( Y N | X 2 N , W 0 ) - H ( Y N | X 2 N , X 1 N ) + δ ( P e 1 ) ) = ( 3 ) 1 N ( i = 1 N ( H ( Y i | X 2 N , W 0 , Y i - 1 ) - H ( Y i | X 1 , i , X 2 , i ) ) + δ ( P e 1 ) ) 1 N ( i = 1 N ( H ( Y i | X 2 , i , W 0 , Y i - 1 ) - H ( Y i | X 1 , i , X 2 , i ) ) + δ ( P e 1 ) ) = ( 4 ) 1 N ( i = 1 N ( H ( Y i | X 2 , i , W 0 , Y i - 1 ) - H ( Y i | X 1 , i , X 2 , i , W 0 , Y i - 1 ) ) + δ ( P e 1 ) ) = ( 5 ) H ( Y | X 2 , U ) - H ( Y | X 1 , X 2 , U ) + 1 N δ ( P e 1 ) = I ( X 1 ; Y | X 2 , U ) + 1 N δ ( P e 1 )
where (1) is from the Fano’s inequality and the fact that W 1 is independent of X 2 N and W 0 ; (2) is from ( W 0 , W 1 ) ( X 2 N , X 1 N ) Y N ; (3) is from the discrete memoryless property of the channel; (4) is from ( W 0 , Y i - 1 ) ( X 1 , i , X 2 , i ) Y i ; and (5) is from the definitions that X 1 X 1 , J , X 2 X 2 , J , Y Y J , U ( W 0 , Y J - 1 , J ) , where J is a random variable (uniformly distributed over { 1 , 2 , . . . , N } ), and is independent of X 1 , i , X 2 , i , W 0 , Y i - 1 and Y i .
Letting ϵ 0 and note that P e 1 ϵ , H ( W 1 ) N R 1 - ϵ , it is easy to see that 0 R 1 I ( X 1 ; Y | X 2 , U ) .
(Proof of 0 R 2 I ( X 2 ; Y | X 1 , U ) ) The proof is similar to the proof of 0 R 1 I ( X 1 ; Y | X 2 , U ) , and it is omitted here.
(Proof of 0 R 1 + R 2 I ( X 1 , X 2 ; Y | U ) )
1 N H ( W 1 , W 2 ) ( a ) 1 N ( I ( W 1 , W 2 ; Y N | W 0 ) + δ ( P e 1 ) ) = 1 N ( H ( Y N | W 0 ) - H ( Y N | W 0 , W 1 , W 2 ) + δ ( P e 1 ) ) 1 N ( H ( Y N | W 0 ) - H ( Y N | W 0 , W 1 , W 2 , X 1 N , X 2 N ) + δ ( P e 1 ) ) = ( b ) 1 N ( H ( Y N | W 0 ) - H ( Y N | X 1 N , X 2 N ) + δ ( P e 1 ) ) = ( c ) 1 N ( i = 1 N ( H ( Y i | W 0 , Y i - 1 ) - H ( Y i | X 1 , i , X 2 , i ) ) + δ ( P e 1 ) ) = ( d ) 1 N ( i = 1 N ( H ( Y i | W 0 , Y i - 1 ) - H ( Y i | X 1 , i , X 2 , i , W 0 , Y i - 1 ) ) + δ ( P e 1 ) ) = ( e ) H ( Y | U ) - H ( Y | X 1 , X 2 , U ) + 1 N δ ( P e 1 ) ) = I ( X 1 , X 2 ; Y | U ) + δ ( P e 1 ) N
where (a) is from the Fano’s inequality and the fact that W 0 is independent of W 1 and W 2 ; (b) is from ( W 0 , W 1 , W 2 ) ( X 1 N , X 2 N ) Y N ; (c) is from the discrete memoryless property of the channel; (d) is from ( W 0 , Y i - 1 ) ( X 1 , i , X 2 , i ) Y i ; and (e) is from the definitions that X 1 ( X 1 , J , J ) , X 2 ( X 2 , J , J ) , Y Y J and U ( W 0 , Y J - 1 , J ) .
Letting ϵ 0 and note that P e 1 ϵ , H ( W 1 ) N R 1 - ϵ , H ( W 2 ) N R 2 - ϵ and W 1 is independent of W 2 , it is easy to see that 0 R 1 + R 2 I ( X 1 , X 2 ; Y | U ) .
(Proof of R 0 + R 1 I ( X 1 , X 2 ; Y | V 2 ) ) The proof is obtained by the following Equation (A4).
1 N H ( W 0 , W 1 ) ( a ) 1 N ( I ( W 0 , W 1 ; Y N | W 2 ) + δ ( P e 1 ) ) = 1 N ( H ( Y N | W 2 ) - H ( Y N | W 0 , W 1 , W 2 ) + δ ( P e 1 ) ) 1 N ( H ( Y N | W 2 ) - H ( Y N | W 0 , W 1 , W 2 , X 1 N , X 2 N ) + δ ( P e 1 ) ) = ( b ) 1 N ( H ( Y N | W 2 ) - H ( Y N | X 1 N , X 2 N ) + δ ( P e 1 ) ) = ( c ) 1 N ( i = 1 N ( H ( Y i | W 2 , Y i - 1 ) - H ( Y i | X 1 , i , X 2 , i ) ) + δ ( P e 1 ) ) = ( d ) 1 N ( i = 1 N ( H ( Y i | W 2 , Y i - 1 ) - H ( Y i | X 1 , i , X 2 , i , W 2 , Y i - 1 ) ) + δ ( P e 1 ) ) = ( e ) H ( Y | V 2 ) - H ( Y | X 1 , X 2 , V 2 ) + 1 N δ ( P e 1 ) ) = I ( X 1 , X 2 ; Y | V 2 ) + δ ( P e 1 ) N
where (a) is from the Fano’s inequality and the fact that W 2 is independent of W 0 and W 1 ; (b) is from ( W 0 , W 1 , W 2 ) ( X 1 N , X 2 N ) Y N ; (c) is from the discrete memoryless property of the channel; (d) is from ( W 2 , Y i - 1 ) ( X 1 , i , X 2 , i ) Y i ; and (e) is from the definitions that X 1 ( X 1 , J , J ) , X 2 ( X 2 , J , J ) , Y Y J and V 2 ( W 2 , Y J - 1 , J ) .
Letting ϵ 0 and note that P e 1 ϵ , H ( W 1 ) N R 1 - ϵ , H ( W 0 ) N R 0 - ϵ and W 0 is independent of W 1 , it is easy to see that R 0 + R 1 I ( X 1 , X 2 ; Y | V 2 ) .
(Proof of R 0 + R 2 I ( X 1 , X 2 ; Y | V 1 ) ) The proof is analogous to the proof of R 0 + R 1 I ( X 1 , X 2 ; Y | V 2 ) , and it is omitted here.
(Proof of R 0 + R 1 + R 2 I ( X 1 , X 2 ; Y ) )
1 N H ( W 0 , W 1 , W 2 ) ( a ) 1 N ( I ( W 0 , W 1 , W 2 ; Y N ) + δ ( P e 1 ) ) = 1 N ( H ( Y N ) - H ( Y N | W 0 , W 1 , W 2 ) + δ ( P e 1 ) ) 1 N ( H ( Y N ) - H ( Y N | W 0 , W 1 , W 2 , X 1 N , X 2 N ) + δ ( P e 1 ) ) = ( b ) 1 N ( H ( Y N ) - H ( Y N | X 1 N , X 2 N ) + δ ( P e 1 ) ) = ( c ) 1 N ( i = 1 N ( H ( Y i | Y i - 1 ) - H ( Y i | X 1 , i , X 2 , i ) ) + δ ( P e 1 ) ) 1 N ( i = 1 N ( H ( Y i ) - H ( Y i | X 1 , i , X 2 , i ) ) + δ ( P e 1 ) ) ( d ) H ( Y ) - H ( Y | X 1 , X 2 , U ) + 1 N δ ( P e 1 ) ) = I ( X 1 , X 2 ; Y ) + δ ( P e 1 ) N
where (a) is from the Fano’s inequality; (b) is from ( W 0 , W 1 , W 2 ) ( X 1 N , X 2 N ) Y N ; (c) is from the discrete memoryless property of the channel; and (d) is from the definitions that X 1 ( X 1 , J , J ) , X 2 ( X 2 , J , J ) , Y Y J .
Letting ϵ 0 and note that P e 1 ϵ , H ( W 0 ) N R 0 - ϵ , H ( W 1 ) N R 1 - ϵ , H ( W 2 ) N R 2 - ϵ , it is easy to see that R 0 + R 1 + R 2 I ( X 1 , X 2 ; Y ) .
The Markov chain, ( U , V 1 , V 2 ) ( X 1 , X 2 ) Y Z , is directly obtained from the definitions U ( W 0 , Y J - 1 , J ) , V 1 ( W 1 , Y J - 1 , J ) , V 2 ( W 2 , Y J - 1 , J ) X 1 ( X 1 , J , J ) , X 2 ( X 2 , J , J ) , Y Y J and Z Z J .
The proof of the converse part of Theorem 2 is completed.

B. Proof of the Direct Part of Theorem 2

In this section, we establish the direct part of Theorem 2 (about existence). Suppose ( R 0 , R 1 , R 2 ) R A , we will show that ( R 0 , R 1 , R 2 ) is achievable.
The coding scheme for Theorem 2 is in the following Figure A1. Now, the remainder of this section is organized as follows. Some preliminaries about typical sequences are introduced in Subsection B.1. The construction of the code is introduced in Subsection B.2. For any given ϵ > 0 , the proofs of log W 1 N R 1 - ϵ , log W 2 N R 2 - ϵ , log W 0 N R 0 - ϵ , P e 1 ϵ and P e 2 ϵ are given in Subsection B.3.
Figure A1. Coding scheme for MAC-DBC.
Figure A1. Coding scheme for MAC-DBC.
Entropy 15 00678 g011
B.1. Preliminaries
  • Given a probability mass function, p V ( v ) , for any η > 0 , let T V N ( η ) be the strong typical set of all v N , such that | p V ( v ) - c v N ( v ) N | < η for all v V , where c v N ( v ) is the number of occurences of the letter v in the v N . We say that the sequences, v N T V N ( η ) , are V-typical.
  • Analogously, given a joint probability mass function, p V W ( v , w ) , for any η > 0 , let T V W N ( η ) be the joint strong typical set of all pairs ( v N , w N ) , such that | p V W ( v , w ) - c v N , w N ( v , w ) N | < η for all v V and w W , where c v N , w N ( v , w ) is the number of occurences of ( v , w ) in the pair of sequences ( v N , w N ) . We say that the pairs of sequences, ( v N , w N ) T V W N ( η ) are V W -typical.
  • Moreover, w N is called W | V -generated by v N iff v N is V- typical and ( v N , w N ) T V W N ( η ) . For any given v N T V N ( η ) , define T W | V N ( η ) = { w N : w N i s W | V - generated by v N } .
  • Lemma 1 For any v N T V N ( η ) ,
    2 - N ( H ( V ) + η * ) p V N ( v N ) 2 - N ( H ( V ) - η * )
    where η * 0 as η 0 .
B.2. Coding Construction
Given a triple ( R 0 , R 1 , R 2 ) , choose a joint probability mass function, p U , V 1 , V 2 , X 1 , X 2 , Y , Z ( u , v 1 , v 2 , x 1 , x 2 , y , z ) , such that
0 R 1 I ( X 1 ; Y | X 2 , U ) , 0 R 2 I ( X 2 ; Y | X 1 , U ) , R 1 + R 2 I ( X 1 , X 2 ; Y | U ) ,
R 0 I ( U ; Z ) , R 0 + R 1 I ( X 1 , X 2 ; Y | V 2 ) , R 0 + R 2 I ( X 1 , X 2 ; Y | V 1 ) ,
R 0 + R 1 + R 2 I ( X 1 , X 2 ; Y )
The message sets, W 0 , W 1 and W 2 , satisfy the following conditions:
1 N log W 0 = R 0 , 1 N log W 1 = R 1 1 N log W 2 = R 2
Code-book generation:
  • For a given w 0 W 0 , generate a corresponding u N ( w 0 ) i.i.d., according to the probability mass function p U ( u ) .
  • For a given w 1 W 1 , generate a corresponding v 1 N ( w 1 ) i.i.d., according to the probability mass function p V 1 ( v 1 ) .
  • For a given w 2 W 2 , generate a corresponding v 2 N ( w 2 ) i.i.d., according to the probability mass function p V 2 ( v 2 ) .
  • x 1 N ( w 0 , w 1 ) is generated according to a new discrete memoryless channel (DMC), with inputs v 1 N ( w 1 ) and u N ( w 0 ) , and output x 1 N ( w 0 , w 1 ) . The transition probability of this new DMC is p X 1 | V 1 , U ( x 1 | v 1 , u ) .
    Similarly, x 2 N ( w 0 , w 2 ) is generated according to a new discrete memoryless channel (DMC), with inputs v 2 N ( w 2 ) and u N ( w 0 ) , and output x 2 N ( w 0 , w 2 ) . The transition probability of this new DMC is p X 2 | V 2 , U ( x 2 | v 2 , u ) .
Decoding scheme:
  • (Receiver 1) Receiver 1 declares that messages, w ^ 0 , w ^ 1 and w ^ 2 , are sent if they are the unique messages, such that ( u N ( w ^ 0 ) , v 1 N ( w ^ 1 ) , v 2 N ( w ^ 2 ) , x 1 N ( w ^ 0 , w ^ 1 ) , x 2 N ( w ^ 0 , w ^ 2 ) , y N ) T U V 1 V 2 X 1 X 2 Y N ( ϵ ) , otherwise, it declares an error.
  • (Receiver 2) Receiver 2 declares that a message w ^ 0 is sent if it is the unique message, such that ( u N ( w ^ 0 ) , z N ) T U Z N ( ϵ ) ; otherwise it declares an error.
B.3. Achievability Proof
By using the above Equation (A6), it is easy to verify that log W 0 N R 0 - ϵ , log W 1 N R 1 - ϵ and log W 2 N R 2 - ϵ . It remains to show that P e 1 ϵ and P e 2 ϵ ; see the following.
Without loss of generality, assume that w 0 = 1 , w 1 = 1 and w 2 = 1 are sent.

2.3.1. P e 2 ϵ

For receiver 2, define the events:
B 1 = { ( u N ( 1 ) , z N ) T U Z N ( η ) }
B 2 = { ( u N ( w 0 ) , z N ) T U Z N ( η ) } for some w 0 1
The probability of error for receiver 2 is then upper bounded by
P e 2 = P r { B 1 B 2 } P r { B 1 } + P r { B 2 }
By using LLN, the first term P r { B 1 } 0 as N . On the other hand, by using the packing lemma [15, p. 53-54], P r { B 2 } 0 as N if R 0 I ( U ; Z ) .
Therefore, by choosing sufficiently large N, we have P e 2 ϵ .

2.3.2. P e 1 ϵ

The proof of P e 1 ϵ is as follows.
Define the sets
A 1 = { ( u N ( 1 ) , v 1 N ( 1 ) , v 2 N ( 1 ) , x 1 N ( 1 , 1 ) , x 2 N ( 1 , 1 ) , y N ) T U V 1 V 2 X 1 X 2 Y N ( ϵ ) }
A 2 = { ( u N ( 1 ) , v 1 N ( 1 ) , v 2 N ( w 2 ) , x 1 N ( 1 , 1 ) , x 2 N ( 1 , w 2 ) , y N ) T U V 1 V 2 X 1 X 2 Y N ( ϵ ) } for some w 2 1
A 3 = { ( u N ( 1 ) , v 1 N ( w 1 ) , v 2 N ( 1 ) , x 1 N ( 1 , w 1 ) , x 2 N ( 1 , 1 ) , y N ) T U V 1 V 2 X 1 X 2 Y N ( ϵ ) } for some w 1 1
A 4 = { ( u N ( 1 ) , v 1 N ( w 1 ) , v 2 N ( w 2 ) , x 1 N ( 1 , w 1 ) , x 2 N ( 1 , w 2 ) , y N ) T U V 1 V 2 X 1 X 2 Y N ( ϵ ) } for some w 1 1 , w 2 1
A 5 = { ( u N ( w 0 ) , v 1 N ( 1 ) , v 2 N ( 1 ) , x 1 N ( w 0 , 1 ) , x 2 N ( w 0 , 1 ) , y N ) T U V 1 V 2 X 1 X 2 Y N ( ϵ ) } for some w 0 1
A 6 = { ( u N ( w 0 ) , v 1 N ( 1 ) , v 2 N ( w 2 ) , x 1 N ( w 0 , 1 ) , x 2 N ( w 0 , w 2 ) , y N ) T U V 1 V 2 X 1 X 2 Y N ( ϵ ) } for some w 0 1 , w 2 1
A 7 = { ( u N ( w 0 ) , v 1 N ( w 1 ) , v 2 N ( 1 ) , x 1 N ( w 0 , w 1 ) , x 2 N ( w 0 , 1 ) , y N ) T U V 1 V 2 X 1 X 2 Y N ( ϵ ) } for some w 0 1 , w 1 1
A 8 = { ( u N ( w 0 ) , v 1 N ( w 1 ) , v 2 N ( w 2 ) , x 1 N ( w 0 , w 1 ) , x 2 N ( w 0 , w 2 ) , y N ) T U V 1 V 2 X 1 X 2 Y N ( ϵ ) } for some ( w 0 , w 1 , w 2 ) 1
The probability of error for receiver 1 is then upper bounded by:
P e 2 = P r { A 1 A 2 A 3 A 4 A 5 A 6 A 7 A 8 } (A8) i = 1 8 P r { A i }
By using LLN, the first term, P r { A 1 } 0 as N .
For the second term, by using the packing lemma ([15][p. 53–54]), P r { A 2 } 0 as N if
R 2 I ( V 2 , X 2 ; Y | U , V 1 , X 1 ) (A9) = ( 1 ) H ( Y | U , X 1 ) - H ( Y | X 1 , X 2 , U ) = I ( X 2 ; Y | U , X 1 )
where (1) is from V 1 ( U , X 1 ) Y and ( V 1 , V 2 ) ( X 1 , X 2 , U ) Y .
Analogously, for the third term, by using the packing lemma, P r { A 3 } 0 as N if R 1 I ( X 1 ; Y | U , X 2 ) .
For the fourth term, by using the packing lemma, P r { A 4 } 0 as N if R 1 + R 2 I ( X 1 ; Y | U , X 2 ) .
R 1 + R 2 I ( V 1 , V 2 , X 1 , X 2 ; Y | U ) (A10) = ( 2 ) H ( Y | U ) - H ( Y | X 1 , X 2 , U ) = I ( X 1 , X 2 ; Y | U )
where (2) is from ( V 1 , V 2 ) ( X 1 , X 2 , U ) Y .
For the fifth term, P r { A 5 } 0 as N if R 0 I ( U ; Y ) .
For the sixth term, by using the packing lemma, P r { A 6 } 0 as N if
R 0 + R 2 I ( U , V 2 , X 1 , X 2 ; Y | V 1 ) (A11) = ( 3 ) H ( Y | V 1 ) - H ( Y | X 1 , X 2 , V 1 ) = I ( X 1 , X 2 ; Y | V 1 )
where (3) is from ( U , V 2 ) ( X 1 , X 2 , V 1 ) Y .
Analogously, the seventh term, P r { A 7 } 0 as N if R 0 + R 1 I ( X 1 , X 2 ; Y | V 2 ) .
For the eighth term, by using the packing lemma, P r { A 8 } 0 as N if R 0 + R 1 + R 2 I ( X 1 , X 2 ; Y ) .
Therefore, by choosing sufficiently large N, we have P e 1 ϵ .
The proof of the direct part of Theorem 2 is completed.

C. Proof of the Convexity of R A

Let ( R 0 , R 1 , R 2 ) R ( A ) , i.e., ( R 0 , R 1 , R 2 ) , satisfy the following conditions:
R 1 I ( X 1 ( 1 ) ; Y ( 1 ) | X 2 ( 1 ) , U ( 1 ) ) , R 2 I ( X 2 ( 1 ) ; Y ( 1 ) | X 1 ( 1 ) , U ( 1 ) ) , R 1 + R 2 I ( X 1 ( 1 ) , X 2 ( 1 ) ; Y ( 1 ) | U ( 1 ) ) ,
0 R 0 I ( U ( 1 ) ; Z ( 1 ) ) , R 0 + R 1 I ( X 1 ( 1 ) , X 2 ( 1 ) ; Y ( 1 ) | V 2 ( 1 ) ) , R 0 + R 2 I ( X 1 ( 1 ) , X 2 ( 1 ) ; Y ( 1 ) | V 1 ( 1 ) ) ,
R 0 + R 1 + R 2 I ( X 1 ( 1 ) , X 2 ( 1 ) ; Y ( 1 ) )
Let ( R 0 , R 1 , R 2 ) R ( A ) , i.e., ( R 0 , R 1 , R 2 ) , satisfy the following conditions:
R 1 I ( X 1 ( 2 ) ; Y ( 2 ) | X 2 ( 2 ) , U ( 2 ) ) , R 2 I ( X 2 ( 2 ) ; Y ( 2 ) | X 1 ( 2 ) , U ( 2 ) ) , R 1 + R 2 I ( X 1 ( 2 ) , X 2 ( 2 ) ; Y ( 2 ) | U ( 2 ) )
0 R 0 I ( U ( 2 ) ; Z ( 2 ) ) , R 0 + R 1 I ( X 1 ( 2 ) , X 2 ( 2 ) ; Y ( 2 ) | V 2 ( 2 ) ) , R 0 + R 2 I ( X 1 ( 2 ) , X 2 ( 2 ) ; Y ( 2 ) | V 1 ( 2 ) )
R 0 + R 1 + R 2 I ( X 1 ( 2 ) , X 2 ( 2 ) ; Y ( 2 ) )
Let Q be a switch function, such that P r { Q = 1 } = θ and P r { Q = 2 } = 1 - θ , where 0 θ 1 . Q is independent of all the random variables.
Define V 1 = V 1 ( Q ) Q , V 2 = V 2 ( Q ) Q , U = U ( Q ) Q , X 1 = X 1 ( Q ) Q , X 2 = X 2 ( Q ) Q , Y = Y ( Q ) , Z = Z ( Q ) . Then we have
I ( X 1 ; Y | X 2 , U ) = I ( X 1 ( Q ) ; Y ( Q ) | X 2 ( Q ) , U ( Q ) , Q ) = θ I ( X 1 ( 1 ) ; Y ( 1 ) | X 2 ( 1 ) , U ( 1 ) , Q = 1 ) + ( 1 - θ ) I ( X 1 ( 2 ) ; Y ( 2 ) | X 2 ( 2 ) , U ( 2 ) , Q = 2 ) (A12) = θ I ( X 1 ( 1 ) ; Y ( 1 ) | X 2 ( 1 ) , U ( 1 ) ) + ( 1 - θ ) I ( X 1 ( 2 ) ; Y ( 2 ) | X 2 ( 2 ) , U ( 2 ) )
I ( X 2 ; Y | X 1 , U ) = I ( X 2 ( Q ) ; Y ( Q ) | X 1 ( Q ) , U ( Q ) , Q ) = θ I ( X 2 ( 1 ) ; Y ( 1 ) | X 1 ( 1 ) , U ( 1 ) , Q = 1 ) + ( 1 - θ ) I ( X 2 ( 2 ) ; Y ( 2 ) | X 1 ( 2 ) , U ( 2 ) , Q = 2 ) (A13) = θ I ( X 2 ( 1 ) ; Y ( 1 ) | X 1 ( 1 ) , U ( 1 ) ) + ( 1 - θ ) I ( X 2 ( 2 ) ; Y ( 2 ) | X 1 ( 2 ) , U ( 2 ) )
I ( X 1 , X 2 ; Y ) I ( X 1 ( Q ) , X 2 ( Q ) ; Y ( Q ) | Q ) = θ I ( X 1 ( 1 ) , X 2 ( 1 ) ; Y ( 1 ) | Q = 1 ) + ( 1 - θ ) I ( X 1 ( 2 ) , X 2 ( 2 ) ; Y ( 2 ) | Q = 2 ) (A14) = θ I ( X 1 ( 1 ) , X 2 ( 1 ) ; Y ( 1 ) ) + ( 1 - θ ) I ( X 1 ( 2 ) , X 2 ( 2 ) ; Y ( 2 ) )
I ( X 1 , X 2 ; Y | U ) = I ( X 1 ( Q ) , X 2 ( Q ) ; Y ( Q ) | U ( Q ) , Q ) = θ I ( X 1 ( 1 ) , X 2 ( 1 ) ; Y ( 1 ) | U ( 1 ) , Q = 1 ) + ( 1 - θ ) I ( X 1 ( 2 ) , X 2 ( 2 ) ; Y ( 2 ) | U ( 2 ) , Q = 2 ) (A15) = θ I ( X 1 ( 1 ) , X 2 ( 1 ) ; Y ( 1 ) | U ( 1 ) ) + ( 1 - θ ) I ( X 1 ( 2 ) , X 2 ( 2 ) ; Y ( 2 ) | U ( 2 ) )
I ( X 1 , X 2 ; Y | V 1 ) = I ( X 1 ( Q ) , X 2 ( Q ) ; Y ( Q ) | V 1 ( Q ) , Q ) = θ I ( X 1 ( 1 ) , X 2 ( 1 ) ; Y ( 1 ) | V 1 ( 1 ) , Q = 1 ) + ( 1 - θ ) I ( X 1 ( 2 ) , X 2 ( 2 ) ; Y ( 2 ) | V 1 ( 2 ) , Q = 2 ) (A16) = θ I ( X 1 ( 1 ) , X 2 ( 1 ) ; Y ( 1 ) | V 1 ( 1 ) ) + ( 1 - θ ) I ( X 1 ( 2 ) , X 2 ( 2 ) ; Y ( 2 ) | V 1 ( 2 ) )
I ( X 1 , X 2 ; Y | V 2 ) = I ( X 1 ( Q ) , X 2 ( Q ) ; Y ( Q ) | V 2 ( Q ) , Q ) = θ I ( X 1 ( 1 ) , X 2 ( 1 ) ; Y ( 1 ) | V 2 ( 1 ) , Q = 1 ) + ( 1 - θ ) I ( X 1 ( 2 ) , X 2 ( 2 ) ; Y ( 2 ) | V 2 ( 2 ) , Q = 2 ) (A17) = θ I ( X 1 ( 1 ) , X 2 ( 1 ) ; Y ( 1 ) | V 2 ( 1 ) ) + ( 1 - θ ) I ( X 1 ( 2 ) , X 2 ( 2 ) ; Y ( 2 ) | V 2 ( 2 ) )
I ( U ( Q ) ; Z ( Q ) ) I ( U ( Q ) ; Z ( Q ) | Q ) = θ I ( U ( 1 ) ; Z ( 1 ) | Q = 1 ) + ( 1 - θ ) I ( U ( 2 ) ; Z ( 2 ) | Q = 2 ) (A18) = θ I ( U ( 1 ) ; Z ( 1 ) ) + ( 1 - θ ) I ( U ( 2 ) ; Z ( 2 ) )
From Equations (A12)–(A18), it is easy to see that ( θ R 1 + ( 1 - θ ) R 1 , θ R 2 + ( 1 - θ ) R 2 , θ R 0 + ( 1 - θ ) R 0 ) R ( A ) , and therefore, R ( A ) is convex.

D. Size Constraints of the Auxiliary Random Variables in Theorem 2

By using the support lemma (see [14], p.310), it suffices to show that the random variables U, A and K can be replaced by new ones, preserving the Markovity ( U , V 1 , V 2 ) ( X 1 , X 2 ) Y Z and the characters I ( U ; Z ) , I ( X 2 ; Y | X 1 , U ) , I ( X 1 ; Y | X 2 , U ) , I ( X 1 , X 2 ; Y | U ) , I ( X 1 , X 2 ; Y | V 2 ) , I ( X 1 , X 2 ; Y | V 1 ) , and furthermore, the range of the new U, A and K satisfies:
U X 1 X 2 + 2
V 1 X 1 X 2
V 2 X 1 X 2
The proof of which is in the reminder of this section.
Let
p ¯ = p X 1 X 2 ( x 1 , x 2 )
Define the following continuous scalar functions of p ¯ :
f X 1 X 2 ( p ¯ ) = p X 1 X 2 ( x 1 , x 2 ) , f Y ( p ¯ ) = H ( Y ) , f Y | X 1 ( p ¯ ) = H ( Y | X 1 ) , f Y | X 2 ( p ¯ ) = H ( Y | X 2 )
Since there are X 1 X 2 - 1 functions of f X 1 X 2 ( p ¯ ) , the total number of the continuous scalar functions of p ¯ is X 1 X 2 +2.
Let p ¯ X 1 X 2 | U = P r { X 1 = x 1 , X 2 = x 2 | U = u } . With these distributions p ¯ X 1 X 2 | U , we have
p X 1 X 2 ( x 1 , x 2 ) = u U p ( U = u ) f X 1 X 2 ( p ¯ X 1 X 2 | U )
H ( Y | U ) = u U p ( U = u ) f Y ( p ¯ X 1 X 2 | U )
H ( Y | X 1 , U ) = u U p ( U = u ) f Y | X 1 ( p ¯ X 1 X 2 | U )
H ( Y | X 2 , U ) = u U p ( U = u ) f Y | X 2 ( p ¯ X 1 X 2 | U )
According to the support lemma ([14], p.310), the random variable, U, can be replaced by new ones, such that the new U takes at most X 1 X 2 + 2 different values and the expressions (A20)–(A23) are preserved.
Similarly, we can prove that V 1 X 1 X 2 and V 2 X 1 X 2 . The proof is omitted here.

E. Proof of Theorem 3

E.1. Proof of the Achievability

The achievability proof follows by computing the mutual information terms in Theorem 2 with the following joint distributions:
U N ( 0 , α ( P 1 + P 2 ) ) , V 1 N ( 0 , ( 1 - α ) P 1 ) and V 2 N ( 0 , ( 1 - α ) P 2 )
X 1 = P 1 P 1 + P 2 U + V 1 and X 2 = P 2 P 1 + P 2 U + V 2
U is independent of V 1 and V 2 .

E.2. Proof of the Converse

The proof of R 0 1 2 log ( 1 + α ( σ x , 1 2 + σ x , 2 2 ) ( 1 - α ) ( σ x , 1 2 + σ x , 2 2 ) + σ n , 1 2 + σ n , 2 2 ) and R 1 + R 2 1 2 log ( 1 + ( 1 - α ) ( σ x , 1 2 + σ x , 2 2 ) σ n , 1 2 ) are from the proof of the Gaussian broadcast channel [15], and it is omitted here.
The proof of R 1 1 2 log ( 1 + ( 1 - α ) σ x , 1 2 σ n , 1 2 ) , R 2 1 2 log ( 1 + ( 1 - α ) σ x , 2 2 σ n , 1 2 ) and R 0 + R 1 + R 2 1 2 log ( 1 + σ x , 1 2 + σ x , 2 2 σ n , 1 2 ) are from the proof of the Gaussian multiple-access channel [16], and it is omitted here.
Then, it remains to show that R 0 + R 1 1 2 log ( 1 + σ x , 1 2 + α σ x , 2 2 σ n , 1 2 ) and R 0 + R 2 1 2 log ( 1 + σ x , 2 2 + α σ x , 1 2 σ n , 1 2 ) . The proof of these two inequalities are analogous to ([17] [p. 1000–1001]), and therefore, we omit the proof here.
The proof of Theorem 3 is completed.

Share and Cite

MDPI and ACS Style

Dai, B.; Vinck, A.J.H.; Luo, Y.; Zhuang, Z. Capacity Region of a New Bus Communication Model. Entropy 2013, 15, 678-697. https://doi.org/10.3390/e15020678

AMA Style

Dai B, Vinck AJH, Luo Y, Zhuang Z. Capacity Region of a New Bus Communication Model. Entropy. 2013; 15(2):678-697. https://doi.org/10.3390/e15020678

Chicago/Turabian Style

Dai, Bin, A. J. Han Vinck, Yuan Luo, and Zhuojun Zhuang. 2013. "Capacity Region of a New Bus Communication Model" Entropy 15, no. 2: 678-697. https://doi.org/10.3390/e15020678

APA Style

Dai, B., Vinck, A. J. H., Luo, Y., & Zhuang, Z. (2013). Capacity Region of a New Bus Communication Model. Entropy, 15(2), 678-697. https://doi.org/10.3390/e15020678

Article Metrics

Back to TopTop