Next Article in Journal
A Novel Underwater Acoustic Target Identification Method Based on Spectral Characteristic Extraction via Modified Adaptive Chirp Mode Decomposition
Next Article in Special Issue
Utility–Privacy Trade-Offs with Limited Leakage for Encoder
Previous Article in Journal
Complexity of Recent Earthquake Swarms in Greece in Terms of Non-Extensive Statistical Physics
Previous Article in Special Issue
On the Asymptotic Capacity of Information-Theoretic Privacy-Preserving Epidemiological Data Collection
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

On Decoder Ties for the Binary Symmetric Channel with Arbitrarily Distributed Input

1
Department of Electrical Engineering, Yuan Ze University, Taoyuan 32003, Taiwan
2
Institute of Communications Engineering, National Yang-Ming Chiao-Tung University, Taipei 112304, Taiwan
3
Department of Mathematics and Statistics, Queen’s University, Kingston, ON K7L 3N6, Canada
*
Author to whom correspondence should be addressed.
Entropy 2023, 25(4), 668; https://doi.org/10.3390/e25040668
Submission received: 20 March 2023 / Revised: 13 April 2023 / Accepted: 15 April 2023 / Published: 16 April 2023
(This article belongs to the Special Issue Advances in Information and Coding Theory)

Abstract

:
The error probability of block codes sent under a non-uniform input distribution over the memoryless binary symmetric channel (BSC) and decoded via the maximum a posteriori (MAP) decoding rule is investigated. It is proved that the ratio of the probability of MAP decoder ties to the probability of error grows most linearly in blocklength when no MAP decoding ties occur, thus showing that decoder ties do not affect the code’s error exponent. This result generalizes a similar recent result shown for the case of block codes transmitted over the BSC under a uniform input distribution.

1. Introduction

Consider the classical channel coding context, where we send a block code through the memoryless binary symmetric channel (BSC) with crossover probability 0 < p < 1 / 2 . Given a sequence of binary codes { C n } n 1 with n being the blocklength, we denote the sequence of corresponding minimal probabilities of decoding error under maximum a posteriori (MAP) decoding by { a n } n 1 . The following result was recently shown in [1] when the channel input selects codewords from C n according to a uniform distribution.
Theorem 1
([1]). For any sequence of codes { C n } n 1 of blocklength n and size | C n | = M with C n { 0 , 1 } n , sent over the BSC with crossover probability 0 < p < 1 / 2 under a uniform channel input distribution over C n , its minimum probability of decoding error a n satisfies
b n a n 1 + ( 1 p ) p n b n ,
where
b n = P X n , Y n ( x n , y n ) X n × Y n : P X n | Y n ( x n | y n ) < max u n C n \ { x n } P X n | Y n ( u n | y n ) ,
where P X n , Y n is the joint input–output distribution that X n = ( X 1 , X 2 , , X n ) X n { 0 , 1 } n is sent over the BSC (via n uses) and Y n = ( Y 1 , Y 2 , , Y n ) Y n { 0 , 1 } n is received.
Noting that b n in (2) is the probability that a decoding error occurs without inducing decoder ties (which occur when two or more codewords in C n are identified by the decoder as the estimated transmitted codeword; i.e., when more than one codeword in C n maximize P X n | Y n ( · | y n ) for a given received word y n ), the above result in (1) directly implies that decoder ties do not affect the error exponent of a n . The error exponent or reliability function of a block coding communication system represents the largest rate of exponential decay of the system’s probability of decoding error as the coding blocklength grows to infinity (e.g., see [2,3,4,5,6,7,8,9,10,11,12,13,14]).
It is known that uniformly distributed data achieves the largest entropy rate and leaves no room for data compression. Thus, ideally compressed data should exhibit uniform distribution for all blocklengths n. However, this setting is often impractical due to the sub-optimality of the implemented data compression schemes. Instead, we generally have non-uniformly distributed data after compression in the form of residual redundancy such as in speech or image coding (e.g., [15,16]). Furthermore, one may have a compressed source that can be divided into several groups, within each of which the symbols are equally probable. Decoder ties can thus occur with respect to two (or more) codewords corresponding to symbols within the same group.
In this paper, we consider a non-uniform prior distribution over C n and prove that decoder ties, under optimal MAP decoding, still have linear and hence sub-exponential impact on the error probability a n , thus extending Theorem 1 established for the case of a uniform prior distribution over C n . Since our problem falls within the general framework of joint source-channel coding for point-to-point communication systems, we refer the reader to [14,15,16,17,18,19,20,21] (Section 4.6) and the references therein for theoretical studies on this subject as well as practical designs that outperform separate source and channel coding under complexity or delay constraints.
The proof technique used in [1] to show (1) above is based on the observation that there are two types of decoding errors. One is that the received tuple at the channel output induces no decoder ties but the corresponding decoder decision is wrong. The other is that the received tuple at the channel output causes a decoder tie, but the decoder picks the wrong codeword. As a result, the MAP error probability a n can be upper bounded by the sum of two terms, b n and δ n , where b n is the probability of the first type of decoding errors as given in (2), and δ n is the probability of decoder ties regardless of whether the tie breaker misses the correct codeword or not. Under the assumption that the channel input is uniformly distributed over block code C n for each blocklength n and an arbitrary sequence of codes { C n } n 1 , it was shown in [1] that flipping a properly selected bit component of the channel output that causes a decoder tie can produce a unique channel output that leads to the first type of decoding errors. An analysis of this bit-flipping manipulation shows that the ratio δ n / b n grows at most linearly in n and hence yields the upper bound in (1). However, this flipping technique no longer works when non-uniform channel inputs are considered. To tackle this problem, we judiciously separate the channel output tuples that induce decoder ties into two groups, one group consisting of output tuples that do not fulfill the above flipping manipulation property and the other group composed of the remaining output tuples (i.e., the complement group). We then show that the probability of the former group is upper bounded by that of the latter group, and therefore δ n / b n remains growing at most linearly in blocklength n under arbitrary channel input statistics. Note that the group that fails the flipping property is an empty set when channel input is uniformly distributed over C n , thereby making the result of Theorem 1 a special case of the extended result in this paper. The rest of the paper is organized as follows. Section 2 presents the main result and highlights the key steps of the proof to facilitate its understanding. The proof is then provided in full detail, along with illustrative examples, in Section 3 and Appendix A and Appendix B. Finally, conclusions and future directions are given in Section 4.
Throughout the paper, we denote [ M ] { 1 , 2 , , M } for positive integer M and set d ( x n , y n | S ) to be the Hamming distance between n-tuples x n = ( x 1 , x 2 , , x n ) and y n = ( y 1 , y 2 , , y n ) with the indices of the tuples restricted to S [ n ] . By convention, we set d ( x n , y n | S ) = 0 when S = and use d ( x n , y n ) to represent d ( x n , y n | [ n ] ) .

2. Main Result

Consider a binary code C n { 0 , 1 } n with fixed blocklength n and size M to be used over the memoryless BSC with crossover probability 0 < p < 1 2 . Denote the prior probability on C n by P X n and hence P X n ( C n ) = 1 . Without loss of generality, we assume that all codewords in C n occur with positive probability, i.e., P X n ( x n ) > 0 for all x n C n ; hence, C n is the support of P X n .
It is known the minimal probability of decoding error is achieved by the MAP decoder, which upon the reception of the channel output y n { 0 , 1 } n estimates the codeword x n C n according to
e ( y n ) = arg max u n C n P X n | Y n ( u n | y n ) ,
where P X n | Y n is the posterior conditional distribution of X n given Y n . We can see from (3) that if more than one u n C n achieves the maximum value of P X n | Y n ( u n | y n ) for a given y n , a decoder tie occurs, in which case the set of these u n , denoted conveniently as { e ( y n ) } , contains more than one element. As a result, an erroneous MAP decision is made if one of the two situations occurs: i ) the transmitted codeword does not belong to { e ( y n ) } ; i i ) the transmitted codeword belong to { e ( y n ) } and | { e ( y n ) } | > 1 , but the tie breaker picks the wrong one from { e ( y n ) } . By conveniently denoting
C n = { c 1 , c 2 , , c M } ,
the probability of the first situation acts as a lower bound b n for a n (i.e., b n a n ), where b n is given in (2) and can be written as
b n = i = 1 M P X n ( c i ) P Y n | X n y n { 0 , 1 } n : P X n | Y n ( c i | y n ) < max r [ M ] \ { i } P X n | Y n ( c r | y n ) .
It is shown in [22] that b n exactly equals the generalized Poor–Verdú (lower) bound [23,24] as its tilting parameter approaches infinity. The probability of the second situation is bounded above by the probability that the transmitted codeword belongs to { e ( y n ) } and | { e ( y n ) } | > 1 , disregarding whether the tie breaker picks the wrong codeword or not, and this upper bound can be expressed as
δ n i = 1 M P X n ( c i ) P Y n | X n y n { 0 , 1 } n : P X n | Y n ( c i | y n ) = max r [ M ] \ { i } P X n | Y n ( c r | y n ) .
We thus have
b n a n b n + δ n .
By proving the inequality
δ n 2 q n b n ,
where
q 1 p p > 1 ,
we have our main result as follows.
Theorem 2.
For any sequence of binary codes { C n } n 1 and prior probabilities { P X n } n 1 used over the BSC, we have
b n a n 1 + 2 q n b n .
Remark 1.
Theorem 2 implies that the relative deviation of a n from b n is at most linear in the blocklength n and the impact of decoder ties in (6) to a n is only sub-exponential. Consequently, a n and b n must have the same error exponent. Note also that the upper bound in (10) differs from the result in Theorem 1 by an additional multiplicative factor of 2 in the q n term. As explained in the introduction section, this is a consequence of the fact that the probability of the group of channel output tuples that cause decoder ties but fail the flipping manipulation property is upper bounded by that of the remaining tie-inducing channel outputs. The full technical details are provided in Section 3.2. Finally, we emphasize that Theorem 2 holds for arbitrary binary codes, including “bad” codes for which high probability codewords have small Hamming distance between them. Hence, tightening the upper bound in (10) by restricting the analysis for “sufficiently good" codes, in the sense that their most likely codewords sit “sufficiently” far apart in { 0 , 1 } n , is an interesting future direction.
List of Main Symbols: Before providing an overview of the main steps of the proof of Theorem 2 (which is presented in full detail in the next section), we describe in Table 1 the main symbols used in the paper and indicate the equation where they are first introduced. We emphasize that sets T j | i , N j | i and S 1 , j ( m ) are defined differently from their counterparts in [1] that use the same notation.
We also visually illustrate in Figure 1 some of the main sets defined in Table 1 under the setting of Example 1, which is presented in Section 3 below for a non-uniformly distributed binary code with M = 4 codewords and blocklength n = 4 given by C 4 = { c 1 , c 2 , c 3 , c 4 } = { 0000 , 0101 , 0110 , 0111 } . More specifically, we only show the non-empty component subsets in Y n = { 0 , 1 } 4 corresponding to codewords c 1 and c 2 ; refer to Table A2 in Appendix A for a detailed listing of all component subsets in { 0 , 1 } 4 (including empty ones).
Overview of the Proof: Given that codeword c i is sent over the channel, i [ M ] , let T i denote the set of output tuples y n that result in MAP decoding ties:
T i y n { 0 , 1 } n : P X n | Y n ( c i | y n ) = max r [ M ] \ { i } P X n | Y n ( c r | y n )
  = y n { 0 , 1 } n : P X n , Y n ( c i , y n ) = max r [ M ] \ { i } P X n , Y n ( c r , y n ) ,
where (12) holds because P X n | Y n ( x n | y n ) = P X n , Y n ( x n , y n ) P Y n ( y n ) . Then, δ n in (6) can be rewritten as
δ n = i [ M ] P X n ( c i ) P Y n | X n T i | c i = i [ M ] P X n , Y n ( c i , T i ) .
Similarly, let N i denote the set of output tuples y n which guarantee a tie-free MAP decoding error when c i is transmitted over the channel:
N i y n { 0 , 1 } n : P X n | Y n ( c i | y n ) < max r [ M ] \ { i } P X n | Y n ( c r | y n )
  = y n { 0 , 1 } n : P X n , Y n ( c i , y n ) < max r [ M ] \ { i } P X n , Y n ( c r , y n ) .
Hence, b n in (5) can be rewritten as:
b n = i [ M ] P X n ( c i ) P Y n | X n ( N i | c i ) = i [ M ] P X n , Y n ( c i , N i ) .
Note if δ n = 0 , then (7) is tight and (10) holds trivially; so, without loss of generality, we assume in the proof that δ n > 0 , which implies that there exists at least one non-empty T i for i [ M ] . Then, according to (13) and (16), we have that
δ n b n = i [ M ] P X n , Y n ( c i , T i ) i [ M ] P X n , Y n ( c i , N i ) i [ M ] : T i P X n , Y n ( c i , T i ) i [ M ] : T i P X n , Y n ( c i , N i ) .
We can upper-bound (17) by
i [ M ] : T i P X n , Y n ( c i , T i ) i [ M ] : T i P X n , Y n ( c i , N i ) max 1 [ M ] : T i P X n , Y n ( c i , T i ) P X n , Y n ( c i , N i ) ,
where for convenience we will refer to an inequality of the form given in (18) as the ratio-sum inequality. As a result, Theorem 2 holds if we can substantiate that 2 q n is an upper bound for (18). To this end, we will find a proper partition of T i and an equal number of disjoint subsets of N i , of which the individual probabilities can be evaluated. For ease of notation, we denote the individual probabilities corresponding to the K-partition of T i and K disjoint subsets of N i by { α k } k = 1 K and { β k } k = 1 K , respectively. Then, we obtain that
P X n , Y n ( c i , T i ) P X n , Y n ( c i , N i ) k = 1 K α k k = 1 K β k .
By showing that each individual ratio α k / β k , k [ K ] is bounded above by 2 q n , the ratio-sum inequality can again be applied to complete the proof.

3. Proof of Theorem 2

In [1], where a uniformly distributed prior probability P X n over C n is assumed, one can flip a properly selected bit in the output y n T i to convert it to a corresponding element in N i . In light of this connection, one can evaluate the ratio P Y n | X n ( T i | c i ) P Y n | X n ( N i | c i ) . This approach, however, no longer works when a non-uniformly distributed prior probability is considered. Therefore, we have to devise a more judicious approach to extend the result in [1] for a general prior probability.

3.1. A Partition of Non-Empty T i and Corresponding Disjoint Subsets of N i

In this section, instead of finding a disjoint covering of the set of decoder ties T i as in [1], we establish a proper partition of T i from Definitions 1 and 2. This is one of the key differences from the techniques used in [1]. Example 1 is given after Proposition 1 to illustrate Definitions 1 and 2.
Given y n T i defined in (12), there exists at least one m [ M ] \ { i } such that
P X n , Y n ( c i , y n ) = P X n , Y n ( c m , y n ) = max r [ M ] \ { i } P X n , Y n ( c r , y n ) .
We collect the indices m that satisfy (20) in I i ( y n ) as follows:
I i ( y n ) h [ M ] \ { i } : P X n , Y n ( c i , y n ) = P X n , Y n ( c h , y n ) = max r [ M ] \ { i } P X n , Y n ( c r , y n ) .
Remark 2.
First, we note that I i ( y n ) is not empty as long as y n T i . Also, for any y n T i , we can infer from (21) that h I i ( y n ) if and only if y n T h .
In Definitions 1 and 2 that follow, we will assign each y n T i to a subset indexed by j I i ( y n ) . These subsets will form a partition of T i as stated in Proposition 1.
Definition 1.
For j [ M ] \ { i } , denoting by S i , j the set of indices where the bit components of c i and c j differ, we define
{ (22a) T j | i y n T i : j = min r I i ( y n ) : d ( c i , y n | S i , r ) < | S i , r | r ; N j | i { y n N i : P X n , Y n ( c i , y n ) · q = P X n , Y n ( c j , y n ) · 1 q (22b) and P X n , Y n ( c j , y n ) P X n , Y n ( c r , y n ) for r [ j 1 ] \ { i } } .
Since there may exist y n T i satisfying d ( c i , y n | S i , r ) = | S i , r | for all r I i ( y n ) , the collection of all elements in j [ M ] \ { i } T j | i may not exhaust the elements in T i (see Example 1). We thus go on to collect the remaining elements in T i \ j [ M ] \ { i } T j | i as follows.
Definition 2.
Define for j [ M ] \ { i } ,
T j | i y n T i \ h [ M ] \ { i } T h | i : j = min r I i ( y n ) r .
With the sets defined in Definitions 1 and 2, a partition of T i and disjoint subsets of N i are constructed as proven in the following proposition.
Proposition 1.
For non-empty T i , the following two properties hold.
(i) 
The collection { T j | i T j | i } j [ M ] \ { i } forms a (disjoint) partition of T i .
(ii) 
{ N j | i } j [ M ] \ { i } is a collection of disjoint subsets of N i .
Before proving Proposition 1, we provide the following example to illustrate the above sets.
Example 1.
This example illustrates the necessity of introducing T j | i as a companion to T j | i . Suppose C 4 = { c 1 , c 2 , c 3 , c 4 } = { 0000 , 0101 , 0110 , 0111 } . Let P X 4 ( c 1 ) = q 2 2 + q 2 + q 2 , P X 4 ( c 2 ) = P X 4 ( c 3 ) = 1 2 + q 2 + q 2 and P X 4 ( c 4 ) = q 2 2 + q 2 + q 2 . Then, y 4 = 0111 satisfies
P X 4 , Y 4 ( c 1 , y 4 ) q 2 ( 2 + q 2 + q 2 ) p 4 q 1 = max r [ 4 ] \ { 1 } P X 4 , Y 4 ( c r , y 4 ) = P X 4 , Y 4 ( c 2 , y n ) 1 ( 2 + q 2 + q 2 ) p 4 q 3 = P X 4 , Y 4 ( c 3 , y 4 ) 1 ( 2 + q 2 + q 2 ) p 4 q 3 > P X 4 , Y 4 ( c 4 , y 4 ) q 2 ( 2 + q 2 + q 2 ) p 4 q 4 ,
where the probabilities P X n , Y n ( x n , y n ) are written in the form
P X n , Y n ( x n , y n ) = P X n ( x n ) P Y n | X n ( y n | x n ) = P X n ( x n ) p n q n d ( x n , y n ) .
Note that the first equality in (24) indicates 0111 T 1 and the last two equalities and the right-most inequality jointly imply I 1 ( 0111 ) = { 2 , 3 } . In light of Proposition 1, this 0111 must lie in one and only one of { T j | 1 T j | 1 } j [ 4 ] \ { 1 } as shown in Table A1 and Table A2 of Appendix A. Since there exist no integers h in I 1 ( 0111 ) fulfilling d ( c 1 , 0111 | S 1 , h ) < | S 1 , h | , this 0111 belongs to T j | 1 with j = min r I 1 ( 0111 ) = 2 . Recall that in [1], an element in N j | 1 can be obtained if flipping a zero of y n T j | 1 can make it further away from c 1 but closer to c j . However, for y 4 = 0111 in this example if we flip the only zero to one, it gets further away from both c 1 and c h for any h = 2 , 3 , 4 . Therefore, the bit-flipping manipulation fails.
With y 4 = 0111 , we also have
P X 4 , Y 4 ( c 2 , y 4 ) 1 ( 2 + q 2 + q 2 ) p 4 q 3 = max r [ 4 ] \ { 2 } P X 4 , Y 4 ( c r , y 4 ) = P X 4 , Y 4 ( c 1 , y 4 ) q 2 ( 2 + q 2 + q 2 ) p 4 q = P X 4 , Y 4 ( c 3 , y n ) 1 ( 2 + q 2 + q 2 ) p 4 q 3 > P X 4 , Y 4 ( c 4 , y 4 ) 1 ( 2 + q 2 + q 2 ) p 4 q 4 ,
where the first equality indicates 0111 T 2 and the remaining parts in (26) jointly imply that I 2 ( 0111 ) = { 1 , 3 } . Proposition 1 then states that this 0111 lies in one and only one of { T j | 2 T j | 2 } j [ 4 ] \ { 2 } . Since 1 I 2 ( 0111 ) = { 1 , 3 } and d ( c 2 , 0111 | S 2 , 1 ) = 0 < | S 2 , 1 | = 2 , we have 0111 T 1 | 2 according to (22a). Thus, we can flip a bit in 0111 to get further away from c 2 and closer to c j simultaneously. More specifically, the bit-flipping manipulation produces either 0110 or 0011, which lies in N 1 | 2 as y 4 = 0111 is in T 1 | 2 . Therefore, we can associate the element in T 1 | 2 with an element in N 1 | 2 via a single flipping operation. For completeness, a full list of the sets T i , N i , T j | i , T j | i and N j | i for i [ 4 ] and j [ 4 ] \ { i } , is given in Appendix A.
Proof of Proposition 1.
First, we note that by the definitions in (22a) and (23), { T j | i } j [ M ] \ { i } are disjoint and so is { T j | i } j [ M ] \ { i } . Additionally, (23) implies T j | i T h | i = for arbitrary j , h [ M ] \ { i } . Furthermore, according to Definitions 1 and 2, for any y n T i , we have either y n T h | i or y n T h | i for some h [ M ] \ { i } . Consequently, { T j | i T j | i } j [ M ] \ { i } forms a partition of T i .
On the other hand, the inequality in (22b) prevents multiple inclusions of an element from the previous collections. Therefore, { N j | i } j [ M ] \ { i } are a collection of disjoint subsets of N i . □
Remark 3.
When channel inputs are uniformly distributed as considered in [1], it follows that
I i ( y n ) = h [ M ] \ { i } : d ( c i , y n ) = d ( c h , y n ) = max r [ M ] \ { i } d ( c r , y n ) ,
and d ( c i , y n | S i , j ) = | S i , j | 2 < | S i , j | for every j I i ( y n ) . Therefore, (22a) is reduced to
T j | i = y n T i : j = min r I i ( y n ) r ,
and
T j | i = .
We then have the following two remarks. First, we note that the T j | i newly defined via (22a) and reduced to (28) in the regime considered in [1] is more restrictive than the T j | i introduced in [1] (Equation (16a)). As a consequence, { T j | i } j [ M ] \ { i } forms a partition of T i in this paper while those introduced in [1] (Equation (16a)) are a disjoint covering of T i under uniform channel inputs. Second, (29) shows that [1] does not need to consider a companion T j | i to T j | i , but this paper does.
Based on Proposition 1, we continue the derivation from (17) and obtain:
δ n b n i [ M ] P X n , Y n c i , j [ M ] \ { i } ( T j | i T j | i ) i [ M ] P X n , Y n c i , j [ M ] \ { i } N j | i
= i [ M ] j [ M ] \ { i } P X n , Y n c i , T j | i + i [ M ] j [ M ] \ { i } P X n , Y n c i , T j | i i [ M ] j [ M ] \ { i } P X n , Y n c i , N j | i ,
where (31) holds because { T j | i } j [ M ] \ { i } and { T j | i } j [ M ] \ { i } are disjoint, and the same applies to { N j | i } j [ M ] \ { i } . An additional upper bound for (31) requires the verification of the inequality:
i [ M ] j [ M ] \ { i } P X n , Y n c i , T j | i j [ M ] i [ M ] \ { j } P X n , Y n c j , T i | j ,
which is an immediate consequence of the proposition to be proven in the next section (Proposition 2), stating that
y n T j | i and h I i ( y n ) y n T | h for some I h ( y n ) and P X n , Y n ( c i , y n ) = P X n , Y n ( c h , y n ) .

3.2. Verification of (32)

Recall that the main technique used in [1] is to associate every element in T i with a corresponding element in N i via the bit-flipping manipulation. By this bit-flipping association, the probability ratio of the elements and corresponding elements respectively in T i and N i can be evaluated. However, as Example 1 indicates, for an element in T j | i , the bit-flipping association no longer works. This reveals the challenge of generalizing the results in [1] from uniform channel inputs to arbitrarily distributed channel inputs. A solution is to subdivide the elements in T i into two groups { T j | i } j [ M ] \ { i } and { T j | i } j [ M ] \ { i } , where the bit-flipping association to { N j | i } j [ M ] \ { i } works for the former group but not for the latter. The inequality in (32) can then be used to exclude the latter group with an upper bound:
δ n b n i [ M ] j [ M ] \ { i } P X n , Y n c i , T j | i + i [ M ] j [ M ] \ { i } P X n , Y n c i , T j | i i [ M ] j [ M ] \ { i } P X n , Y n c i , N j | i
2 i [ M ] j [ M ] \ { i } P X n , Y n c i , T j | i i [ M ] j [ M ] \ { i } P X n , Y n c i , N j | i .
Since uniform channel inputs as considered in [1] guarantee (29), it can be seen from (35) that the multiplicative factor of 2 can be reduced to 1 as observed in Remark 1. For general arbitrary channel inputs, we have the factor of 2 since the set T j | i may not be empty. The validity of (32) can be confirmed by the next proposition.
Proposition 2.
Suppose y n T j | i . Then, for every h I i ( y n ) , we have
y n T | h for some I h ( y n ) and P X n , Y n ( c i , y n ) = P X n , Y n ( c h , y n ) .
Proof. 
Suppose y n T j | i . Then, d ( c i , y n | S i , h ) = | S i , h | for every h I i ( y n ) . We therefore have:
P X n , Y n ( c i , y n ) = P X n , Y n ( c h , y n ) = max r [ M ] \ { i } P X n , Y n ( c r , y n ) .
We can rewrite (37) as
P X n , Y n ( c h , y n ) = P X n , Y n ( c i , y n ) = max r [ M ] \ { h } P X n , Y n ( c r , y n ) ,
implying y n T h and i I h ( y n ) . Noting that d ( c h , y n | S h , i ) = 0 < | S h , i | because d ( c i , y n | S i , h ) = | S i , h | and S h , i = S i , h , we conclude that the smallest integer I h ( y n ) satisfying d ( c h , y n | S h , ) < | S h , | exists, and therefore y n T | h . □
Remark 4.
Two observations can be made based on Proposition 2. First, Proposition 2 indicates that every y n T j | i must appear at least once in the sum h [ M ] [ M ] \ { h } P X n , Y n c h , T | h , contributing the same probability mass P X n , Y n ( c h , y n ) as P X n , Y n ( c i , y n ) . Second, Proposition 2 also implies that every y n T j | i cannot be contained in h [ M ] r [ M ] \ { h } T r | h \ T j | i . This observation can be substantiated as follows. For every h I i ( y n ) , Proposition 2 implies y n T | h for some I h ( y n ) and hence Definition 2 immediately gives y n T r | h for all r [ M ] \ { h } . For h I i ( y n ) , we have y n T h and therefore y n T r | h T h for all r [ M ] \ { h } as pointed out in Remark 2. As a result, every y n T j | i appears exactly once in the sum h [ M ] r [ M ] \ { h } P X n , Y n c h , T r | h . Combining the two observations leads to:
i [ M ] j [ M ] \ { i } y n T j | i P X n , Y n c i , y n j [ M ] [ M ] \ { j } y n T | j P X n , Y n c j , y n .
To flesh out the above inequality, we give the next example.
Example 2.
Proceeding from Example 1, we observe from Table A1 and Table A2 in Appendix A that 0111 is contained in T 2 | 1 , T 1 | 2 and T 1 | 3 . Hence, it appears once in the sum j [ 4 ] i [ 4 ] \ { j } P X n , Y n c j , T i | j while it contributes twice in the sum j [ 4 ] i [ 4 ] \ { j } P X n , Y n c j , T i | j . We then confirm from (A35) that:
i [ 4 ] j [ 4 ] \ { i } P X 4 , Y 4 c i , T j | i i [ 4 ] j [ 4 ] \ { i } P X 4 , Y 4 c i , T j | i .
We continue the derivation from (35) and obtain
δ n b n 2 i [ M ] j [ M ] \ { i } : T j | i P X n , Y n c i , T j | i i [ M ] j [ M ] \ { i } : T j | i P X n , Y n c i , N j | i
2 max i [ M ] and j [ M ] \ { i } : T j | i P X n , Y n c i , T j | i P X n , Y n c i , N j | i ,
where we add the restriction T j | i in (41) to exclude the cases of zero dividing by zero in (42), and (42) follows the ratio-sum inequality in (18).
In the next section, we introduce a number of delicate decompositions of non-empty T j | i and an equal number of disjoint subsets of N j | i to facilitate the bit-flipping association of the pairs.

3.3. Atomic Decomposition of Non-Empty T j | i and the Corresponding Disjoint Subsets of N j | i

To simplify the exposition, we assume without loss of generality that c 1 is the all-zero codeword (It is known that we can simultaneously flip the same position of all codewords to yield a new code of equal performance over the BSC. Thus, via a number of flipping manipulations, we can transform any code to a code of equal performance with the first codeword being all-zero.) Below we present the proof for i = 1 . The proof for general i > 1 follows annalagously.
Since c 1 is the all-zero codeword, S 1 , j is the set containing the indices of the non-zero components of c j . To facilitate the investigation of the structure of c j relative to the remaining codewords { c r } r [ M ] \ { 1 , j } , we first partition S 1 , j into 2 M 2 subsets according to whether each index in S 1 , j is in S 1 , 2 , , S 1 , j 1 , S 1 , j + 1 , , S 1 , M or not, as follows:
S 1 , j ( m ) r = 2 j 1 S r ; λ r r = j + 1 M S r ; λ r S 1 , j for m 1 + r = 2 j 1 λ r · 2 r 2 + r = j + 1 M λ r · 2 r 3 ,
where S r ; 1 S 1 , r and S r ; 0 [ n ] \ S 1 , r = S 1 , r c , and each λ r { 0 , 1 } . An example of the partition is given below.
Example 3.
Suppose C 4 = { 00000 , 11001 , 01111 , 01101 } . For j = 3 and S 1 , j = { 2 , 3 , 4 , 5 } , we obtain 2 4 2 = 4 subsets as
S 1 , 3 ( m ) = S 1 , 3 ( 1 ) = S 1 , 2 c S 1 , 4 c S 1 , 3 = { 4 } , if ( λ 4 , λ 2 ) = ( 0 , 0 ) ; S 1 , 3 ( 2 ) = S 1 , 2 S 1 , 4 c S 1 , 3 = , if ( λ 4 , λ 2 ) = ( 0 , 1 ) ; S 1 , 3 ( 3 ) = S 1 , 2 c S 1 , 4 S 1 , 3 = { 3 } , if ( λ 4 , λ 2 ) = ( 1 , 0 ) ; S 1 , 3 ( 4 ) = S 1 , 2 S 1 , 4 S 1 , 3 = { 2 , 5 } , if ( λ 4 , λ 2 ) = ( 1 , 1 ) .
As c 1 is the all-zero codeword, the components of c r with indices in S 1 , j ( m ) can now be unambiguously identified and must all equal λ r . As a result,
d c 1 , c r | S 1 , j ( m ) = | S 1 , j ( m ) | , λ r = 1 ; 0 , λ r = 0 .
Example 4.
Proceeding from Example 3, we have
d c 1 , c 2 | S 1 , 3 ( 1 ) = 0 because λ 2 = 0 ; d c 1 , c 2 | S 1 , 3 ( 2 ) = | S 1 , 3 ( 2 ) | = 0 because λ 2 = 1 ; d c 1 , c 2 | S 1 , 3 ( 3 ) = 0 because λ 2 = 0 ; d c 1 , c 2 | S 1 , 3 ( 4 ) = | S 1 , 3 ( 4 ) | = 2 because λ 2 = 1 ,
and
d c 1 , c 4 | S 1 , 3 ( 1 ) = 0 because λ 4 = 0 ; d c 1 , c 4 | S 1 , 3 ( 2 ) = 0 because λ 4 = 0 ; d c 1 , c 4 | S 1 , 3 ( 3 ) = | S 1 , 3 ( 3 ) | = 1 because λ 4 = 1 ; d c 1 , c 4 | S 1 , 3 ( 4 ) = | S 1 , 3 ( 4 ) | = 2 because λ 4 = 1 .
It should be emphasized that S 1 , j ( m ) in this paper is defined differently from that in [1]. While the one defined in [1] partitions S 1 , j only according to codewords with indices less than j, the one defined in this paper considers all other M 2 codewords in the partition manipulation, and hence the order of codewords becomes irrelevant.
Next, to decompose T j | 1 , we further define a sequence of incremental sets:
S 1 , j ( m ) h = 1 m S 1 , j ( h ) , m [ 2 M 2 ] ,
and set S 1 , j ( 0 ) . Let 1 , j | S 1 , j | and 1 , j ( m ) | S 1 , j ( m ) | respectively denote the sizes of S 1 , j and S 1 , j ( m ) and note that 0 = 1 , j ( 0 ) 1 , j ( 1 ) 1 , j ( 2 ) 1 , j ( 2 M 2 ) = 1 , j .
The idea behind the partition of T j | 1 into 1 , j subsets, indexed by k [ 1 , j 1 ] { 0 } , is as follows. Pick one y n T j | 1 . We start by examining whether d ( c 1 , y n | S 1 , j ( 1 ) ) is strickly less than 1 , j ( 1 ) 1 . If the answer is negative, we continue examining whether d ( c 1 , y n | S 1 , j ( 2 ) ) is strictly less than 1 , j ( 2 ) 1 . Proceed until we reach the smallest m such that d ( c 1 , y n | S 1 , j ( m ) ) < 1 , j ( m ) 1 holds. Setting k to be equal to k = d ( c 1 , y n | S 1 , j ( m ) ) , we assign this y n to the subset T j | 1 ( k ) . Notably, there exists no such number m [ 2 M 2 ] that satisfies d ( c 1 , y n | S 1 , j ( m ) ) < 1 , j ( m ) 1 if and only if d ( c 1 , y n | S 1 , j ) = 1 , j 1 ; in this case, we find the smallest m satisfying S 1 , j ( m ) = S 1 , j and assign this element to T j | 1 ( 1 , j 1 ) as d ( c 1 , y n | S 1 , j ( m ) ) = 1 , j 1 . For ease of describing the above algorithmic partition process, we introduce a mapping from k [ 1 , j 1 ] { 0 } to m [ 2 M 2 ] as follows:
η k min m [ 2 M 2 ] : k < 1 , j ( m ) 1 , 0 k < 1 , j 1 ; min m [ 2 M 2 ] : k = 1 , j ( m ) 1 , k = 1 , j 1 .
We can see that for 0 k < 1 , j 1 , we have 1 , j ( η k 1 ) 1 k < 1 , j ( η k ) 1 . Therefore, if y n T j | 1 is assigned to T j | 1 ( k ) for some k < 1 , j 1 , we must have
1 , j ( η k 1 ) 1 d c 1 , y n | S 1 , j ( η k 1 ) d c 1 , y n | S 1 , j ( η k ) = k < 1 , j ( η k ) 1 .
On the other hand, if y n T j | 1 is collected in T j | 1 ( 1 , j 1 ) , then S 1 , j ( η k ) = S 1 , j and
1 , j ( η k 1 ) 1 d c 1 , y n | S 1 , j ( η k 1 ) d c 1 , y n | S 1 , j ( η k ) = 1 , j 1 .
A formal definition of T j | 1 ( k ) is given next, where the corresponding subsets N j | 1 ( k ) of N j | 1 are also introduced.
Definition 3.
Define for k = 0 , 1, …, 1 , j 1 ,
{ (52a) T j | 1 ( k ) y n T j | 1 : 1 , j ( η k 1 ) 1 d c 1 , y n | S 1 , j ( η k 1 ) and d c 1 , y n | S 1 , j ( η k ) = k ; (52b) N j | 1 ( k ) y n N j | 1 : 1 , j ( η k 1 ) = d c 1 , y n | S 1 , j ( η k 1 ) and d c 1 , y n | S 1 , j ( η k ) = k + 1 ,
where η k is defined in (49).
With Definition 3, we have the following proposition.
Proposition 3.
For non-empty T j | 1 , the following two properties hold.
(i) 
{ T j | 1 ( k ) } k [ 1 , j 1 ] { 0 } forms a partition of T j | 1 ;
(ii) 
{ N j | 1 ( k ) } k [ 1 , j 1 ] { 0 } is a collection of disjoint subsets of N j | 1 .
Proof. 
It can be seen from the definitions of { T j | 1 ( k ) } k [ 1 , j 1 ] { 0 } and { N j | 1 ( k ) } k [ 1 , j 1 ] { 0 } that they are collections of mutually disjoint subsets of T j | 1 and N j | 1 , respectively. It remains to argue that every element in T j | 1 belongs to T j | 1 ( k ) for some k [ 1 , j 1 ] { 0 } . Noting that the element y n in T j | 1 satisfies d ( c 1 , y n | S 1 , j ) 1 , j 1 , we differentiate two cases: d ( c 1 , y n | S 1 , j ) 1 , j 2 and d ( c 1 , y n | S 1 , j ) = 1 , j 1 . For the former case, d ( c 1 , y n | S 1 , j ( m ) ) < 1 , j ( m ) 1 must hold for m = η k ; hence, this y n will be contained in T j | 1 ( k ) . For the latter case, y n will be included in T j | 1 ( 1 , j 1 ) . The proof is thus completed. □
In light of Proposition 3, we can apply the ratio-sum inequality to obtain
P X n , Y n ( c 1 , T j | 1 ) P X n , Y n ( c 1 , N j | 1 ) k = 0 : T j | 1 ( k ) 1 , j 1 P X n , Y n c 1 , T j | 1 ( k ) k = 0 : T j | 1 ( k ) 1 , j 1 P X n , Y n c 1 , N j | 1 ( k )
max k [ 1 , j 1 ] { 0 } : T j | 1 ( k ) P X n , Y n c 1 , T j | 1 ( k ) P X n , Y n c 1 , N j | 1 ( k ) .
We continue to construct a fine partition of T j | 1 ( k ) and the corresponding disjoint subsets of N j | 1 ( k ) in Proposition 4 after giving the next definition.
Definition 4.
Define for u n T j | 1 ( k ) ,
{ (55a) T j | 1 ( u n ; k ) y n T j | 1 ( k ) : d u n , y n | S 1 , j ( η k ) c = 0 ; (55b) N j | 1 ( u n ; k ) y n N j | 1 ( k ) : d u n , y n | ( S 1 , j ( η k ) ) c = 0 ,
where η k is given in (49).
Note from Definition 4 that for one element u n in non-empty T j | 1 ( k ) , we can find a group of elements that have identical bit components to u n with indices in ( S 1 , j ( η k ) ) c . We denote this group as T j | 1 ( u n ; k ) . We continue this grouping manipulation until all elements in T j | 1 ( k ) are exhausted as summarized below.
Proposition 4.
For non-empty T j | 1 ( k ) , there exists a representative subset U j | 1 ( k ) T j | 1 ( k ) such that the following two properties hold.
(i) 
T j | 1 ( u n ; k ) u n U j | 1 ( k ) forms a (non-empty) partition of T j | 1 ( k ) ;
(ii) 
N j | 1 ( u n ; k ) u n U j | 1 ( k ) is a collection of (non-empty) disjoint subsets of N j | 1 ( k ) .
Since the above proposition can be self-validated via its sequential selection manipulation of each u n from T j | 1 ( k ) , we omit the proof. Interested readers can find the details in [1] (Section III-C).
From Proposition 4, using again the ratio-sum inequality, we obtain that for non-empty T j | 1 ( k ) ,
P X n , Y n c 1 , T j | 1 ( k ) P X n , Y n c 1 , N j | 1 ( k ) u n U j | 1 ( k ) P X n , Y n c 1 , T j | 1 ( u n ; k ) u n U j | 1 ( k ) P X n , Y n c 1 , N j | 1 ( u n ; k )
max u n U j | 1 ( k ) P X n , Y n c 1 , T j | 1 ( u n ; k ) P X n , Y n c 1 , N j | 1 ( u n ; k ) .
Noting that the above result can be similarly conducted for general i > 1 , we combine (42), (54) and (57) to conclude that
δ n b n 2 max i [ M ] and j [ M ] \ { i } : T j | i max k [ i , j 1 ] { 0 } : T j | i ( k ) max u n U j | i ( k ) P X n , Y n c i , T j | i ( u n ; k ) P X n , Y n c i , N j | i ( u n ; k ) .
The final task is to evaluate P X n , Y n c i , T j | i ( u n ; k ) / P X n , Y n c i , N j | i ( u n ; k ) in order to characterize a linear upper bound for δ n / b n .

3.4. Characterization of a Linear Upper Bound for δ n / b n

We again focus on i = 1 with c 1 being the all-zero codeword for simplicity. The definitions of T j | 1 ( u n ; k ) in (55a) and N j | 1 ( u n ; k ) in (55b) indicate that when dealing with the ratio P X n , Y n ( c 1 , T j | 1 ( u n ; k ) ) / P X n , Y n ( c 1 , N j | 1 ( u n ; k ) ) , we only need to consider those bits with indices in S 1 , j ( η k ) because the remaining bits of all tuples in T j | 1 ( u n ; k ) and N j | 1 ( u n ; k ) have identical values as u n . Note that all | T j | 1 ( u n ; k ) | elements in T j | 1 ( u n ; k ) have exactly k ones with indices in S 1 , j ( η k ) , and all | N j | 1 ( u n ; k ) | elements in N j | 1 ( u n ; k ) have exactly k + 1 ones with indices in S 1 , j ( η k ) , we can immediately infer that
P X n , Y n ( c 1 , T j | 1 ( u n ; k ) ) P X n , Y n ( c 1 , N j | 1 ( u n ; k ) ) = P X n ( c 1 ) · P Y n | c 1 ( T j | 1 ( u n ; k ) | c 1 ) P X n ( c 1 ) · P Y n | c 1 ( N j | 1 ( u n ; k ) | c 1 ) = ( 1 p ) p · | T j | 1 ( u n ; k ) | | N j | 1 ( u n ; k ) | .
The cardinalities of T j | 1 ( u n ; k ) and N j | 1 ( u n ; k ) then decide the ratio in (59) as verified in the next proposition, based on which the proof of Theorem 2 can be completed from (58).
Proposition 5.
For u n T j | 1 ( k ) , we have
P X n , Y n ( c 1 , T j | 1 ( u n ; k ) ) P X n , Y n ( c 1 , N j | 1 ( u n ; k ) ) ( 1 p ) p n .
Proof. 
Recall from (22a), (52a) and (55a) that y n T j | 1 ( u n ; k ) T j | 1 ( k ) T j | 1 if and only if
{ (61a) P X n , Y n ( c 1 , y n ) = P X n , Y n ( c j , y n ) = max h [ M ] \ { 1 } P X n , Y n ( x ( h ) n , y n ) and d ( c 1 , y n | S 1 , j ) < | S i , j | ; (61b) 1 , j ( η k 1 ) 1 d c 1 , y n | S 1 , j ( η k 1 ) and d c 1 , y n | S 1 , j ( η k ) = k ; (61c) d u n , y n | ( S 1 , j ( η k ) ) c = 0 .
Thus, the number of elements in T j | 1 ( u n ; k ) is exactly the number of channel outputs y n fulfilling the above three conditions. We then examine the number of y n satisfying (61b) and (61c). Noting that these y n have either 1 , j ( η k 1 ) 1 ones or 1 , j ( η k 1 ) ones with indices in S 1 , j ( η k 1 ) , we know that there are at most
1 , j ( η k 1 ) 1 , j ( η k 1 ) 1 j ( η k ) 1 , j ( η k 1 ) k ( 1 , j ( η k 1 ) 1 ) + 1 , j ( η k 1 ) 1 , j ( η k 1 ) j ( η k ) 1 , j ( η k 1 ) k 1 , j ( η k 1 )
of y n tuples satisfying (61b) and (61c). Disregarding (61a), we get that the number of elements in T j | 1 ( u n ; k ) is upper-bounded by (62).
On the other hand, from (22b), (52b) and (55b), we obtain that w n N j | 1 ( u n ; k ) N j | 1 ( k ) N j | 1 if and only if
{ (63a) P X n , Y n ( c 1 , w n ) · q 2 = P X n , Y n ( c j , w n ) ; (63b) P X n , Y n ( c 1 , w n ) · q 2 P X n , Y n ( c r , w n ) for r [ j 1 ] \ { 1 } ; (63c) 1 , j ( η k 1 ) = d c 1 , w n | S 1 , j ( η k 1 ) and d c 1 , w n | S 1 , j ( η k ) = k + 1 ; (63d) d u n , w n | ( S 1 , j ( η k ) ) c = 0 .
We then claim that any w n satisfying (63c) and (63d) directly validate (63a) and (63b). Note that the validity of the claim, which we prove in Appendix B, immediately implies that the number of elements in N j | 1 ( u n ; k ) can be determined by (63c) and (63d), and hence
| N j | 1 ( u n ; k ) | = j ( η k ) 1 , j ( η k 1 ) k + 1 1 , j ( η k 1 ) .
Under this claim, (62) and (64) result in
| T j | 1 ( u n ; k ) | | N j | 1 ( u n ; k ) | 1 , j ( η k 1 ) 1 , j ( η k 1 ) 1 j ( η k ) 1 , j ( η k 1 ) k ( 1 , j ( η k 1 ) 1 ) + 1 , j ( η k 1 ) 1 , j ( η k 1 ) j ( η k ) 1 , j ( η k 1 ) k 1 , j ( η k 1 ) j ( η k ) 1 , j ( η k 1 ) k + 1 1 , j ( η k 1 )
= 1 , j ( η k 1 ) + k + 1 1 , j ( η k 1 ) j ( η k ) k
1 , j ( η k 1 ) + j ( η k ) 1 , j ( η k 1 ) 1
n ,
where (67) holds because k j ( η k ) 1 by (49), and (68) follows from 1 , j ( η k ) 1 , j n . The proof of the proposition is thus completed by (59) and (68). □

4. Conclusions

In this paper, we analyzed the error probability of block codes sent over the memoryless BSC under an arbitrary (not necessarily uniform) input distribution and used in conjunction with (optimal) MAP decoding. We showed that decoder ties do not affect the error exponent of the probability of error, thus extending a similar result recently established in [1] for uniformly distributed channel inputs. This result was obtained by proving that the relative deviation of the error probability from the probability of error grows no more than linearly in blocklength when no MAP decoding ties occur, directly implying that decoder ties have only a sub-exponential effect on the error probability as blocklength grows without bound. Future work includes further extending this result for more general channels used under arbitrary input statistics, such as non-binary symmetric channels (Note that the result of Theorem 1 can be extended for non-binary (q-ary, q > 2 ) codes sent over q-ary symmetric memoryless channels under a uniform input distribution; see [25] (Theorem 2).) and binary non-symmetric channels. Studying how to sharpen the upper bound derived in (10) for “sufficiently good” codes as highlighted in Remark 1 and for codes with small blocklengths are other worthwhile future directions.

Author Contributions

Conceptualization, L.-H.C., P.-N.C. and F.A.; Formal analysis, L.-H.C.; Writing—original draft, L.-H.C.; Writing—review & editing, P.-N.C. and F.A. All authors have read and agreed to the published version of the manuscript.

Funding

The work of Ling-Hua Chang is supported by the Ministry of Science and Technology, Taiwan under Grant MOST 109-2221-E-155-035-MY3. The work of Po-Ning Chen is supported by the Ministry of Science and Technology, Taiwan, under Grant MOST 110-2221-E-A49-024-MY3. The work of Fady Alajaji is supported by the Natural Sciences and Engineering Research Council (NSERC) of Canada.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Supplement to Example 1

Under distribution
P X 4 ( c 1 ) = q 2 2 + q 2 + q 2
P X 4 ( c 2 ) = P X 4 ( c 3 ) = 1 2 + q 2 + q 2
P X 4 ( c 4 ) = q 2 2 + q 2 + q 2
over the code C 4 = { c 1 , c 2 , c 3 , c 4 } = { 0000 , 0101 , 0110 , 0111 } , we obtain:
T 1 = y 4 { 0 , 1 } 4 : P X 4 , Y 4 ( c 1 , y 4 ) = max r [ 4 ] \ { 1 } P X 4 , Y 4 ( c r , y 4 )
= y 4 { 0 , 1 } 4 : P X 4 ( c 1 ) q d ( c 1 , y 4 ) = max P X 4 ( c 2 ) q d ( c 2 , y 4 ) , P X 4 ( c 3 ) q d ( c 3 , y 4 ) , P X 4 ( c 4 ) q d ( c 4 , y 4 )
= y 4 { 0 , 1 } 4 : d ( c 1 , y 4 ) 2 = min d ( c 2 , y 4 ) , d ( c 3 , y 4 ) , d ( c 4 , y 4 ) + 2
= 0101 , 0110 , 0111 , 1101 , 1110 , 1111 ,
where (A5) follows from (25), and
T j | 1 y 4 T 1 : j = min r I 1 ( y 4 ) : d ( c 1 , y 4 | S 1 , r ) < | S 1 , r | r = for j = 2 , 3 , 4 ,
and
T j | 1 y 4 T 1 \ h [ 4 ] \ { 1 } T h | 1 : j = min r I 1 ( y n ) r
= { 0101 , 0111 , 1101 , 1111 } , j = 2 ; { 0110 , 1110 } , j = 3 ; , j = 4 .
The above derivations are verified via Table A1. Continuing with the same setting, we obtain
T 2 = y 4 { 0 , 1 } 4 : P X 4 , Y 4 ( c 2 , y 4 ) = max r [ 4 ] \ { 2 } P X 4 , Y 4 ( c r , y 4 )
= y 4 { 0 , 1 } 4 : P X 4 ( c 2 ) q d ( c 2 , y 4 ) = max P X 4 ( c 1 ) q d ( c 1 , y 4 ) , P X 4 ( c 3 ) q d ( c 3 , y 4 ) , P X 4 ( c 4 ) q d ( c 4 , y 4 )
= y 4 { 0 , 1 } 4 : d ( c 2 , y 4 ) = min d ( c 1 , y 4 ) 2 , d ( c 3 , y 4 ) , d ( c 4 , y 4 ) + 2
= 0101 , 0111 , 1101 , 1111 ,
T j | 2 y 4 T 2 : j = min r I 2 ( y 4 ) : d ( c 2 , y 4 | S 2 , r ) < | S 2 , r | r = T 2 , j = 1 ; , j = 3 , 4
and
T j | 2 y 4 T 2 \ h [ 4 ] \ { 1 } T h | 2 : j = min r I 2 ( y n ) r = for j = 1 , 3 , 4 ,
where the above derivations are also confirmed via Table A1. Based on Table A1, we further have
T 3 = y 4 { 0 , 1 } 4 : P X 4 , Y 4 ( c 3 , y 4 ) = max r [ 4 ] \ { 3 } P X 4 , Y 4 ( c r , y 4 )
= y 4 { 0 , 1 } 4 : P X 4 ( c 3 ) q d ( c 3 , y 4 ) = max P X 4 ( c 1 ) q d ( c 1 , y 4 ) , P X 4 ( c 2 ) q d ( c 2 , y 4 ) , P X 4 ( c 4 ) q d ( c 4 , y 4 )
= y 4 { 0 , 1 } 4 : d ( c 3 , y 4 ) = min d ( c 1 , y 4 ) 2 , d ( c 2 , y 4 ) , d ( c 4 , y 4 ) + 2
= 0110 , 0111 , 1110 , 1111 ,
T j | 3 y 4 T 3 : j = min r I 3 ( y 4 ) : d ( c 3 , y 4 | S 3 , r ) < | S 3 , r | r = T 3 , j = 1 ; , j = 2 , 4 ,
and
T j | 3 y 4 T 3 \ h [ 4 ] \ { 3 } T h | 3 : j = min r I 3 ( y n ) r = for j = 1 , 2 , 4 .
Furthermore, we establish from Table A1 that
T 4 = y 4 { 0 , 1 } 4 : P X 4 , Y 4 ( c 4 , y 4 ) = max r [ 4 ] \ { 4 } P X 4 , Y 4 ( c r , y 4 )
= y 4 { 0 , 1 } 4 : P X 4 ( c 4 ) q d ( c 4 , y 4 ) = max P X 4 ( c 1 ) q d ( c 1 , y 4 ) , P X 4 ( c 2 ) q d ( c 2 , y 4 ) , P X 4 ( c 3 ) q d ( c 3 , y 4 )
= y 4 { 0 , 1 } 4 : d ( c 4 , y 4 ) + 2 = min d ( c 1 , y 4 ) 2 , d ( c 2 , y 4 ) , d ( c 3 , y 4 )
= ,
T j | 4 y 4 T 4 : j = min r I 4 ( y 4 ) : d ( c 4 , y 4 | S 4 , r ) < | S 4 , r | r = for j = 1 , 2 , 3 ,
and
T j | 4 y 4 T 4 \ h [ 4 ] \ { 4 } T h | 4 : j = min r I 4 ( y n ) r = for j = 1 , 2 , 3 .
After summarizing all sets derived above in Table A2, we remark that
T 2 | 1 T 1 | 2 and P X 4 , Y 4 ( c 1 , y 4 ) = P X 4 , Y 4 ( c 2 , y 4 ) for every y 4 T 2 | 1 ; T 3 | 1 T 1 | 3 and P X 4 , Y 4 ( c 1 , y 4 ) = P X 4 , Y 4 ( c 2 , y 4 ) for every y 4 T 3 | 1 .
Note that { T j | i } i [ M ] , j [ M ] \ { i } are disjoint as confirmed in Remark 4 such that every element y n T j | i appears only once in the following summation:
i [ 4 ] j [ 4 ] \ { i } P X 4 , Y 4 c i , T j | i = P X 4 , Y 4 c 1 , 0101 + P X 4 , Y 4 c 1 , 0110 + P X 4 , Y 4 c 1 , 0111 + P X 4 , Y 4 c 1 , 1101 + P X 4 , Y 4 c 1 , 1110 + P X 4 , Y 4 c 1 , 1111
= p 4 q 2 2 + q 2 + q 2 q 2 + q 2 + q + q + q + 1
= p 4 q 2 2 + q 2 + q 2 2 q 2 + 3 q + 1 .
Additionally,
i [ 4 ] j [ 4 ] \ { i } P X 4 , Y 4 c i , T j | i = P X 4 , Y 4 c 2 , 0101 + P X 4 , Y 4 c 2 , 0111 + P X 4 , Y 4 c 2 , 1101 + P X 4 , Y 4 c 2 , 1111 + P X 4 , Y 4 c 3 , 0110 + P X 4 , Y 4 c 3 , 0111 + P X 4 , Y 4 c 3 , 1110 + P X 4 , Y 4 c 3 , 1111
= p 4 2 + q 2 + q 2 q 4 + q 3 + q 3 + q 2 + q 4 + q 3 + q 3 + q 2
= i [ 4 ] j [ 4 ] \ { i } P X 4 , Y 4 c i , T j | i + p 4 q 2 2 + q 2 + q 2 q + 1 .
Finally, we have
N 1 = y 4 { 0 , 1 } 4 : P X 4 ( c 1 ) q d ( c 1 , y 4 ) < max P X 4 ( c 2 ) q d ( c 2 , y 4 ) , P X 4 ( c 3 ) q d ( c 3 , y 4 ) , P X 4 ( c 4 ) q d ( c 4 , y 4 )
= y 4 { 0 , 1 } 4 : d ( c 1 , y 4 ) 2 > min d ( c 2 , y 4 ) , d ( c 3 , y 4 ) , d ( c 4 , y 4 ) + 2
= ,
N 2 = y 4 { 0 , 1 } 4 : d ( c 2 , y 4 ) > min d ( c 1 , y 4 ) 2 , d ( c 3 , y 4 ) , d ( c 4 , y 4 ) + 2 = { 0 , 1 } 4 \ T 2 ,
N 3 = y 4 { 0 , 1 } 4 : d ( c 3 , y 4 ) > min d ( c 2 , y 4 ) , d ( c 1 , y 4 ) 2 , d ( c 4 , y 4 ) + 2 = { 0 , 1 } 4 \ T 3 ,
N 4 = y 4 { 0 , 1 } 4 : d ( c 4 , y 4 ) + 2 > min d ( c 2 , y 4 ) , d ( c 3 , y 4 ) , d ( c 1 , y 4 ) 2 = { 0 , 1 } 4 ,
N j | 1 = for j = 2 , 3 , 4 ,
N j | 2 = { y 4 N 2 : P X 4 , Y 4 ( c 2 , y 4 ) · q = P X 4 , Y 4 ( c j , y 4 ) · 1 q
and P X 4 , Y 4 ( c j , y 4 ) P X 4 , Y 4 ( c r , y 4 ) for r [ j 1 ] \ { 2 } } = { y 4 N 2 : P X 4 ( c 2 ) q d ( c 2 , y 4 ) 1 = P X 4 ( c j ) q d ( c j , y 4 ) + 1 and P X 4 ( c j ) q d ( c j , y 4 ) P X 4 ( c r ) q d ( c r , y 4 )
for r [ j 1 ] \ { 2 } }
= y 4 N 2 : P X 4 ( c 2 ) q d ( c 2 , y 4 ) 1 = P X 4 ( c 1 ) q d ( c 1 , y 4 ) + 1 , j = 1 ; y 4 N 2 : P X 4 ( c 2 ) q d ( c 2 , y 4 ) 1 = P X 4 ( c 3 ) q d ( c 3 , y 4 ) + 1 and P X 4 ( c 3 ) q d ( c 3 , y 4 ) P X 4 ( c 1 ) q d ( c 1 , y 4 ) , j = 3 ; { y 4 N 2 : P X 4 ( c 2 ) q d ( c 2 , y 4 ) 1 = P X 4 ( c 4 ) q d ( c 4 , y 4 ) + 1 and P X 4 ( c 4 ) q d ( c 4 , y 4 ) P X 4 ( c r ) q d ( c r , y 4 ) for r [ 3 ] \ { 2 } } , j = 4
= y 4 N 2 : d ( c 2 , y 4 ) = d ( c 1 , y 4 ) , j = 1 ; y 4 N 2 : d ( c 2 , y 4 ) = d ( c 3 , y 4 ) + 2 and d ( c 2 , y 4 ) d ( c 1 , y 4 ) , j = 3 ; { y 4 N 2 : d ( c 2 , y 4 ) = d ( c 4 , y 4 ) + 4 , d ( c 2 , y 4 ) d ( c 1 , y 4 ) and d ( c 2 , y 4 ) d ( c 3 , y 4 ) + 2 } , j = 4
= 0001 , 0100 , 0011 , 0110 , 1001 , 1100 , 1011 , 1110 , j = 1 ; 0010 , 1010 , j = 3 ; , j = 4 ,
N j | 3 = { y 4 N 3 : P X 4 , Y 4 ( c 3 , y 4 ) · q = P X 4 , Y 4 ( c j , y 4 ) · 1 q
and P X 4 , Y 4 ( c j , y 4 ) P X 4 , Y 4 ( c r , y 4 ) for r [ j 1 ] \ { 3 } } = { y 4 N 3 : P X 4 ( c 3 ) q d ( c 3 , y 4 ) 1 = P X 4 ( c j ) q d ( c j , y 4 ) + 1 and P X 4 ( c j ) q d ( c j , y 4 ) P X 4 ( c r ) q d ( c r , y 4 )
for r [ j 1 ] \ { 3 } }
= y 4 N 3 : P X 4 ( c 3 ) q d ( c 3 , y 4 ) 1 = P X 4 ( c 1 ) q d ( c 1 , y 4 ) + 1 , j = 1 ; y 4 N 3 : P X 4 ( c 3 ) q d ( c 3 , y 4 ) 1 = P X 4 ( c 2 ) q d ( c 2 , y 4 ) + 1 and P X 4 ( c 2 ) q d ( c 2 , y 4 ) P X 4 ( c 1 ) q d ( c 1 , y 4 ) , j = 2 ; { y 4 N 3 : P X 4 ( c 3 ) q d ( c 3 , y 4 ) 1 = P X 4 ( c 4 ) q d ( c 4 , y 4 ) + 1 and P X 4 ( c 4 ) q d ( c 4 , y 4 ) P X 4 ( c r ) q d ( c r , y 4 ) for r [ 3 ] \ { 3 } } , j = 4 = y 4 N 3 : d ( c 3 , y 4 ) = d ( c 1 , y 4 ) , j = 1 ; y 4 N 3 : d ( c 3 , y 4 ) = d ( c 2 , y 4 ) + 2 and d ( c 3 , y 4 ) d ( c 1 , y 4 ) , j = 3 ; { y 4 N 3 : d ( c 3 , y 4 ) = d ( c 4 , y 4 ) + 4 , d ( c 3 , y 4 ) d ( c 1 , y 4 ) and d ( c 3 , y 4 ) d ( c 2 , y 4 ) + 2 } , j = 4
= 0010 , 0100 , 0011 , 0101 , 1010 , 1100 , 1011 , 1101 , j = 1 ; 0001 , 1001 , j = 3 ; , j = 4 ,
N j | 4 = { y 4 N 4 : P X 4 , Y 4 ( c 4 , y 4 ) · q = P X 4 , Y 4 ( c j , y 4 ) · 1 q
and P X 4 , Y 4 ( c j , y 4 ) P X 4 , Y 4 ( c r , y 4 ) for r [ j 1 ] \ { 4 } } = y 4 N 4 : P X 4 ( c 4 ) q d ( c 4 , y 4 ) 1 = P X 4 ( c j ) q d ( c j , y 4 ) + 1 and P X 4 ( c j ) q d ( c j , y 4 ) P X 4 ( c r ) q d ( c r , y 4 ) for r [ j 1 ]
= y 4 N 4 : P X 4 ( c 4 ) q d ( c 4 , y 4 ) 1 = P X 4 ( c 1 ) q d ( c 1 , y 4 ) + 1 , j = 1 ; y 4 N 4 : P X 4 ( c 4 ) q d ( c 4 , y 4 ) 1 = P X 4 ( c 2 ) q d ( c 2 , y 4 ) + 1 and P X 4 ( c 2 ) q d ( c 2 , y 4 ) P X 4 ( c 1 ) q d ( c 1 , y 4 ) , j = 2 ; { y 4 N 4 : P X 4 ( c 4 ) q d ( c 4 , y 4 ) 1 = P X 4 ( c 3 ) q d ( c 3 , y 4 ) + 1 and P X 4 ( c 3 ) q d ( c 3 , y 4 ) P X 4 ( c r ) q d ( c r , y 4 ) for r [ 2 ] } , j = 3
= y 4 N 4 : d ( c 4 , y 4 ) = d ( c 1 , y 4 ) 2 , j = 1 ; y 4 N 4 : d ( c 4 , y 4 ) = d ( c 2 , y 4 ) and d ( c 4 , y 4 ) d ( c 1 , y 4 ) 2 , j = 2 ; { y 4 N 4 : d ( c 4 , y 4 ) = d ( c 3 , y 4 ) , d ( c 4 , y 4 ) d ( c 1 , y 4 ) 2 and d ( c 4 , y 4 ) d ( c 2 , y 4 ) } , j = 3
= for j = 1 , 2 , 3 .
Table A1. Measures used in Example 1.
Table A1. Measures used in Example 1.
d ( 0000 , y 4 ) 2 d ( 0101 , y 4 ) d ( 0110 , y 4 ) d ( 0111 , y 4 ) + 2 I 1 ( y 4 ) I 2 ( y 4 ) I 3 ( y 4 ) I 4 ( y 4 )
y 4 = 0000 2 225
y 4 = 0001 1 134
y 4 = 0010 1 314
y 4 = 0100 1 114
y 4 = 1000 1 336
y 4 = 0011 0223
y 4 = 0101 0023 { 2 } { 1 }
y 4 = 0110 0203 { 3 } { 1 }
y 4 = 1001 0245
y 4 = 1010 0425
y 4 = 1100 0225
y 4 = 0111 1112 { 2 , 3 } { 1 , 3 } { 1 , 2 }
y 4 = 1011 1334
y 4 = 1101 1134 { 2 } { 1 }
y 4 = 1110 1314 { 3 } { 1 }
y 4 = 1111 2223 { 2 , 3 } { 1 , 3 } { 1 , 2 }
Table A2. List of T i , N i , T j | i , T j | i and N j | i for i [ 4 ] and j [ 4 ] \ { i } in Example 1.
Table A2. List of T i , N i , T j | i , T j | i and N j | i for i [ 4 ] and j [ 4 ] \ { i } in Example 1.
T 1 0101 , 0110 , 0111 , 1101 , 1110 , 1111 N 1
T 2 0101 , 0111 , 1101 , 1111 N 2 { 0 , 1 } 4 \ T 2
T 3 0110 , 0111 , 1110 , 1111 N 3 { 0 , 1 } 4 \ T 3
T 4 N 4 { 0 , 1 } 4
T 2 | 1 T 2 | 1 { 0101 , 0111 , 1101 , 1111 } N 2 | 1
T 3 | 1 T 3 | 1 { 0110 , 1110 } N 3 | 1
T 4 | 1 T 4 | 1 N 4 | 1
T 1 | 2 { 0101 , 0111 , 1101 , 1111 } T 1 | 2 N 1 | 2 N 2 \ { 0000 , 0010 , 1000 , 1010 }
T 3 | 2 T 3 | 2 N 3 | 2 { 0010 , 1010 }
T 4 | 2 T 4 | 2 N 1 | 2
T 1 | 3 { 0110 , 0111 , 1110 , 1111 } T 1 | 3 N 1 | 3 N 3 \ { 0000 , 0001 , 1000 , 1001 }
T 2 | 3 T 2 | 3 N 2 | 3 { 0001 , 1001 }
T 4 | 3 T 4 | 3 N 4 | 3
T 1 | 4 T 1 | 4 N 1 | 4
T 2 | 4 T 2 | 4 N 2 | 4
T 3 | 4 T 3 | 4 N 3 | 4

Appendix B. The Proof of the Claim Supporting Proposition 5

We validate the claim that (63c) and (63d) imply (63a) and (63b) via the construction of an auxiliary v n N j | 1 ( u n ; k ) from u n T j | 1 ( u n ; k ) . This auxiliary v n will be defined differently according to whether d c 1 , u n | S 1 , j ( η k 1 ) equals 1 , j ( η k 1 ) or 1 , j ( η k 1 ) 1 as follows.
(i) 
d ( c 1 , u n | S 1 , j ( η k 1 ) ) = 1 , j ( η k 1 ) : In this case, u n has no zero components with indices in S 1 , j ( η k 1 ) . Moreover, d ( c 1 , u n | S 1 , j ( η k ) ) = k j ( η k ) 1 indicates that
u n has at least one zero component with its index in S 1 , j ( η k ) \ S 1 , j ( η k 1 ) = S 1 , j ( η k ) .
Therefore, we flip arbitrarily a zero component of u n with its index in S 1 , j ( η k ) to construct a v n such that
d ( c 1 , v n ) = d ( c 1 , u n ) + 1 and d ( c j , v n ) = d ( c j , u n ) 1 ,
which implies
P X n , Y n ( c 1 , v n ) = P X n , Y n ( c 1 , u n ) · 1 q and P X n , Y n ( c j , v n ) = P X n , Y n ( c j , u n ) · q .
Then, v n must fulfill (63a), (63c) and (63d) (with w n replaced by v n ) as u n satisfies (61a), (61b) and (61c). We next declare that v n also fulfills (63b) and will prove this declaration by contradiction.
Proof of the declaration: Suppose there exists a r [ j 1 ] \ { 1 } satisfying
P X n , Y n ( c 1 , v n ) · q 2 = P X n , Y n ( c r , v n ) .
We then recall from (45) that d ( c 1 , c r | S 1 , j ( η k ) ) is either 0 or | S 1 , j ( η k ) | . Thus, (A58) can be disproved by differentiating two subcases: ( 1 ) d ( c 1 , c r | S 1 , j ( η k ) ) = 0 , and ( 2 ) d ( c 1 , c r | S 1 , j ( η k ) ) = | S 1 , j ( η k ) | (Since 1 , j ( η k 1 ) < 1 , j ( η k ) as can be seen from (50) and (51), we have | S 1 , j ( η k ) | = 1 , j ( η k ) 1 , j ( η k 1 ) > 0 , i.e., S 1 , j ( η k ) non-empty).
In Subcase ( 1 ) , v n that is obtained by flipping a zero component of u n with index in S 1 , j ( η k ) must satisfy d ( c 1 , v n ) = d ( c 1 , u n ) + 1 and d ( c r , v n ) = d ( c r , u n ) + 1 , which is equivalent to
P X n , Y n ( c 1 , v n ) · q = P X n , Y n ( c 1 , u n ) and P X n | Y n ( c r | v n ) · q = P X n | Y n ( c r | u n ) .
Then, (A58) implies
P X n , Y n ( c 1 , u n ) · q 2 = P X n , Y n ( c r , u n ) .
Hence,
P X n , Y n ( c 1 , u n ) < P X n , Y n ( c r , u n ) max h [ M ] \ { 1 } P X n , Y n ( x ( h ) n , u n ) .
A contradiction to the fact that u n T j | 1 ( u n ; k ) satisfies (61a) (with y n replaced by u n ) is obtained.
In Subcase ( 2 ) , we note that d ( c 1 , c r | S 1 , j ( η k ) ) = | S 1 , j ( η k ) | implies S 1 , j ( η k ) S 1 , r . Therefore, (A55) leads to
d ( c 1 , u n | S 1 , r ) < | S 1 , r | .
The flipping manipulation on u n results in d ( c 1 , v n ) = d ( c 1 , u n ) + 1 and d ( c r , v n ) = d ( c r , u n ) 1 , which is equivalent to
P X n , Y n ( c 1 , v n ) · q = P X n , Y n ( c 1 , u n ) and P X n , Y n ( c r , v n ) = P X n , Y n ( c r , u n ) · q .
Therefore, (A58) implies
P X n , Y n ( c 1 , u n ) = P X n , Y n ( c r , u n ) ,
which together with max h [ M ] \ { 1 } P X n , Y n ( c h , u n ) = P X n , Y n ( c 1 , u n ) and (A62) result in u n T r | 1 because r < j . This contradicts u n T j | 1 . Accordingly, v n must also fulfill (63b); hence, v n N j | 1 ( u n ; k ) . This completes the proof of the declaration.
With this auxiliary v n , we are ready to prove that every w n satisfying (63c) and (63d) also validates (63a) and (63b). Toward this end, we need to prove
P X n , Y n ( c r , w n ) = P X n , Y n ( c r , v n ) for all r [ M ] .
Note that
{ (A66a) d w n , v n | S 1 , j ( η k 1 ) = 0 ; (A66b) d c r , w n | S 1 , j ( η k ) = d c r , v n | S 1 , j ( η k ) for all r [ M ] ; (A66c) d ( w n , v n | ( S 1 , j ( η k 1 ) ) c = 0 ,
where (A66a) holds because both v n and w n satisfy (63c), implying that all components of v n and w n with indices in S 1 , j ( η k 1 ) are equal to one; (A66b) holds because when considering only those portions with indices in (non-empty) S 1 , j ( η k ) , c r gives either all ones or all zeros according to (45), and both w n and v n have exactly k + 1 1 , j ( η k 1 ) ones according to (63c); and (A66c) is valid since both v n and w n satisfy (63d). Based on (A66a)–(A66c), we remark that d ( c r , w n ) = d ( c r , v n ) for all r [ M ] , which implies P Y n | X n ( w n | c r ) = P Y n | X n ( v n | c r ) (equivalently, P X n , Y n ( c r , w n ) = P X n , Y n ( c r , v n ) ) for all r [ M ] ).
(ii) 
d ( c 1 , u n | S 1 , j ( η k 1 ) ) = 1 , j ( η k 1 ) 1 : In this case, there is only one zero component of u n with its index in S 1 , j ( η k 1 ) . Suppose the index of such zero component lie in S 1 , j ( h ) S 1 , j ( η k 1 ) , where h η k 1 . The flipping manipulation to u n leads to v n , which has all one components with respect to S 1 , j ( η k 1 ) . Then, v n must fulfill (63a), (63c), and (63d) as u n satisfies (61a), (61b), and (61c). With the components of c r with respect to (non-empty) S 1 , j ( h ) being either all zeros or all ones, the same contradiction argument between (A58) and (A64), with η k replaced by h, can disprove the validity of (A58) for this v n and for any r [ j 1 ] \ { 1 } . Therefore, v n also fulfills (63b), implying v n N j | 1 ( u n ; k ) . With this auxiliary v n , we can again verify (A66a)–(A66c) via the same argument. The claim that w n satisfying (63c) and (63d) validates (63a) and (63b) is thus confirmed.

References

  1. Chang, L.H.; Chen, P.N.; Alajaji, F.; Han, Y.S. Decoder Ties Do Not Affect the Error Exponent of the Memoryless Binary Symmetric Channel. IEEE Trans. Inf. Theory 2022, 68, 3501–3510. [Google Scholar] [CrossRef]
  2. Shannon, C.E.; Gallager, R.G.; Berlekamp, E.R. Lower bounds to error probability for coding on discrete memoryless channels—I. Inf. Control 1967, 10, 65–103. [Google Scholar] [CrossRef]
  3. Shannon, C.E.; Gallager, R.G.; Berlekamp, E.R. Lower bounds to error probability for coding on discrete memoryless channels—II. Inf. Control 1967, 10, 522–552. [Google Scholar] [CrossRef]
  4. McEliece, R.J.; Omura, J.K. An improved upper bound on the block coding error exponent for binary-input discrete memoryless channels. IEEE Trans. Inf. Theory 1977, 23, 611–613. [Google Scholar] [CrossRef]
  5. Gallager, R.G. Information Theory and Reliable Communication; Wiley: New York, NY, USA, 1968. [Google Scholar]
  6. Viterbi, A.J.; Omura, J.K. Principles of Digital Communication and Coding; McGraw-Hill: New York, NY, USA, 1979. [Google Scholar]
  7. Csiszár, I.; Körner, J. Information Theory: Coding Theorems for Discrete Memoryless Systems; Academic Press: New York, NY, USA, 1981. [Google Scholar]
  8. Blahut, R. Principles and Practice of Information Theory; Addison-Wesley Longman Publishing Co., Inc.: Albany, NY, USA, 1988. [Google Scholar]
  9. Barg, A.; McGregor, A. Distance distribution of binary codes and the error probability of decoding. IEEE Trans. Inf. Theory 2005, 51, 4237–4246. [Google Scholar] [CrossRef]
  10. Haroutunian, E.A.; Haroutunian, M.E.; Harutyunyan, A.N. Reliability Criteria in Information Theory and in Statistical Hypothesis Testing. In Foundations and Trends in Communications and Information Theory; Now Publishers Inc.: Delft, The Netherlands, 2007; Volume 4, pp. 97–263. [Google Scholar]
  11. Dalai, M. Lower bounds on the probability of error for classical and classical-quantum channels. IEEE Trans. Inf. Theory 2013, 59, 8027–8056. [Google Scholar] [CrossRef]
  12. Burnashev, M.V. On the BSC reliability function: Expanding the region where it is known exactly. Probl. Inf. Transm. 2015, 51, 307–325. [Google Scholar] [CrossRef]
  13. Csiszár, I. Joint source-channel error exponent. Probl. Control. Inf. Theory 1980, 9, 315–328. [Google Scholar]
  14. Zhong, Y.; Alajaji, F.; Campbell, L. On the joint source-channel coding error exponent for discrete memoryless systems. IEEE Trans. Inf. Theory 2006, 52, 1450–1468. [Google Scholar] [CrossRef]
  15. Alajaji, F.; Phamdo, N.; Fuja, T. Channel codes that exploit the residual redundancy in CELP-encoded speech. IEEE Trans. Speech Audio Process. 1996, 4, 325–336. [Google Scholar] [CrossRef]
  16. Xu, W.; Hagenauer, J.; Hollmann, J. Joint source-channel decoding using the residual redundancy in compressed images. In Proceedings of the Proceedings of the International Conference on Communications, Washington, DC, USA, 25–28 February 1996; Volume 1, pp. 142–148. [Google Scholar]
  17. Hagenauer, J. Source-controlled channel decoding. IEEE Trans. Commun. 1995, 43, 2449–2457. [Google Scholar] [CrossRef]
  18. Goertz, N. Joint Source-Channel Coding of Discrete-Time Signals with Continuous Amplitudes; World Scientific: Singapore, 2007. [Google Scholar]
  19. Duhamel, P.; Kieffer, M. Joint Source-Channel Decoding: A Cross-Layer Perspective with Applications in Video Broadcasting; Academic Press: Cambridge, MA, USA, 2009. [Google Scholar]
  20. Fresia, M.; Pérez-Cruz, F.; Poor, H.V.; Verdú, S. Joint source and channel coding. IEEE Signal Process. Mag. 2010, 27, 104–113. [Google Scholar] [CrossRef]
  21. Alajaji, F.; Chen, P.N. An Introduction to Single-User Information Theory; Springer: Berlin/Heidelberg, Germany, 2018. [Google Scholar]
  22. Chang, L.H.; Chen, P.N.; Alajaji, F.; Han, Y.S. The asymptotic generalized Poor-Verdú bound achieves the BSC error exponent at zero rate. In Proceedings of the IEEE International Symposium on Information Theory, Los Angeles, CA, USA, 21–26 June 2020. [Google Scholar]
  23. Chen, P.N.; Alajaji, F. A generalized Poor-Verdú error bound for multihypothesis testings. IEEE Trans. Inf. Theory 2012, 58, 311–316. [Google Scholar] [CrossRef]
  24. Poor, H.V.; Verdú, S. A lower bound on the probability of error in multihypothesis testing. IEEE Trans. Inf. Theory 1995, 41, 1992–1994. [Google Scholar] [CrossRef]
  25. Chang, L.H.; Chen, P.N.; Alajaji, F.; Han, Y.S. Tightness of the asymptotic generalized Poor-Verdú error bound for the memoryless symmetric channel. arXiv 2020, arXiv:2007.04080v1. [Google Scholar]
Figure 1. An illustration, based on the setting in Example 1 for a non-uniformly distributed binary code (with M = n = 4 ) given by C 4 = { c 1 , c 2 , c 3 , c 4 } = { 0000 , 0101 , 0110 , 0111 } of the non-empty component subsets of Y n defined in Table 1 and corresponding to codewords c 1 = 0000 (left figure) and c 2 = 0101 (right figure).
Figure 1. An illustration, based on the setting in Example 1 for a non-uniformly distributed binary code (with M = n = 4 ) given by C 4 = { c 1 , c 2 , c 3 , c 4 } = { 0000 , 0101 , 0110 , 0111 } of the non-empty component subsets of Y n defined in Table 1 and corresponding to codewords c 1 = 0000 (left figure) and c 2 = 0101 (right figure).
Entropy 25 00668 g001
Table 1. Summary of the main symbols used in this paper.
Table 1. Summary of the main symbols used in this paper.
SymbolDescriptionDefined in
[ M ] A shorthand for { 1 , 2 , , M }
C n The code c 1 , c 2 , , c n with c 1 being the all-zero codeword
d ( u n , v n | S ) The Hamming distance between the portions of u n and v n with indices in S
All terms below are functions of C n (this dependence is not explicitly shown to simplify notation)
T i The set of channel outputs y n inducing a decoder tie when c i is sent(12)
N i The set of channel outputs y n leading to a tie-free decoder decision error when c i is sent(15)
I i ( y n ) The set { m [ M ] \ { i } : y n T m } for y n T i (21)
S i , j The set of indices for which the components of c i and c j differ
i , j The size of S i , j , i.e., | S i , j |
T j | i The subset of T i consisting of channel outputs y n such that j is the minimal(22a)
number r in I i ( y n ) satisfying d ( c i , y n | S i , r ) < i , r
N j | i The subset of N i consisting of channel outputs y n that satisfy P X n , Y n ( c i , y n ) · q = (22b)
P X n , Y n ( c j , y n ) · 1 q and that are not included in N r | i for r [ j 1 ] { i }
T j | i The subset of T i \ h [ M ] \ { i } T h | i consisting of channel outputs y n (23)
such that j is the minimal number in I i ( y n )
S 1 , j ( m ) The subset of S 1 , j defined according to whether each index in S 1 , j is in each(43)
of S 1 , 2 , , S 1 , j 1 , S 1 , j + 1 , , S 1 , M
S 1 , j ( m ) The union of S 1 , j ( 1 ) , S 1 , j ( 2 ) , , S 1 , j ( m ) (48)
1 , j ( m ) The size of S 1 , j ( m ) , i.e., | S 1 , j ( m ) |
η k The mapping from k { 0 , 1 , , j 1 } to [ 2 M 2 ] used for partitioning T j | 1 into 1 , j (49)
subsets { T j | 1 ( k ) } 0 k < 1 , j
T j | 1 ( k ) The kth partition of T j | 1 for k = 0 , 1, , 1 , j 1 (52a)
N j | 1 ( k ) The kth subset of N j | 1 for k = 0 , 1, , 1 , j 1 (52b)
U j | 1 ( k ) The set of representative elements in T j | 1 ( k ) for partitioning T j | 1 ( k )
T j | 1 ( u n ; k ) The subset of T j | 1 ( k ) associated with u n U j | 1 ( k ) (55a)
N j | 1 ( u n ; k ) The subset of N j | 1 ( k ) associated with u n U j | 1 ( k ) (55b)
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chang, L.-H.; Chen, P.-N.; Alajaji, F. On Decoder Ties for the Binary Symmetric Channel with Arbitrarily Distributed Input. Entropy 2023, 25, 668. https://doi.org/10.3390/e25040668

AMA Style

Chang L-H, Chen P-N, Alajaji F. On Decoder Ties for the Binary Symmetric Channel with Arbitrarily Distributed Input. Entropy. 2023; 25(4):668. https://doi.org/10.3390/e25040668

Chicago/Turabian Style

Chang, Ling-Hua, Po-Ning Chen, and Fady Alajaji. 2023. "On Decoder Ties for the Binary Symmetric Channel with Arbitrarily Distributed Input" Entropy 25, no. 4: 668. https://doi.org/10.3390/e25040668

APA Style

Chang, L. -H., Chen, P. -N., & Alajaji, F. (2023). On Decoder Ties for the Binary Symmetric Channel with Arbitrarily Distributed Input. Entropy, 25(4), 668. https://doi.org/10.3390/e25040668

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop