Next Article in Journal
Competitive Conditions in Global Value Chain Networks: An Assessment Using Entropy and Network Analysis
Next Article in Special Issue
Systematic Encoding and Shortening of PAC Codes
Previous Article in Journal
Optimal Control of Hydrogen Atom-Like Systems as Thermodynamic Engines in Finite Time
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Performance Analysis of Identification Codes

Lehr- und Forschungseinheit für Nachrichtentechnik, Technische Universität München, 80333 Munich, Germany
*
Author to whom correspondence should be addressed.
Entropy 2020, 22(10), 1067; https://doi.org/10.3390/e22101067
Submission received: 15 July 2020 / Revised: 8 September 2020 / Accepted: 22 September 2020 / Published: 23 September 2020
(This article belongs to the Special Issue Information Theory for Channel Coding)

Abstract

:
In this paper, we analyze the construction of identification codes. Identification codes are based on the question: “Is the message I have just received the one I am interested in?”, as opposed to Shannon’s transmission, where the receiver is interested in not only one, but any, message. The advantage of identification is that it allows rates growing double exponentially in the blocklength at the cost of not being able to decode every message, which might be beneficial in certain applications. We focus on a special identification code construction based on two concatenated Reed-Solomon codes and have a closer look at its implementation, analyzing the trade-offs of identification with respect to transmission and the trade-offs introduced by the computational cost of identification codes.

1. Introduction

Shannon’s transmission [1] is the standard model for communication. For some special tasks, however, more effective models can be used. After Shannon’s transmission theory, there have been some proposals on achieving a better capacity, using schemes that are outside of Shannon’s description. These advanced techniques for communications are sometimes referred to as Post-Shannon communication models. In 1989, Ahlswede and Dueck proposed one such new model for communication: identification [2,3]. The main differences between transmission and identification are the following questions: The receiver in transmission is interested in “What is the message I just received?”, whereas the receiver in identification is merely interested in “Is the message I have just received the one I am interested in?” (Figure 1 and Figure 2). In Shannon’s transmission, there is usually one receiver, and he is ready to decode any codewords in the codebook he receives into messages. However, there can be multiple receivers in identification, and each of them would only be interested in one message of the codebook.
A real-life application of identification can be online sales. Since web platforms track user data in their surfing or product viewing behaviors, they can identify whether the user is interested in a certain product type or company. Various categories are possible for data tracking. We can think of each element in these categories as a receiver of an identification scheme. According to the information of the user’s interests gathered with the help of identification, the platform can use optimized advertising when targeting that user [4].
There are advantages and disadvantages to identification. It allows a negligible amount of overlapping between the decoding sets of codewords. This helps to fit in exponentially more messages [2,3], but the drawback is that we cannot decode every message anymore. The additional error introduced by the overlap can still be made arbitrarily close to zero in the asymptotic case, namely for codewords which have increasingly large blocklengths.
Identification codes can be constructed around transmission codes with the addition of some pre-processing at the sender and some post-processing at the receiver. Rather than associating a codeword and a decoding set to each identification message, or identity, the identification scheme associates a function to each identification message. Given an identity, the identification encoder then picks an input to the function at random, computes the corresponding output and sends the input-output pair to the transmission encoder so that it can be sent through the channel. Given any other identity, the identification decoder first gets an estimate for the input-output pair from the transmission decoder, and then it computes its own output on the received input using the function associated to the given identity. The identity given to the identification decoder is verified if the computed output matches the received output and rejected otherwise. Beyond the errors that can be introduced by the transmission code use to send the input-output pair, an error can occur also when transmission was successful if the two identities provided to the identification encoder and decoder are distinct, but nonetheless produce the same output on the randomly chosen input. In such a case, the identity at the decoder would be erroneously accepted as verified.
The identities given to the identification encoder and decoders are uniquely determined by their associated function. One of the goals of constructing identification codes is to find functions distinct enough so that the probability of erroneously verifying an identity in the manner just described is negligible. These functions turn out to be simply error-correction codes that are used in a different manner than how they are used for the correction or error in transmission schemes. In order to achieve identification capacity, some conditions must be satisfied [5], and the existence of codes satisfying this condition is guaranteed by either probabilistic arguments [3] or by the Gilbert-Varshamov bound ([6] [Section III.B] and [5] [Appendix]). In this paper, we focus on a capacity-achieving construction of identification codes based on the concatenation of two Reed-Solomon codes [5]. The codewords made with the help of the concatenated Reed-Solomon codes are those mapping functions we have mentioned above. The positions of the symbols in the codewords are the inputs, and the outputs are nothing but the symbols themselves.
The paper is structured as follows. In the next section (Section 2), we introduce all the necessary concepts for an implementation of the identification code based on the concatenated Reed-Solomon code. For further details on identification, we refer also to an upcoming survey of identification [7]. In Section 3.3, we describe how we compare such construction to transmission and how the simulation is implemented, with particular focus on the efficiency of the implementation. We also give an analytic formula for how much larger is the growth of number of identities compare to the number of transmission messages. Finally, in Section 4, we discuss the numerical results obtained in the simulation.

2. Preliminaries

Channels are mediums in which the information-carrying codewords are transmitted. Noisy channels disrupt the codewords they are carrying, unless we are using an ideal channel.
Shannon’s transmission is a system where a party communicates messages to another via an often noisy channel. The receiver is interested in any message the sender sends; therefore, the receiver’s motivation can be efficiently stated with the question: ”What is the message?”
To counteract the noise of the channel, the messages are channel-coded into codewords, adding redundancy. Each message is encoded by an encoder before the channel, and, after it passes through the channel, the codeword is decoded by a decoder (Figure 3). The decoder is defined by decoding sets, one for each possible message, which are disjoint (Or at least the size of the intersection is bounded as the blocklength n increases in the case of soft decoding.) If a codeword falls into a decoding set after passing through the channel, the decoder recognizes the message associated to the decoding set and outputs that particular message. In this way, the message has been decoded and transmitted to the sink, which is the communication partner waiting on the receiving end.

2.1. Identification

Identification differs from Shannon’s transmission in how it carries information between two communication partners. This time the receiver is not interested in finding which message i was sent, but only whether one of them, e.g., in message j chosen by the receiver itself, was sent or not. The identification messages can be called identities in order to distinguish them from the transmission messages. Compared to transmission, the role of the decoder then becomes that of a verifier, which either accepts or rejects the received output as being compatible with the chosen message (Figure 4).
Identification can be achieved by adding some pre-processing and post-processing around transmission (Figure 5). These new steps allow for a much larger number of possible identities than the possible transmission messages, while adding a small new kind of error. The idea is that the high number of identities can be beneficial if we can keep the new kind of error below a certain threshold and send it to zero in the asymptotic case. This also allows us to first fix the errors in the channel using a transmission code, thus approximating a noiseless channel, and then code for identification. For this reason, it is enough to study identification for the noiseless channel, which is the case for this work.
The advantage of identification over transmission, given the added complexity, lies in the number of possible identities in an identification scheme. By allowing the decoding sets to collide, the maximum number of identities grows doubly exponentially, while the number of messages in the transmission scheme grows only exponentially. In short, in identification, we allow a small error to happen and we sacrifice decoding for an extremely large number of identities.

2.1.1. Types of Errors

In identification, we have to bound two kinds of errors independently: missed identification, also known as the error of the first kind, and false identification, also known as the error of the second kind.
The missed identification is a regular transmission error, i.e., the codeword i is not recognized as codeword i on the receiver end. This happens when the receiver is interested in the same message i sent by the sender, and when the message is disrupted in the channel in such a way that the receiver does not identify the received output as message i. This error is absent when coding for the noiseless channel with the identification codes that we consider.
The false identification will be our main interest in this paper. If the sender sends the message i and the receiver is only interested in a different message j, the receiver may still identify the message as message j. This is a false identification or a false identification error. There are two possible contributions to this error probability:
  • A transmission error (the same reason as for the missed identification) happens; and the channel gives an output of message j although having received the message i.
  • The scheme uses a randomized identification code (see Section 2.1.2), and the randomly selected codeword lies in the overlapping section of the decoding sets of message i and message j.
Since we only look at identification over the noiseless channel, the first contribution is absent.

2.1.2. Identification Codes

A deterministic code is a sequence of encoding codewords and decoding sets, or better verifiers, { u i , D i } i = 1 , , N for each possible identity. A randomized code is a sequence of encoding probability distributions and verifiers { Q i , D i } i = 1 , , N for each possible message. The randomization is necessary to achieve a high number of identities in the discrete memoryless channel [2], which is in contrast to transmission, where having local randomness does not increase capacity [8]. For this reason, we focus only on the randomized codes. The construction we work on in this paper also makes use of a randomized identification code.
In both cases, the verifier sets are subsets of the possible outputs, namely a verifier set D i is a subset of the Y n alphabet, where an output codeword is recognized as the i-th codeword, if it is an element of D i . The larger the decoding sets are, the more tolerance the code has against transmission errors. The difference between the verifier sets of identification and the decoding sets of transmission is that the do not need to be disjoint, and a priori no restriction is put on the intersection between two distinct verifier sets.
As already mentioned before, we have to treat two types of error independently. Indeed, in the constructions sketched in the introduction, we have seen that an error in the post-processing can happen even if there was no error in the transmission. ( n , N , λ 1 , λ 2 ) randomized identification codes are defined as randomized codes with blocklength n, N total number of identities and λ 1 , λ 2 bounds on the two types of errors. Namely, the codewords on the sender side are selected randomly, using a conditional probability distribution Q such that:
μ 1 i = x n X n Q ( x n | i ) W n ( D i c | x n ) λ 1 , i ,
μ 2 i , j = x n X n Q ( x n | j ) W n ( D i | x n ) λ 2 , i j ,
where Q i X n represents the encoder of the i-th identity, and D j Y n is the verifier set on the receiver side for the j-th identity. μ 1 are the probabilities of the error of the first kind, and μ 2 are the probabilities of the error of the second kind. Note that the error of the first kind concerns only one codeword, where the error of the second kind happens between two codewords. λ 1 is thus a bound on the maximum probability of the error of the first kind, and λ 2 is a bound on the maximum probability of the error of the second kind.
As a final note, we point out that, at least for false identification, the error to be bound must be the maximum and not the average error. It has been proven that bad codes exist if we require only the average of the false identification probability to be small; these codes allow for an infinite identification rate [9].

2.1.3. Identification Codes from Transmission Codes

A common way of constructing identification codes is the way explained in the introduction, by having all the channel errors corrected by a transmission code and performing identification for the almost-noiseless channel provided by the transmission code. In the noiseless channel, then, identification can be done by associating to each identity i a function T i and by sending a single input-output pair ( j , T i ( j ) ) for verification, with j picked uniformly at random using the local randomness highlighted in Figure 4 and Figure 5, through the channel. The receiver, given another identity i , can verify whether the two chosen identities are the same by computing its own output T i ( j ) on the input part j of the received input-output pair ( j , t ) and verifying that the received and computed output are the same ( T i ( j ) = t ?). If the channel is noiseless, then ( j , t ) = ( j , T i ( j ) ) ; this scheme has zero missed identification error (if i = i , then the computed outputs will match) and false identification bounded by the fraction of inputs any two identities map to the same output (if i i , then an incorrect accept will happen only on those j such that T i ( j ) = T i ( j ) ) [3,5]. In case of a noisy channel, the error probability of the transmission code (the probability that ( j , t ) ( j , T i ( j ) ) ) adds to the missed- and false-identification error probabilities of the identification code [2,5]. Arguably, in such cases, the transmission code should have the error probability comparable to the error probability of the identification code. This analysis is, however, left for future work.
Observation 1. A set of N functions { T i } i = 1 , , N from inputs of size M to outputs of size q such that, for any two of these functions, their outputs are equal on at most M d inputs, corresponds to an error correction code of size N, blocklength M, alphabet of size q, and distance d, and vice versa. To each function corresponds a codeword, obtained by concatenating all the possible function outputs in order, as displayed in the example of Figure 6. Similarly, given any correction code, each codeword defines a function, namely the function that map positions (the size of the blocklength) to the symbols of the codeword.
For the sake of clarity, we repeat the identification pre- and post-processing using error-correction codes. Given an error-correction code of size N, blocklength n, alphabet of size q, and minimum Hamming distance d, namely a ( n , N , d ) q error-correction code, an identification code is constructed as follows. The error-correction code is not used to correct error but instead is used in a different manner. The codewords c i of the error-correction code are associated each to an identity i. The identification sender of identity i randomly and uniformly chooses a position j from 1 , 2 , , n , and then sends j and the jth letter c i j to the receiver, using a transmission code if the channel is noisy. The receiver must make a choice on what identity he is interested, say i . Upon receiving j , c i j , the receiver checks whether jth letter of the codeword associated with his interested message i th is c i . Namely, it checks whether c i j = c i j and says “yes/accept i ” if so, or “no/reject i ” otherwise. In the noiseless case, the only possible error is the false identification error (error of the second kind), which occurs only when the receiver is interested in a different message than the one sent, and two codewords c i and c i have the same letter as in the ith position. Thus, the probability of false identification is bounded by
λ 2 1 d n .
Again, we highlight that this is a bound on the false identification error only in the absence of a transmission error.
In order to uniquely refer to the input-output pairs ( j , T i ( j ) ) produced in the pre-processing, we will call j the randomness and T i ( j ) the tag, for reason that should become clear next. For convenience we may then call the functions T i tagging functions. Traditionally, the identities have been called messages [2,3]. However, the identities are not the messages that are sent through the transmission code. Furthermore, transmission can be performed in parallel to identification without trade-off, meaning that the capacity of both can be achieved at the same time [9]. This is because the goal of the pre-processing is to use all the capacity of the channel to send the local randomness j and produce common randomness between the sender and receiver [10]. A small tag size (the size of T i ( j ) or c i j ) that asymptotically does not use any capacity [3] is then enough to allow a rate of identities that grows doubly exponentially in the blocklength. The intuition is that identification is performed by verifying the two tags, the senders and the receivers, of a random challenge. In other works [4,11,12], the randomness has been called a coloring number, the tag a color, and the identity a coloring.
Observation 2. In Reference [5,13], error-correction codes are used to construct constant-weight binary codes that are then used to construct identification codes following the method in Reference [5]. This, at first glance, might seem like a different way of constructing identification codes; however, it is implicitly the same use of error-correction codes presented above. The way the constant-weight binary codes are further encoded in Reference [5] (via the use of the incidence matrix of the binary code), it is an implicit encoding of the randomness-tag pair. Both in Reference [5,13] and in the scheme above, the information that is sent through the channel via the transmission code is a randomness-tag pair.

2.2. Error-Correction Codes

We have seen error-correction codes used for identification. However, their initially intended purpose is to allow the receiver to fix the damage a noisy channel does to the codewords, by adding some redundancy to the original message. This intended purpose has a huge influence on how the encoding is generally performed. The result is that computing the codewords in the usual way (using the generator matrix) and then computing a tag for identification as explained above results in an extremely inefficient implementation. For this reason, we briefly review the background in error-correction codes and Reed-Solomon codes in particular, in order to highlight where we obtain the more efficient implementations.
Linear block codes are error-correction codes over an alphabet such that any linear combination of its codewords is also a codeword. The [ n , k , d ] q codes denote the set of linear codes over an alphabet of size q (commonly omitted for q = 2 ), mapping length k message to blocklength n codewords of minimum Hamming distance d. The size of the a block code is q k . The linearity of the code implies that we can see messages as k dimensional vectors, that we can build a generator matrix G, and that we can compute any codeword for any message via matrix multiplication as
m ̲ · G = c ̲ ,
where m is the message, i.e., the original bitstream, G is the generator matrix of the linear code, and c is the resulting codeword after the encoding process. G will have full rank and the dimensions k × n in order to map messages to codewords uniquely.

2.2.1. Reed-Solomon Codes

Reed-Solomon codes [14] are linear block-codes based on either prime fields or extension fields. The key observation of this subsection is that Reed-Solomon codes naturally define tagging functions before constructing the error-correction codewords. Reed-Solomon codes also achieve the Singleton Bound
d n k + 1
with equality, which means that they are maximum distance separable codes. This is relevant for tag codes because greater scaling in the code distance means a smaller false identification probability in the asymptotic case.
The codewords of Reed-Solomon codes are generated as the evaluation of polynomials on the elements of F q , also called code locators. To each message corresponds a polynomial, where the symbols of the message are the coefficients of the polynomial. So, if we let m ̲ = { m 0 , m 1 , m 2 , m 3 , m 4 , , m k 1 } be the message vector and have β F q , then the evaluation of β is defined as:
T m ( β ) = i = 0 k 1 m i · β i ,
with the convention that 0 0 = 1 (we use the same codes as Reference [5]). For a Reed-Solomon code of length n q , there are n such evaluations. One evaluation gives us one symbol of the codeword, whereas n such evaluations give us the whole codeword. If we let c ̲ = { c 0 , c 1 , c 2 , c 3 , c 4 , , c n 1 } be the codeword and f j F q i , where j shows us the order of the picked code locator, then the codewords are:
c ̲ ( m ) = { i = 0 k 1 m i · f 0 i , i = 0 k 1 m i · f 1 i , i = 0 k 1 m i · f 2 i , , i = 0 k 1 m i · f n 1 i } ,
= { T m ( f 0 ) , T m ( f 1 ) , T m ( f 2 ) , , T m ( f n 1 ) } .
These codewords can be generated from the message m = m 1 m k 1 with the k × n generator matrix:
G = f 0 0 f 1 0 f 2 0 f 3 0 f n 1 0 f 0 1 f 1 1 f 2 1 f 3 1 f n 1 1 f 0 2 f 1 2 f 2 2 f 3 2 f n 1 2 f 0 3 f 1 3 f 2 3 f 3 3 f n 1 3 f 0 k 1 f 1 k 1 f 2 k 1 f 3 k 1 f n 1 k 1 .
Two distinct polynomials of degree less than k agree on at most k inputs; therefore, the distance of the code constructed this way in n k + 1 . We denote with
( n , k ) q R S
the [ n , k , n k + 1 ] q Reed-Solomon block code. We will be interested only in the ( q , k ) q R S codes.
If we use the Reed-Solomon code for identification, then the polynomials T m are the tagging functions, the randomness is j { 0 , , n 1 } , F q are the possible tags, and ( j , T m ( j ) ) is the randomness-tag pair sent through the transmission code.

2.2.2. Concatenated Codes

We will see that using a single Reed-Solomon code as tagging functions does not allow to achieve the identification capacity [5]. A way around this is to concatenate multiple codes [5]. When concatening two codes, we distinguish the two as inner code C i and outer codes C o with the usual convention that the outer code’s symbols are encoded by the inner code. The subscripts “ i ” and “ o ” stand for the inner and the outer code, respectively. Let the inner code be an [ n i , k i , d i ] q i code and the outer code be an [ n o , k o , d o ] q o code, satisfying q i = q o k o . Then, the concatenated error-correction code C i C o is an
[ n c , k c , d c ] q c = [ n i n o , k i k o , d i d o ] q i
code. The alphabet of the concatenated code is the alphabet of the inner code q i . The number of codewords in the concatenated code is the number of codewords in the outer code, q i k i k o = q o k o .
The following is an example, taken from Reference [15], of concatenated codes where each symbol of the outer code is encoded by the inner code, albeit without using block codes (these are, however, not block codes; thus, k c = k i · k o is not valid for this example). C i = { 0120112 , 1202102 , 2100211 , 1201120 } with n i = 7 , q i = 3 , | C i | = 4 and C o = { a d , b c , a c , c c , d b , a b } with n o = 2 , a l p h a b e t A o = { a , b , c , d } ( q o = 4 ), | C o | = 6 . Then:
C i C o = { 0120112 1201120 ,
1202102 2100211 ,
0120112 2100211 ,
2100211 2100211 ,
1201120 1202102 ,
0120112 1202102 } ,
where n = 14 , q = 3 , | C i C o | = 6 . The outer code’s alphabet size and the inner code’s number of codewords in the codebook are the same so that each codeword in the inner code represents one symbol in the alphabet of the outer code, e.g., 0120112 represents a, 1202102 represents b, 2100211 represents c, and 1201120 represents d. This is clear in the codebook of the concatenated code.

2.3. Error-Correction Codes Achieving Identification Capacity

There are three conditions which a capacity-achieving identification code constructed on transmission codes needs to meet. If we are considering a sequence of block codes [ M ( n ) , k ( n ) , d ( n ) ] q ( n ) to construct identification codes, the following conditions are necessary in order to achieve the identification capacity [5] [Definition 9]:
  • The size of the block code, and thus the size of the identification code, must be exponential in the size of the randomness:
    lim n log k ( n ) log M ( n ) 1 ;
  • The size of the tag in bits must be negligible in the size of the randomness in bits:
    lim n log q ( n ) log M ( n ) 0 ;
  • The distance of the code must grow as the blocklength, so that the error of the second kind goes to zero:
    lim n d ( n ) M ( n ) 1 .
Recall that, after computing the tag, the tuple must be channel coded. The first two conditions together guarantee that the entire capacity of the channel is used to send the randomness and that the number of identities grows doubly exponential in the channel capacity. The last condition guarantees that we can send the false identification error to zero.
The Gilbert-Varshamov bound guarantees the existence of such codes, but it is actually possible to explicitly construct, given a capacity achieving code for transmission, identification codes that achieve the identification capacity [5]. One such construction involves the concatenation of two Reed-Solomon codes, where one is based on a prime field and the other is based on an extension field. The construction is defined as:
( q , k , δ ) R S 2 : = ( q , k ) q R S ( q k , q k δ ) q k R S ,
which gives a
[ q k + 1 , k q k δ , ( q k + 1 ) ( q k q k δ + 1 ) ] q
block code. We will refer to it either a the double or concatenated Reed-Solomon code.
To achieve the identification capacity as the blocklenght n grows, q, k must be increasing functions of n satisfying [5] [Proposition 3]
q ( n ) k ( n ) δ ( n ) > 0
( δ can stay constant). These conditions allow the concatenated Reed-Solomon code to satisfy the capacity achieving conditions below. The first condition, Equation (16), can be easily verified. The second condition, Equation (17), is also obvious because q k + 1 will always be bigger and also grow faster than q itself. In particular, these two conditions are met already, by expanding the extension field into the base field for the outer code only. The crucial condition where the inner code is needed is the third one, Equation (18). Applying the code distance of the double Reed-Solomon code construction to the third condition gives:
d M = ( q k + 1 ) ( q k q k δ + 1 ) q k + 1 ( q k ) ( q k q k δ ) q k + 1 ,
q k + 1 k q k q k δ + 1 q k + 1 = 1 k q q δ 1 ,
given the assumption on the scaling of q, k and δ from Equation (21). As we can see in Equation (23), we can satisfy the conditions to achieve identification capacity if we assume that q grows to infinity faster that k.
Identification based on single Reed-Solomon codes is also an option [16]; however, they do not fulfil the conditions to achieve capacity. A full size ( n = q ) single Reed-Solomon code would mean a [ n , · , · ] n block code, which does not allow us to satisfy Equation (17):
lim n log q ( n ) log M ( n ) = lim n log n log n 1 ,
meaning that the tag is too big compared to the randomness. Shorter Reed-Solomon codes ( n < q ) can only make this limit larger.

2.4. Example

We now proceed with an example of how such concatenated tagging functions are built. For simplicity and clarity reasons, we will use the smallest possible code with the parameter set ( 3 , 2 , 1 ) R S 2 = ( 3 , 2 ) 3 R S ( 3 2 , 3 ) 3 2 R S . The concatenated code is an
[ q k + 1 , k q k δ , ( q k + 1 ) ( q k q k δ + 1 ) ] q = [ 27 , 6 , 14 ] 3
error-correction code.
In order to obtain a tag from the concatenated Reed-Solomon code, we first need an identity. For example purposes, we pick the 587 th identity among the 729 total identities. In symbols of the field F 3 2 , this is the string
m ̲ = 587 = 722 F 3 2 ,
where the first string is the number in base 10, while the second is the same number written in base F 3 2 . Now, we can use the numbers 7, 2 and 2 as the orders of the field elements. We can order the elements of F 3 2 = { 0 , α 0 , α 1 , α 2 , α 3 , α 4 , α 5 , α 6 , α 7 } . With this order, we can rewrite our identity as
m ̲ = 722 F 3 2 = ( α 6 α 1 α 1 ) = ( α 6 α α ) ,
where α is the primitive element of F 3 2 . Now, we are ready to compute the tag function for the identity m ̲ . It is time to introduce the generator matrix of the outer code ( 3 2 , 3 ) 3 2 R S with what we have seen in Section 2:
G o = α 0 α 0 α 0 α 0 α 0 α 0 α 0 α 0 α 0 0 α 0 α 1 α 2 α 3 α 4 α 5 α 6 α 7 0 α 0 α 2 α 4 α 6 α 0 α 2 α 4 α 6 .
Multiplying m ̲ with G o , we get the outer codeword
c ̲ ( m ̲ ) = m ̲ · G o
= ( α 6 α α ) · G o
= ( α 6 α 7 α 3 0 0 α 6 α 4 α 3 α 4 ) F 3 2 9 .
Before calculating the final codeword, we need two preparation steps. First, we need to expand the symbols of the first codeword by transforming it’s base field from F 3 2 to F 3 . The reason is that the next Reed-Solomon code ( 3 2 , 3 ) 3 R S is based on field F 3 . The way to do this is through representing the symbols back in their string form. If we apply this to the first codeword in our example, the outer codeword will become:
c ̲ ( m ̲ ) = ( ( 22 ) ( 21 ) ( 12 ) ( 00 ) ( 00 ) ( 22 ) ( 20 ) ( 12 ) ( 20 ) ) F 3 18 .
Now, our codeword is ready to be concatenated with the inner code. But since a concatenation means the symbols will be encoded one by one, instead of encoding only the codeword, we need a larger generator matrix than the regular generator matrix of ( 3 , 2 ) 3 R S . We need the direct sum of this generator matrix with itself by 9 times, as the second preparation step. The specified block sum looks like this:
G B = = 0 8 G i = = 0 8 1 1 1 0 1 2 ,
where G i is the generator matrix of the inner code ( 3 , 2 ) 3 R S . Now, the multiplication
T m ̲ = c ̲ ( m ̲ ) · G B
will yield the final codeword T m ̲ . Every symbol of the first codeword c will be encoded with G i . Since each symbol now has two trits and G i transforms two trits into three trits, we get
T 587 = ( 2 1 0 2 0 1 1 0 2 0 0 0 0 0 0 2 1 0 2 2 2 1 0 2 2 2 2 ) .
At this point, we have the tagging function of the 587 th identity. If, for instance, the corresponding tag of the randomness j = 5 in the i = 587 identity is required, we simply calculate this tagging function as done above and look at position j = 5 . The preprocessing step is completed as we use the concatenated string
( j , T i ( j ) ) = ( 5 , T 587 ( 5 ) ) = ( 5 , 1 )
as encoding of the identity i = 587 , which must now be sent through a channel to the receiver.

3. Comparison of Identification and Transmission

Here, we make a comparison between the identification and the transmission schemes at equal blocklengths. Both the single Reed-Solomon code and the double Reed-Solomon code constructions will be compared to their counterparts in the transmission scheme. Since we are coding for the noiseless channel, the number of bits sent is the size of the transmission scheme. We will thus be comparing the amount of bits used in the construction of the ( j , T i ( j ) ) to the number of identities achieved by the error-correction code used for identification. However, notice that while the identification code will have a certain false identification probability, the error in the transmission code is completely absent. Therefore, part of the comparison is the trade-off between the introduced error and the increase in the number of identities.

3.1. Single Reed-Solomon Code

As already explained, the codewords of an error-correction code represent each identity. Thus there are as many identities as the number of possible codewords. In a ( q , k ) q R S Reed-Solomon code, there are q k codewords and thus that many identities. Therefore, the Single Reed-Solomon construction [16] has q k identification messages. The blocklength and the alphabet size of a ( q , k ) q R S Reed-Solomon code are both q, therefore the transmission ( j , T i ( j ) ) takes q 2 elements. If we use the q 2 elements to send q 2 messages via noiseless transmission, we would achieve a rate
r T = log q 2 n = 2 log q n ,
where n is now the blocklength of the noiseless channel. In comparison, the rate achieved by identification is
r I D = log q k n = k log q n .
Therefore, we have an increase in the rate of
r I D r T = k 2
at the cost of introducing, in the worst case, a false identification error probability of
1 d q = 1 q k + 1 q = k 1 q .
As long as q grows faster than k, in the asymptotic case where they both grow to infinity, we can increase the encoded identification messages and decrease maximum false identification probability λ 2 . This shows that, for the same blocklength, the single Reed-Solomon code construction has polynomially more messages than its transmission counterpart.
By letting k grow with the same order as q, an exponential increase in the rate is still possible [16]; however, as we will see now, the double Reed-Solomon code achieves an increase in the rate that is exponential also in k.

3.2. Double Reed-Solomon Code

We now make the same comparison for the double Reed-Solomon code of Equation (19): we compare the number of identification messages with the size of the randomness/tag pairs. Recall that the double Reed-Solomon code is a
( q , k , δ ) R S 2 [ q k + 1 , k q k δ , ( q k + 1 ) ( q k q k δ + 1 ) ] q .
There are thus q k q k δ identification messages in this construction.
The size of the preprocessed codeword ( j , T i ( j ) ) is computed as follows. This time there are q k + 1 possible randomness values with again q possible tags, for a total of q k + 2 . The rate achieved by transmission is thus
r T = log q k + 2 n = ( k + 2 ) log q n .
In contrast, the rate achieved by identification is
r T = log q k q k δ n = k q ( k δ ) log q n .
The resulting increase in the rate is thus
r I D r T = k q k δ k + 2 q k δ = exp ( k δ ) n log q n = exp ( k δ ) n r T .
The rate of a double Reed-Solomon code construction is thus exponential to the rate achieved with simple transmission.
The trade-off in the false identification error introduced to achieve this rate is given by the distance
d = ( q k + 1 ) ( q k q k δ + 1 ) ( q k ) ( q k q k δ ) q k + 1 k q k q k + 1 δ ,
where we assumed that k > 1 . This gives a maximum false identification probability of
λ 2 1 d q k + 1 = 1 q k + 1 k q k q k + 1 δ q k + 1 ,
= k q + q δ O k q .
Thus, we can still achieve the exponential increase in messages while still sending the error to zero. Notice, however, that the scaling at which the error goes to zero is unchanged and arguably slow compared to the scaling of the amount of identification messages.

3.3. Implementation and Simulation

The major part of our contribution lies in the implementation of the double Reed-Solomon code construction using Sagemath [17]. Coding the steps as presented so far is an inefficient solution. In particular, producing the generator matrices is costly in memory and a waste of computation. Already at small parameters ( 7 , 5 , 2 ) R S 2 the generator matrices become too large to handle. Even storing only one codeword is prohibitive. The only advantage to such a solution would be the instant access to the precomputed tags which is nullified by the impossibility of taking advantage of the doubly exponential growth of the codewords.
Instead, the only way of taking advantage of the doubly exponential growth is to compute each tag on demand, using the Reed-Solomon codes as polynomial evaluations as per the initial definition (see Section 2.2.1), and computing single symbols of each codeword at every transmission of an encoded identity. In the final implementation, a tag is computed as follows:
  • Divide the randomness j, ranging from 0 to q k + 1 1 by q. The quotient j ÷ q shows us which column of the generator matrix of the ( q k , q k δ ) q k R S we need to use. We do not need the other columns to calculate the necessary tag. Using only that column, we will get one symbol t ˜ in the alphabet of size q k . The remainder of the division, j mod q is used later.
  • Expand the symbol t ˜ into k symbols of size q. The list of these q elements will be called the expanded codeword.
  • Find the column - or the code locator - in the generator matrix of ( q , k ) q R S with the index as the remainder j mod q and multiply the expanded codeword with that column scalarly, or simply evaluate it with the picked code locator.
The result of this last evaluation gives us the necessary tag. This method saves us a lot of memory compared to using the generator matrices. The bottleneck of the implementation at this point is listing the elements of the extension field F q k , which renders parameters of the order ( 11 , 8 , 4 ) R S 2 again intractable. Ideally, we would want to index the desired field element as a parameter of the field and obtain the desired element. This technique works on prime fields F q but not on extension fields F q k . On prime fields, F q ( 129 ) returns 129 mod q , which is correct; however, on extension fields, F q k ( 129 ) also returns 129 mod q , which is not correct for q 129 .
Further improvements were obtained by changing the way field elements are generated. The “next()” method of the field class does what its name suggests: it gives the next field element. In this way, we can generate the elements by sequentially producing the next element from the zero element. If, for example, we wanted the 10,234th element of the field, we would start from the 0th element of the field and using the “next()” method 10,234 times to get to the 10,234th element of the field. With this method, the computation time of an element remains feasible up to elements of the order 10 8 , with elements larger that 10 9 taking hours of computation time.
A final improvement was made again, changing the generations method. In Section 2, we see that the primitive element of a field can create the field from scratch by its exponents, where 0 and α 0 would be the first two elements and α q k 2 would be the last element, if we are speaking of the field F q k . Generating the elements with this method resulted in several order of magnitude of improvement in the computation time. With this final method, the boundary of the feasible computations lay at parameters ( 17 , 12 , 6 ) R S 2 for which the computation of a single tag took 1.5 h.
At this point, the achieved number of identities are far beyond the ones achievable with transmission; however, the maximum false identification ratio λ 2 k / q = 12 / 17 60 % is too high to be acceptable. In order to bring the error down, the base field size q must grow much larger than the parameter k. In the final simulations, we decided to keep fixed k = 3 and δ = 2 and simply increased the field size q, studying the performance of only ( q , 3 , 2 ) R S 2 codes. Recall that the increase in rate is
r I D r T = exp ( k δ ) n r T .
By fixing the parameter k, we did not achieve identification capacity, but we still hit an exponential increase in rate compared to transmission. This also allowed for a vast increase in the range of computable values for q and reduced the false identification error as rapidly as we could increase q. The limiting value in this regime is q 10 8 , which provides a false identification error of λ 2 = 10 8 , at the cost of 2.5 h required to calculate one tag. In the next section, we present these results in detail.

4. Results

We used k = 3 and δ = 2 at all times in order for q to grow larger than k. To reach channel capacity, k would also need to grow; however, making both q and k grow together makes the simulation prohibitive, requiring a much greater RAM in the simulation computer.
Regardless of the algorithm used, the time requirement decreases with the complexity, in turn increasing the error of the second kind λ 2 . If a false identification error lower than 10 5 is needed, then with our algorithm this implies a calculation time of at least 1 s. And vice versa; requiring the calculation time to be under 1 ms, we must allow a smaller number of identities and a false identification error of at least 10%. These results are displayed in Figure 7. Simply said, complexity reduces the maximum false identification probability at the cost of calculation time. Note that this calculation time involves everything from building the system with the selected parameters, picking a random identity among all identities and picking the randomness to calculate the corresponding tag.
A similar trade-off exists with the number of identification messages. We saw that there is an exponential relationship between the calculation time of one tag and the total number of identities. When we have a tenfold calculation time of one tag, the exponent of the number of identities becomes tenfold. For examples, the reader can see Figure 8. These results (Note that, in both Figure 8 and Figure 9, the vertical axis expresses the common logarithm of number of identities but not the number of identities themselves. This was the only way we could fit a double exponential growth in a plot.) show us that we can increase the total number of identities exponentially if we can endure a linear increase of the calculation time, depending on the applications.
Another simulation yielded us the relationship between the total number of identities and the error of the second kind. The thing to notice here, is that when the q grows ten-fold, the error of the second kind becomes 10% of the previous value. This is no surprise, since we fixed k to 3 and δ to 2:
λ 2 = 1 d n = 1 ( q k + 1 ) ( q k q k δ + 1 ) q k + 1 ,
= 1 ( q 2 ) ( q 3 q + 1 ) q 4 = 2 q + 1 q 2 3 q 3 + 2 q 4 ,
O ( 2 q ) .
As we see in Equation (49), for large q, the error of the second kind equals, in our example set with ( q , 3 , 2 ) R S 2 , approximately 1 q , which is why the error of the second kind λ 2 and q seem inversely proportional. They are indeed inversely proportional. This brings us to the conclusion that lowering the error probability is expensive and the user of the system needs to prioritize between a slow but mostly accurate system, a fast system which has non-neglectable errors, or a balanced system. Our simulation results can be seen in Figure 9.
We also ran a different simulation: we fix a randomly chosen identity and a randomness value, then we start to pick different identities at random and compare their tags in the same position with the tag of our original identity at the fixed position. We repeat this comparison with 1000 different secondary identities and set the said “false identification ratio” as the number of recognized false identifications over 1000 iterations. This simulation can be seen in Figure 10. It is assuring that no false identification ratio in the simulation exceeds the theoretical limits (Note that in Figure 10, the 3 in the horizontal axis corresponds to a parameter set of ( 3 , 2 , 1 ) R S 2 exceptionally.).
One could make two remarks here. First, that the simulated false identification ratio values are approximately inversely proportional to q. This is no surprise, since the originally-chosen tag is a fixed number from the set { 0 , 1 , 2 , , q 1 } . And yet, the secondary identity’s tag at the same position will also be another number from the same set { 0 , 1 , 2 , , q 1 } . The broad probability that a position possesses a certain value is therefore 1 q , assuming that each possible tag comes up almost the same number of times, which also seems like the case in this last simulation of ours. So, it is no coincidence that we meet the same number in the same position of other identities with the probability 1 q .
The second remark, which is the most important aspect here, is that the simulated false identification ratio is an average error probability, unlike this paper focusing on the maximum error probability.

5. Conclusions

This paper shows that identification can efficiently reach many more messages than transmission at the cost of a manageable additional error and the inability to decode all messages. We can see from the results, that making the maximum probability of the error of the second kind λ 2 smaller also means a trade-off in the computation time, which contributes to the latency of the scheme. So, if one wants smaller errors, our paper shows that one needs a more powerful system to calculate the tags under an acceptable latency. We believe that some improvement in the computational performance is still possible, which is left for future work, together with an analytic computation of the computational complexity of computing the tags.
For future works on identification using tag codes, adding some security element into play is important to protect the information from being eavesdropped. The benefit of protecting a tag code is that we only need to protect the tag part of the codeword because the randomness is uncorrelated with the identity, so it carries no useful information and will not increase the success probability of the eavesdropper.
Furthermore, most of the channel capacity in an identification code is used to produce common randomness on which to compute the tag. In systems where the common randomness is given as a resource, or can be synchronized and produced on demand by other means, the resulting uses of the channel diminish drastically. Then, rather than increasing the size of the messages, an advantage can be achieved by keeping the number of messages comparable and instead exponentially or even doubly exponentially reducing the number of channel uses. In such a scenario we can think of identification as converting the latency of communication into latency of computation, which might become advantageous in applications where the increased number of messages is not the main interest.

Author Contributions

Writing—original draft, S.D., C.D. and R.F. All authors have read and agreed to the published version of the manuscript

Funding

Christian Deppe and Roberto Ferrara were supported by the Bundesministerium für Bildung und Forschung (BMBF) through the grant 16KIS1005.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Shannon, C.E. A mathematical theory of communication. Bell Syst. Tech. J. 1948, 27, 379–423. [Google Scholar] [CrossRef] [Green Version]
  2. Ahlswede, R.; Dueck, G. Identification via channels. IEEE Trans. Inf. Theory 1989, 35, 15–29. [Google Scholar] [CrossRef]
  3. Ahlswede, R.; Dueck, G. Identification in the presence of feedback-a discovery of new capacity formulas. IEEE Trans. Inf. Theory 1989, 35, 30–36. [Google Scholar] [CrossRef] [Green Version]
  4. Boche, H.; Deppe, C. Secure Identification for Wiretap Channels; Robustness, Super-Additivity and Continuity. IEEE Trans. Inf. Forensics Secur. 2018, 13, 1641–1655. [Google Scholar] [CrossRef]
  5. Verdu, S.; Wei, V.K. Explicit construction of optimal constant-weight codes for identification via channels. IEEE Trans. Inf. Theory 1993, 39, 30–36. [Google Scholar] [CrossRef] [Green Version]
  6. Ahlswede, R.; Zhang, Z. New directions in the theory of identification via channels. IEEE Trans. Inf. Theory 1995, 41, 1040–1050. [Google Scholar] [CrossRef] [Green Version]
  7. Ahlswede, A.; Althöfer, I.; Deppe, C.; Tamm, U. Identification and Other Probabilistic Models, Rudolf Ahlswede’s Lectures on Information Theory 6, 1st ed.; Springer: Berlin/Heidelberg, Germany, 2020. [Google Scholar]
  8. Ahlswede, R.; Wolfowitz, J. The structure of capacity functions for compound channels. In Probability and Information Theory; Behara, M., Krickeberg, K., Wolfowitz, J., Eds.; Springer: Berlin/Heidelberg, Germany, 1969; pp. 12–54. [Google Scholar]
  9. Han, T.S.; Verdu, S. New results in the theory of identification via channels. IEEE Trans. Inf. Theory 1992, 38, 14–25. [Google Scholar] [CrossRef]
  10. Ahlswede, R.; Cai, N. Transmission, Identification and Common Randomness Capacities for Wire-Tape Channels with Secure Feedback from the Decoder. In General Theory of Information Transfer and Combinatorics; Springer: Berlin/Heidelberg, Germany, 2006; pp. 258–275. [Google Scholar] [CrossRef] [Green Version]
  11. Boche, H.; Deppe, C. Secure Identification Under Passive Eavesdroppers and Active Jamming Attacks. IEEE Trans. Inf. Forensics Secur. 2019, 14, 472–485. [Google Scholar] [CrossRef]
  12. Boche, H.; Deppe, C.; Winter, A. Secure and Robust Identification via Classical-Quantum Channels. IEEE Trans. Inf. Theory 2019, 65, 6734–6749. [Google Scholar] [CrossRef] [Green Version]
  13. Kurosawa, K.; Yoshida, T. Strongly universal hashing and identification codes via channels. IEEE Trans. Inf. Theory 1999, 45, 2091–2095. [Google Scholar] [CrossRef] [Green Version]
  14. Reed, I.S.; Solomon, G. Polynomial Codes Over Certain Finite Fields. J. Soc. Ind. Appl. Math. 1960, 8, 300–304. [Google Scholar] [CrossRef]
  15. Eswaran, K. Identification via Channels and Constant-Weight Codes. Available online: https://people.eecs.berkeley.edu/~ananth/229BSpr05/Reports/KrishEswaran.pdf (accessed on 23 September 2020).
  16. Moulin, P.; Koetter, R. A framework for the design of good watermark identification codes. In Security, Steganography, and Watermarking of Multimedia Contents VIII; Delp, E.J., III, Wong, P.W., Eds.; International Society for Optics and Photonics, SPIE: Bellingham, WA, USA, 2006; Volume 6072, pp. 565–574. [Google Scholar] [CrossRef] [Green Version]
  17. The Sage Developers. SageMath, the Sage Mathematics Software System (Version 8.1). Available online: https://www.sagemath.org (accessed on 23 September 2020).
Figure 1. An illustration of a transmission receiver interested in any message.
Figure 1. An illustration of a transmission receiver interested in any message.
Entropy 22 01067 g001
Figure 2. An illustration of an identification receiver only interested in one message.
Figure 2. An illustration of an identification receiver only interested in one message.
Entropy 22 01067 g002
Figure 3. General transmission scheme.
Figure 3. General transmission scheme.
Entropy 22 01067 g003
Figure 4. General identification scheme where local randomness at the encoder is necessary to achieve capacity, as opposed to the transmission encoder where it is optional.
Figure 4. General identification scheme where local randomness at the encoder is necessary to achieve capacity, as opposed to the transmission encoder where it is optional.
Entropy 22 01067 g004
Figure 5. The specific class of identification schemes which is the focus of this paper.
Figure 5. The specific class of identification schemes which is the focus of this paper.
Entropy 22 01067 g005
Figure 6. An example of a mapping function T i , which belongs to the identity i. On the right, the 16-digit-long codeword c i obtained by concatenating all the function outputs.
Figure 6. An example of a mapping function T i , which belongs to the identity i. On the right, the 16-digit-long codeword c i obtained by concatenating all the function outputs.
Entropy 22 01067 g006
Figure 7. Trade-off between complexity and reliability.
Figure 7. Trade-off between complexity and reliability.
Entropy 22 01067 g007
Figure 8. The relationship between the number of identities and the calculation time of one tag.
Figure 8. The relationship between the number of identities and the calculation time of one tag.
Entropy 22 01067 g008
Figure 9. The relationship between parameter sets and λ 2 .
Figure 9. The relationship between parameter sets and λ 2 .
Entropy 22 01067 g009
Figure 10. Another simulation by fixing an identity and the randomness.
Figure 10. Another simulation by fixing an identity and the randomness.
Entropy 22 01067 g010

Share and Cite

MDPI and ACS Style

Derebeyoğlu, S.; Deppe, C.; Ferrara, R. Performance Analysis of Identification Codes. Entropy 2020, 22, 1067. https://doi.org/10.3390/e22101067

AMA Style

Derebeyoğlu S, Deppe C, Ferrara R. Performance Analysis of Identification Codes. Entropy. 2020; 22(10):1067. https://doi.org/10.3390/e22101067

Chicago/Turabian Style

Derebeyoğlu, Sencer, Christian Deppe, and Roberto Ferrara. 2020. "Performance Analysis of Identification Codes" Entropy 22, no. 10: 1067. https://doi.org/10.3390/e22101067

APA Style

Derebeyoğlu, S., Deppe, C., & Ferrara, R. (2020). Performance Analysis of Identification Codes. Entropy, 22(10), 1067. https://doi.org/10.3390/e22101067

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop