Next Article in Journal
Wall-Normal Variation of Spanwise Streak Spacing in Turbulent Boundary Layer With Low-to-Moderate Reynolds Number
Next Article in Special Issue
Guessing with Distributed Encoders
Previous Article in Journal
Computing the Partial Correlation of ICA Models for Non-Gaussian Graph Signal Processing
Previous Article in Special Issue
Tight Bounds on the Rényi Entropy via Majorization with Applications to Guessing and Compression
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Detection Games under Fully Active Adversaries

1
Department of Information Engineering and Mathematical Sciences, University of Siena, 53100 Siena, Italy
2
The Andrew and Erna Viterbi Faculty of Electrical Engineering—Israel Institute of Technology Technion City, Haifa 3200003, Israel
*
Author to whom correspondence should be addressed.
Entropy 2019, 21(1), 23; https://doi.org/10.3390/e21010023
Submission received: 23 November 2018 / Revised: 24 December 2018 / Accepted: 25 December 2018 / Published: 29 December 2018

Abstract

:
We study a binary hypothesis testing problem in which a defender must decide whether a test sequence has been drawn from a given memoryless source P 0 , while an attacker strives to impede the correct detection. With respect to previous works, the adversarial setup addressed in this paper considers an attacker who is active under both hypotheses, namely, a fully active attacker, as opposed to a partially active attacker who is active under one hypothesis only. In the fully active setup, the attacker distorts sequences drawn both from P 0 and from an alternative memoryless source P 1 , up to a certain distortion level, which is possibly different under the two hypotheses, to maximize the confusion in distinguishing between the two sources, i.e., to induce both false positive and false negative errors at the detector, also referred to as the defender. We model the defender–attacker interaction as a game and study two versions of this game, the Neyman–Pearson game and the Bayesian game. Our main result is in the characterization of an attack strategy that is asymptotically both dominant (i.e., optimal no matter what the defender’s strategy is) and universal, i.e., independent of P 0 and P 1 . From the analysis of the equilibrium payoff, we also derive the best achievable performance of the defender, by relaxing the requirement on the exponential decay rate of the false positive error probability in the Neyman–Pearson setup and the tradeoff between the error exponents in the Bayesian setup. Such analysis permits characterizing the conditions for the distinguishability of the two sources given the distortion levels.

1. Introduction

Signal processing techniques are routinely applied in the great majority of security-oriented applications. In particular, the detection problem plays a fundamental role in many security-related scenarios, including secure detection and classification, target detection in radar-systems, intrusion detection, biometric-based verification, one-bit watermarking, steganalysis, spam-filtering, and multimedia forensics (see [1]). A unifying characteristic of all these fields is the presence of an adversary explicitly aiming at hindering the detection task, with the consequent necessity of adopting proper models that take into account the interplay between the system designer and the adversary. Specifically, the adopted models should be able to deal, on the one hand, with the uncertainty that the designer has about the attack the system is subject to, and, on the other hand, with the uncertainty of the adversary about the target system (i.e., the system it aims at defeating). Game theory has recently been proposed as a way to study the strategic interaction between the choices made by the system designer and the adversary. Among the fields wherein such an approach has been used, we mention steganalysis [2,3], watermarking [4], intrusion detection [5] and adversarial machine learning [6,7]. With regard to binary detection, game theory and information theory have been combined to address the problem of adversarial detection, especially in the field of digital watermarking (see, for instance, [4,8,9,10]). In all these works, the problem of designing watermarking codes that are robust to intentional attacks is studied as a game between the information hider and the attacker.
An attempt to develop a general theory for the binary hypothesis testing problem in the presence of an adversary is made in [11]. Specifically, in [11], the general problem of binary decision under adversarial conditions is addressed and formulated as a game between two players, the defender and the attacker, which have conflicting goals. Given two discrete memoryless sources, P 0 and P 1 , the goal of the defender is to decide whether a given test sequence has been generated by P 0 (null hypothesis, H 0 ) or P 1 (alternative hypothesis, H 1 ). By adopting the Neyman–Pearson approach, the set of strategies the defender can choose from is the set of decision regions for H 0 ensuring that the false positive error probability is lower than a given threshold. On the other hand, the ultimate goal of the attacker in [11] is to cause a false negative decision, so the attacker acts under H 1 only. In other words, the attacker modifies a sequence generated by P 1 , in attempt to move it into the acceptance region of H 0 . The attacker is subjected to a distortion constraint, which limits his freedom in doing so. Such a struggle between the defender and the attacker is modeled in [11] as a competitive zero-sum game; the asymptotic equilibrium, i.e., the equilibrium when the length of the observed sequence tends to infinity, is derived under the assumption that the defender bases his decision on the analysis of first-order statistics only. In this respect, the analysis conducted in [11] extends the one of [12] to the adversarial scenario. Some variants of this attack-detection game have also been studied; for instance, in [13], the setting is extended to the case where the sources are known to neither the defender nor the attacker, while the training data from both sources are available to both parties. Within this framework, the case where part of the training data available to the defender is corrupted by the attacker has also been studied (see [14]).
The assumption that the attacker is active only under H 1 stems from the observation that in many cases H 0 corresponds to a kind of normal or safe situation and its rejection corresponds to revealing the presence of a threat or that something anomalous is happening. This is the case, for instance, in intrusion detection systems, tampering detection, target detection in radar systems, steganalysis and so on. In these cases, the goal of the attacker is to avoid the possibility that the monitoring system raises an alarm by rejecting the hypothesis H 0 , while having no incentive to act under H 0 . Moreover, in many cases, the attack is not even present under H 0 or it does not have the possibility of manipulating the observations the detector relies on when H 0 holds.
In many other cases, however, it is reasonable to assume that the attacker is active under both hypotheses with the goal of causing both false positive and false negative detection errors. As an example, we may consider the case of a radar target detection system, where the defender wishes to distinguish between the presence and the absence of a target, by considering the presence of a hostile jammer. To maximize the damage caused by his actions, the jammer may decide to work under both the hypotheses: when H 1 holds, to avoid that the defender detects the presence of the target, and in the H 0 case, to increase the number of false alarms inducing a waste of resources deriving from the adoption of possibly expensive countermeasures even when they are not needed. In a completely different scenario, we may consider an image forensic system aiming at deciding whether a certain image has been shot by a given camera, for instance because the image is associated to a crime such as child pornography or terrorism. Even in this case, the attacker may be interested in causing a missed detection event to avoid that he is accused of the crime associated to the picture, or to induce a false alarm to accuse an innocent victim. Other examples come from digital watermarking, where an attacker can be interested in either removing or injecting the watermark from an image or a video, to redistribute the content without any ownership information or in such a way to support a false copyright statement [15], and network intrusion detection systems [16], wherein the adversary may try to both avoid the detection of the intrusion by manipulating a malicious traffic, and to implement an overstimulation attack [17,18] to cause a denial of service failure.
According to the scenario at hand, the attacker may be aware of the hypothesis under which it is operating (hypothesis-aware attacker) or not (hypothesis-unaware attacker). While the setup considered in this paper can be used to address both cases, we focus mainly on the case of an hypothesis-aware attacker, since this represents a worst-case assumption for the defender. In addition, the case of an hypothesis-unaware attacker can be handled as a special case of the more general case of an aware attacker subject to identical constraints under the two hypotheses.
With the above ideas in mind, in this paper, we consider the game-theoretic formulation of the defender–attacker interaction when the attacker acts under both hypotheses. We refer to this scenario as a detection game with a fully active attacker. By contrast, when the attacker acts under hypothesis H 1 only (as in [11,13]), it is referred to as a partially active attacker. As we show, the game in the partially active case turns out to be a special case of the game with fully active, hypothesis-aware, attacker. Accordingly, the hypothesis-aware fully active scenario forms a unified framework that includes the hypothesis-unaware case and the partially active scenario as special cases.
We define and solve two versions of the detection game with fully active attackers, corresponding to two different formulations of the problem: a Neyman–Pearson formulation and a Bayesian formulation. Another difference with respect to [11] is that here the players are allowed to adopt randomized strategies. Specifically, the defender can adopt a randomized decision strategy, while in [11] the defender’s strategies are confined to deterministic decision rules. As for the attack, it consists of the application of a channel, whereas in [11] it is confined to the application of a deterministic function. Moreover, the partially active case of [11] can easily be obtained as a special case of the fully active case considered here. The problem of solving the game and then finding the optimal detector in the adversarial setting is not trivial and may not be possible in general. Thus, we limit the complexity of the problem and make the analysis tractable by confining the decision to depend on a given set of statistics of the observation. Such an assumption, according to which the detector has access to a limited set of empirical statistics of the sequence, is referred to as limited resources assumption (see [12] for an introduction on this terminology). In particular, as already done in related literature [11,13,14], we limit the detection resources to first-order statistics, which are known to be a sufficient statistic for memoryless systems (Section 2.9, [19]). In the setup studied in this paper, the sources are indeed assumed to be memoryless, however one might still be concerned regarding the sufficiency of first-order statistics in our setting, since the attack channel is not assumed memoryless in the first place. Forcing, nonetheless, the defender to rely on first-order statistics is motivated mainly by its simplicity. In addition, the use of first-order statistics is common in a number of application scenarios even if the source under analysis is not memoryless. In image forensics, for instance, several techniques have been proposed which rely on the analysis of the image histogram or a subset of statistics derived from it, e.g., for the detection of contrast enhancement [20] or cut-and-paste [21] processing. As another example, the analysis of statistics derived from the histograms of block-DCT coefficients is often adopted for detecting both single and multiple JPEG compression [22]. More generally, we observe that the assumption of limited resources is reasonable in application scenarios where the detector has a small computational power. Having said that, it should be also emphasized that the analysis in this work can be extended to richer sets of empirical statistics, e.g., higher-order statistics (see the Conclusions Section for a more elaborated discussion on this point).
As a last note, we observe that, although we limit ourselves to memoryless sources, our results can be easily extended to more general models (e.g., Markov sources), as long as a suitable extension of the method of types is available.
One of the main results of this paper is the characterization of an attack strategy that is both dominant (i.e., optimal no matter what the defense strategy is), and universal, i.e., independent of the (unknown) underlying sources. Moreover, this optimal attack strategy turns out to be the same under both hypotheses, thus rendering the distinction between the hypothesis-aware and the hypothesis-unaware scenarios completely inconsequential. In other words, the optimal attack strategy is universal, not only with respect to uncertainty in the source statistics under either hypothesis, but also with respect to the unknown hypothesis in the hypothesis-unaware case. Moreover, the optimal attack is the same for both the Neyman–Pearson and Bayesian games. This result continues to hold also for the partially active case, thus marking a significant difference with respect to previous works [11,13], where the existence of a dominant strategy wasestablished with regard to the defender only.
Some of our results (in particular, the derivation of the equilibrium point for both the Neyman–Pearson and the Bayesian games) have already appeared mostly without proofs in [23]. Here, we provide the full proofs of the main theorems, evaluate the payoff at equilibrium for both the Neyman–Pearson and Bayesian games and include the analysis of the ultimate performance of the games. Specifically, we characterize the so called indistinguishability region (to be defined formally in Section 6), namely the set of the sources for which it is not possible to attain strictly positive exponents for both false positive and false negative probabilities under the Neyman–Pearson and the Bayesian settings. Furthermore, the setup and analysis presented in [23] is extended by considering a more general case in which the maximum allowed distortion levels the attacker may introduce under the two hypotheses are different.
The paper is organized as follows. In Section 2, we establish the notation and introduce the main concepts. In Section 3, we formalize the problem and define the detection game with a fully active adversary for both the Neyman–Pearson and the Bayesian games, and then prove the existence of a dominant and universal attack strategy. The complete analysis of the Neyman–Pearson and Bayesian detection games, namely, the study of the equilibrium point of the game and the computation of the payoff at the equilibrium, are carried out in Section 4 and Section 5, respectively. Finally, Section 6 is devoted to the analysis of the best achievable performance of the defender and the characterization of the source distinguishability.

2. Notation and Definitions

Throughout the paper, random variables are denoted by capital letters and specific realizations are denoted by the corresponding lower case letters. All random variables that denote signals in the system are assumed to have the same finite alphabet, denoted by A . Given a random variable X and a positive integer n, we denote by X = ( X 1 , X 2 , , X n ) , X i A , i = 1 , 2 , , n , a sequence of n independent copies of x. According to the above-mentioned notation rules, a specific realization of X is denoted by x = ( x 1 , x 2 , , x n ) . Sources are denoted by the letter P. Whenever necessary, we subscript P with the name of the relevant random variables: given a random variable x, P X denotes its probability mass function (PMF). Similarly, P X Y denotes the joint PMF of a pair of random variables, ( X , Y ) . For two positive sequences, { a n } and { b n } , the notation a n = · b n stands for exponential equivalence, i.e., lim n 1 / n ln a n / b n = 0 , and a n · b n designates that lim sup n 1 / n ln a n / b n 0 .
For a given real s, we denote [ s ] + = max { s , 0 } . We use notation U ( · ) for the Heaviside step function.
The type of a sequence x A n is defined as the empirical probability distribution P ^ x , that is, the vector { P ^ x ( x ) , x A } of the relative frequencies of the various alphabet symbols in x. A type class T ( x ) is defined as the set of all sequences having the same type as x. When we wish to emphasize the dependence of T ( x ) on P ^ x , we use the notation T ( P ^ x ) . Similarly, given a pair of sequences ( x , y ) , both of length n, the joint type class T ( x , y ) is the set of sequence pairs { ( x , y ) } of length n having the same empirical joint probability distribution (or joint type) as ( x , y ) , P ^ x y , and the conditional type class T ( y | x ) is the set of sequences { y } with P ^ x y = P ^ x y .
Regarding information measures, the entropy associated with P ^ x , which is the empirical entropy of x, is denoted by H ^ x ( X ) . Similarly, H ^ x y ( X , Y ) designates the empirical joint entropy of x and y, and H ^ x y ( X | Y ) is the conditional joint entropy. We denote by D ( P Q ) the Kullback–Leibler (K-L) divergence between two sources, P and Q, with the same alphabet (see [19]).
Finally, we use the letter A to denote an attack channel; accordingly, A ( y | x ) is the conditional probability of the channel output y given the channel input x. Given a permutation-invariant distortion function d : A n × A n I R + (a permutation-invariant distortion function d ( x , y ) is a distortion function that is invariant if the same permutation is applied to both x and y) and a maximum distortion Δ , we define the class C Δ of admissible channels { A ( y | x ) , x , y A n } as those that assign zero probability to every y with d ( x , y ) > n Δ .

2.1. Basics of Game Theory

For the sake of completeness, we introduce some basic definitions and concepts of game theory. A two-player game is defined as a quadruple ( S 1 , S 2 , u 1 , u 2 ) , where S 1 = { s 1 , 1 s 1 , n 1 } and S 2 = { s 2 , 1 s 2 , n 2 } are the sets of strategies from which the first and second player can choose, respectively, and u l ( s 1 , i , s 2 , j ) , l = 1 , 2 , is the payoff of the game for player l, when the first player chooses the strategy s 1 , i and the second one chooses s 2 , j . Each player aims at maximizing its payoff function. A pair of strategies ( s 1 , i , s 2 , j ) is called a profile. When u 1 ( s 1 , i , s 2 , j ) + u 2 ( s 1 , i , s 2 , j ) = 0 , the game is said to be a zero-sum game. For such games, the payoff of the game u ( s 1 , i , s 2 , j ) is usually defined by adopting the perspective of one of the two players: that is, u ( s 1 , i , s 2 , j ) = u 1 ( s 1 , i , s 2 , j ) = u 2 ( s 1 , i , s 2 , j ) if the defender’s perspective is adopted or vice versa. The sets S 1 and S 2 and the payoff functions are assumed known to both players. In addition, we consider strategic games, i.e., games in which the players choose their strategies ahead of time, without knowing the strategy chosen by the opponent.
A common goal in game theory is to determine the existence of equilibrium points, i.e., profiles that in some sense represent a satisfactory choice for both players [24]. The most famous notion of equilibrium is due to Nash [25]. A profile is said to be a Nash equilibrium if no player can improve its payoff by changing its strategy unilaterally.
Despite its popularity, the practical meaning of Nash equilibrium is often unclear, since there is no guarantee that the players will end up playing at the Nash equilibrium. A particular kind of games for which stronger forms of equilibrium exist are the so-called dominance solvable games [24]. The concept of dominance-solvability is directly related to the notion of strict dominance and dominated strategies. In particular, a strategy is said to be strictly dominant for one player if it is the best strategy for this player, i.e., the strategy that maximizes the payoff, no matter what the strategy of the opponent may be. Similarly, we say that a strategy s l , i is strictly dominated by strategy s l , j , if the payoff achieved by player l choosing s l , i is always lower than that obtained by playing s l , j , regardless of the strategy of the other player. Recursive elimination of dominated strategies is a common technique for solving games. In the first step, all the dominated strategies are removed from the set of available strategies, since no rational player (in game theory, a rational player is supposed to act in a way that maximizes its payoff) would ever use them. In this way, a new, smaller game is obtained. At this point, some strategies that were not dominated before, may become dominated in the new, smaller version of the game, and hence are eliminated as well. The process goes on until no dominated strategy exists for either player. A rationalizable equilibrium is any profile which survives the iterated elimination of dominated strategies [26,27]. If at the end of the process only one profile is left, the remaining profile is said to be the only rationalizable equilibrium of the game, which is also the only Nash equilibrium point. Dominance solvable games are easy to analyze since, under the assumption of rational players, we can anticipate that the players will choose the strategies corresponding to the unique rationalizable equilibrium. Another, related, interesting notion of equilibrium is that of dominant equilibrium. A dominant equilibrium is a profile that corresponds to dominant strategies for both players and is the strongest kind of equilibrium that a strategic game may have.

3. Detection Game with Fully Active Attacker

3.1. Problem Formulation

Given two discrete memoryless sources, P 0 and P 1 , defined over a common finite alphabet A , we denote by x = ( x 1 , , x n ) A n a sequence emitted by one of these sources. The sequence x is available to the attacker. Let y = ( y 1 , y 2 , , y n ) A n denote the sequence observed by the defender: when an attack occurs under both H 0 and H 1 , the observed sequence y is obtained as the output of an attack channel fed by x.
In principle, we must distinguish between two cases: in the first, the attacker is aware of the underlying hypothesis (hypothesis-aware attacker), whereas, in the second case, it is not (hypothesis-unaware attacker). In the hypothesis-aware case, the attack strategy is defined by two different conditional probability distributions, i.e., two different attack channels: A 0 ( y | x ) , applied when H 0 holds, and A 1 ( y | x ) , applied under H 1 . Let us denote by Q i ( · ) the PMF of y under H i , i = 0 , 1 . The attack induces the following PMFs on y: Q 0 ( y ) = x P 0 ( x ) A 0 ( y | x ) and Q 1 ( y ) = x P 1 ( x ) A 1 ( y | x ) .
Clearly, in the hypothesis-unaware case, the attacker will apply the same channel under H 0 and H 1 , that is, A 0 = A 1 , and we denote the common attack channel simply by A. Throughout the paper, we focus on the hypothesis-aware case as, in view of this formalism, the hypothesis-unaware case is just a special case. Obviously, the partially active case, where no attack occurs under H 0 can be seen as a degenerate case of the hypothesis-aware fully active one, where A 0 is the identity channel I.
Regarding the defender, we assume a randomized decision strategy, defined by Φ ( H i | y ) , which designates the probability of deciding in favor of H i , i = 0 , 1 , given y. Accordingly, the probability of a false positive (FP) decision error is given by
P FP ( Φ , A 0 ) = y Q 0 ( y ) Φ ( H 1 | y ) ,
and similarly, the false negative (FN) probability assumes the form:
P FN ( Φ , A 1 ) = y Q 1 ( y ) Φ ( H 0 | y ) .
As in [11], due to the limited resources assumption, the defender makes a decision based on first-order empirical statistics of y, which implies that Φ ( · | y ) depends on y only via its type class T ( y ) .
Concerning the attack, to limit the amount of distortion, we assume a distortion constraint. In the hypothesis-aware case, we allow the attacker different distortion levels, Δ 0 and Δ 1 , under H 0 and H 1 , respectively. Then, A 0 C Δ 0 and A 1 C Δ 1 , where, for simplicity, we assume that a common (permutation-invariant) distortion function d ( · , · ) is adopted in both cases.
Figure 1 provides a block diagram of the system with a fully active attacker studied in this paper.

3.2. Definition of the Neyman–Pearson and Bayesian Games

One of the difficulties associated with the fully active setting is that, in the presence of a fully active attacker, both the FP and FN probabilities depend on the attack channels. We therefore consider two different approaches, which lead to different formulations of the detection game: in the first, the detection game is based on the Neyman–Pearson criterion, and, in the second one, the Bayesian approach is adopted.
For the Neyman–Pearson setting, we define the game by assuming that the defender adopts a conservative approach and imposes an FP constraint pertaining to the worst-case attack under H 0 .
Definition 1.
The Neyman–Pearson detection game is a zero-sum, strategic game defined as follows.
  • The set S D of strategies allowed to the defender is the class of randomized decision rules { Φ } that satisfy
    (i) 
    Φ ( H 0 | y ) depends on y only via its type.
    (ii) 
    max A 0 C Δ 0 P F P ( Φ , A 0 ) e n λ for a prescribed constant λ > 0 , independent of n.
  • The set S A of strategies allowed to the attacker is the class of pairs of attack channels ( A 0 , A 1 ) such that A 0 C Δ 0 , A 1 C Δ 1 ; that is, S A = C Δ 0 × C Δ 1 .
  • The payoff of the game is u ( Φ , A 1 ) = P FN ( Φ , A 1 ) ; the attacker is in the quest of maximizing u ( Φ , A 1 ) whereas the defender wishes to minimize it.
In the above definition, we require that the FP probability decays exponentially fast with n, with an exponential rate at least as large as λ . Such a requirement is relatively strong, its main consequence being that the strategy used by the attacker under H 0 is irrelevant, in that the payoff is the same whichever is the channel A 0 C Δ 0 played by the attacker. It is worth observing that, to be more general, we could have defined the problem as a non-zero sum game, where the defender has payoff u D ( ϕ , A 1 ) = P FN ( Φ , A 1 ) , whereas for the attacker we consider a payoff function of the form u A ( ϕ , ( A 0 , A 1 ) ) = β P FP ( Φ , A 0 ) + γ P FN ( Φ , A 1 ) , for some positive constant β and γ . As is made clear in the following section, this non-zero sum version of the game has the same equilibrium strategies of the zero-sum game defined above.
In the case of partially active attack (see the formulation in [23]), the FP probability does not depend on the attack but on the defender only; accordingly, the constraint imposed by the defender in the above formulation becomes P FP ( Φ ) e n λ . Regarding the attacker, we have S A C 0 × C Δ 1 , where C 0 is a singleton that contains the identity channel only.
Another version of the detection game is defined by assuming that the defender follows a less conservative approach, that is, the Bayesian approach, and he tries to minimize a particular Bayes risk.
Definition 2.
The Bayesian detection game is a zero-sum, strategic game defined as follow.
  • The set S D of strategies allowed to the defender is the class of the randomized decision rules { Φ } where Φ ( H 0 | y ) depends on y only via its type.
  • The set S A of strategies allowed to the attacker is S A = C Δ 0 × C Δ 1 .
  • The payoff of the game is
    u ( Φ , ( A 0 , A 1 ) ) = P FN ( Φ , A 1 ) + e a n P F P ( Φ , A 0 ) ,
    for some constant a, independent of n.
We observe that, in the definition of the payoff, the parameter a controls the tradeoff between the two terms in the exponential scale; whenever possible, the optimal defense strategy is expected to yield error exponents that differ exactly by a, so as to balance the contributions of the two terms of Equation (3).
Notice also that, by defining the payoff as in Equation (3), we are implicitly considering for the defender only the strategies Φ ( · | y ) such that P FP ( Φ , A 0 ) · e a n . In fact, any strategy that does not satisfy this inequality yields a payoff u > 1 that cannot be optimal, as it can be improved by always deciding in favor of H 0 regardless of y ( u = 1 ).
As in [11], we focus on the asymptotic behavior of the game as n tends to infinity. In particular, we are interested in the exponents of the FP and FN probabilities, defined as follows:
ε FP = lim sup n 1 n ln P FP ( Φ , A 0 ) ; ε FN = lim sup n 1 n ln P FN ( Φ , A 1 ) .
We say that a strategy is asymptotically optimal (or dominant) if it is optimal (dominant) with respect to the asymptotic exponential decay rate (or the exponent, for short) of the payoff.

3.3. Asymptotically Dominant and Universal Attack

In this subsection, we characterize an attack channel that, for both games, is asymptotically dominant and universal, in the sense of being independent of the unknown underlying sources. This result paves the way to the solution of the two games.
Let u denote a generic payoff function of the form
u = γ P FN ( Φ , A 1 ) + β P FP ( Φ , A 0 ) ,
where β and γ are given positive constants, possibly dependent on n.
We notice that the payoff of the Neyman–Pearson and Bayesian games defined in the previous section can be obtained as particular cases: specifically, γ = 1 and β = 0 for the Neyman–Pearson game and γ = 1 and β = e a n for the Bayesian one.
Theorem 1.
Let c n ( x ) denote the reciprocal of the total number of conditional type classes { T ( y | x ) } that satisfy the constraint d ( x , y ) n Δ for a given Δ > 0 , namely, admissible conditional type classes (from the method of the types, it is known that 1 c n ( x ) ( n + 1 ) | A | · ( | A | 1 ) for any x [19]).
Define:
A Δ * ( y | x ) = c n ( x ) | T ( y | x ) | d ( x , y ) n Δ 0 e l s e w h e r e .
Among all pairs of channels ( A 0 , A 1 ) S A , the pair ( A Δ 0 * , A Δ 1 * ) minimizes the asymptotic exponent of u for every P 0 and P 1 , every γ , β 0 and every permutation-invariant Φ ( H 0 | · ) .
Proof. 
We first focus on the attack under H 1 and therefore on the FN probability.
Consider an arbitrary channel A 1 C Δ 1 . Let Π : A n A n denote a permutation operator that permutes any member of A n according to a given permutation matrix and let
A Π ( y | x ) = A 1 ( Π y | Π x ) .
Since the distortion function is assumed to be permutation-invariant, the channel A Π ( y | x ) introduces the same distortion as A 1 , and hence it satisfies the distortion constraint. Due to the memorylessness of P 1 and the assumption that Φ ( H 0 | · ) belongs to S D (i.e., that Φ ( H 0 | · ) depends on the observed sequence via its type), both P 1 ( y ) and Φ ( H 0 | y ) are invariant to permutations on y. Then, we have:
P FN ( Φ , A Π ) = x , y P 1 ( x ) A Π ( y | x ) Φ ( H 0 | y ) = x , y P 1 ( x ) A 1 ( Π y | Π x ) Φ ( H 0 | y ) = x , y P 1 ( Π x ) A 1 ( Π y | Π x ) Φ ( H 0 | Π y ) = x , y P 1 ( x ) A 1 ( y | x ) Φ ( H 0 | y ) = P FN ( Φ , A 1 ) ,
thus P FN ( Φ , A 1 ) = P FN ( Φ , A ¯ ) , where we have defined
A ¯ ( y | x ) = 1 n ! Π A Π ( y | x ) = 1 n ! Π A 1 ( Π y | Π x ) ,
which also introduces the same distortion as A 1 . Now, notice that this channel assigns the same conditional probability to all sequences in the same conditional type class T ( y | x ) . To see why this is true, we observe that any sequence y T ( y | x ) can be seen as being obtained from y through the application of a permutation Π , which leaves x unaltered. Then, we have:
A ¯ ( y | x ) = A ¯ ( Π y | Π x ) = 1 n ! Π A 1 ( Π ( Π y ) | Π ( Π x ) ) = 1 n ! Π A 1 ( Π y | Π x ) = A ¯ ( y | x ) .
Therefore, since the probability assigned by A ¯ to the sequences in T ( y | x ) is surely less than or equal to 1, we argue that
A ¯ ( y | x ) · 1 | T ( y | x ) | d ( x , y ) n Δ 0 elsewhere = A Δ 1 * ( y | x ) c n ( x ) ( n + 1 ) | A | · ( | A | 1 ) A Δ 1 * ( y | x ) ,
which implies that, P FN ( Φ , A ¯ ) · ( n + 1 ) | A | · ( | A | 1 ) P FN ( A Δ 1 * , Φ ) .
Then,
P FN ( Φ , A 1 ) · ( n + 1 ) | A | · ( | A | 1 ) P FN ( A Δ 1 * , Φ ) ,
or equivalently
P FN ( Φ , A Δ 1 * ) · ( n + 1 ) | A | · ( | A | 1 ) P FN ( A 1 , Φ ) .
We conclude that A Δ 1 * minimizes the error exponent of P FN ( Φ , A 1 ) across all channels in C Δ 1 and for every Φ S D , regardless of P 1 .
A similar argument applies to the FP probability for the derivation of the optimal channel under H 0 ; that is, from the memorylessness of P 0 and the permutation-invariance of Φ ( H 1 | · ) , we have:
P FP ( Φ , A Δ 0 * ) ( n + 1 ) | A | · ( | A | 1 ) P FP ( A 0 , Φ ) ,
for every A 0 C Δ 0 . Accordingly, A Δ 0 * minimizes the error exponent of P FP ( Φ , A 0 ) . We then have:
γ P FN ( Φ , A 1 ) + β P FP ( Φ , A 0 ) ( n + 1 ) | A | · ( | A | 1 ) ( γ P FN ( Φ , A Δ 1 * ) + β P FP ( Φ , A Δ 0 * ) ) γ P FN ( Φ , A Δ 1 * ) + β P FP ( Φ , A Δ 0 * ) ,
for every A 0 C Δ 0 and A 1 C Δ 1 . Notice that, since the asymptotic equality is defined in the exponential scale, Equation (15) holds no matter what the values of β and γ are, including values that depend on n. Hence, the pair of channels ( A Δ 0 * , A Δ 1 * ) minimizes the asymptotic exponent of u for any permutation-invariant decision rule Φ ( H 0 | · ) and for any γ , β 0 . ☐
According to Theorem 1, for every zero-sum game with payoff function of the form in Equation (5), if Φ is permutation-invariant, the pair of attack channels which is the most favorable to the attacker is ( A Δ 0 * , A Δ 1 * ) , which does not depend on Φ . Then, the optimal attack strategy ( A Δ 0 * , A Δ 1 * ) is dominant. The intuition behind the attack channel in Equation (6) is illustrated in Figure 2 and explained below. Given x, generated by a source P X , the set { y : d ( x , y ) n Δ } corresponds to a set of conditional type classes (for a permutation-invariant distortion function, d ( x , y ) is the same for every y T ( y | x ) ). We say that a conditional type class is admissible if it belongs to this set. Then, to generate y which causes a detection error with the prescribed maximum allowed distortion, the attacker cannot do any better than randomly selecting an admissible conditional type class according to the uniform distribution and then choosing at random y within this conditional type class. Since the number of conditional type classes is only polynomial in n, the random choice of the conditional type class does not affect the exponent of the error probabilities; besides, since the decision is the same for all sequences within the same conditional type class, the choice of y within that conditional type class is immaterial.
As an additional result, Theorem 1 states that, whenever an adversary aims at maximizing a payoff function of the form Equation (5), and as long as the defense strategy is confined to the analysis of the first-order statistics, the (asymptotically) optimal attack strategy is universal with respect to the sources P 0 and P 1 , i.e., it depends neither on P 0 nor on P 1 .
We observe that, if Δ 0 = Δ 1 = Δ , the optimal attack consists of applying the same channel A Δ * regardless of the underlying hypothesis and then the optimal attack strategy is fully-universal: the attacker needs to know neither the sources ( P 0 and P 1 ), nor the underlying hypothesis. In this case, it becomes immaterial whether the attacker is aware or unaware of the true hypothesis. As a consequence of this property, in the hypothesis-unaware case, when the attacker applies the same channel under both hypotheses, subject to a fixed maximum distortion Δ , the optimal channel remains A Δ * .
According to Theorem 1, for the partially active case, there exists an (asymptotically) dominant and universal attack channel. This result marks a considerable difference with respect to the results of [11], where the optimal deterministic attack function is found by using the rationalizability argument, that is, by exploiting the existence of a dominant defense strategy, and it is hence neither dominant nor universal.
Finally, we point out that, similar to [11], although the sources are memoryless and only first-order statistics are used by the defender, the output probability distributions induced by the attack channel A Δ * , namely Q 0 and Q 1 , are not necessarily memoryless.

4. The Neyman–Pearson Detection Game

In this section, we study the detection game with a fully active attacker in the Neyman–Pearson setup as defined in Definition 1. From the analysis of Section 3.3, we already know that there exists a dominant attack strategy. Regarding the defender, we determine the asymptotically optimal strategy regardless of the dominant pair of attack channels; in particular, as shown in Lemma 1, an asymptotically dominant defense strategy can be derived from a detailed analysis of the FP constraint. Consequently, the Neyman–Pearson detection game has a dominant equilibrium.

4.1. Optimal Detection and Game Equilibrium

The following lemma characterizes the optimal detection strategy in the Neyman–Pearson setting.
Lemma 1.
For the Neyman–Pearson game of Definition 1, the defense strategy
Φ * ( H 1 | y ) = exp n λ min x : d ( x , y ) n Δ 0 D ( P ^ x P 0 ) + ,
is asymptotically dominant for the defender.
In the special case Δ 0 = 0 , Φ * ( H 1 | y ) in Equation (16) is (asymptotically) equivalent to the Hoeffding test for non-adversarial hypothesis testing [28].
The proof of Lemma 1 appears in Appendix A.1.
We point out that, when the attacker is partially active, it is known from [23] that the optimal defense strategy is
Φ * ( H 1 | y ) = exp n λ D ( P ^ y P 0 ) + ,
which is in line with the one in [11] (Lemma 1), where the class of defense strategies is confined to deterministic decision rules.
Intuitively, the extension from Equations (16) and (17) is explained as follows. In the case of fully active attacker, the defender is subject to a constraint on the maximum FP probability over S A , that is, the set of the admissible channels A C Δ 0 (see Definition 1). From the analysis of Section 3.3, channel A Δ 0 * minimizes the FP exponent over this set. To satisfy the constraint for a given sequence y, the defender must handle the worst-case value (i.e., the minimum) of D ( P ^ x P 0 ) over all the type classes T ( x | y ) which satisfy the distortion constraint, or equivalently, all the sequences x such that d ( x , y ) n Δ 0 .
According to Lemma 1, the best defense strategy is asymptotically dominant. In addition, since Φ * depends on P 0 only, and not on P 1 , it is referred to as semi-universal.
Concerning the attacker, since the payoff is a special case of Equation (5) with γ = 1 and β = 0 , the optimal pair of attack channels is given by Theorem 1 and corresponds to ( A Δ 0 * , A Δ 1 * ) .
The following comment is in order. Since the payoff of the game is defined in terms of the FN probability only, it is independent of A 0 C Δ 0 . Furthermore, since the defender adopts a conservative approach to guarantee the FP constraint for every A 0 , the constraint is satisfied for every A 0 and therefore all channel pairs of the form ( A 0 , A Δ 1 * ) , A 0 S A , are equivalent in terms of the payoff. Accordingly, in the hypothesis-aware case, the attacker can employ any admissible channel under H 0 . In the Neyman–Pearson setting, the sole fact that the attacker is active under H 0 forces the defender to take countermeasures that make the choice of A 0 immaterial.
Due to the existence of dominant strategies for both players, we can immediately state the following theorem.
Theorem 2.
Consider the Neyman–Pearson detection game of Definition 1. Let Φ * and ( A Δ 0 * , A Δ 1 * ) be the strategies defined in Lemma 1 and Theorem 1, respectively. The profile ( Φ * , ( A Δ 0 * , A Δ 1 * ) ) is an asymptotically dominant equilibrium of the game.

4.2. Payoff at the Equilibrium

In this section, we derive the payoff of the Neyman–Pearson game at the equilibrium of Theorem 2. To do this, we assume an additive distortion function, i.e., d ( x , y ) = i = 1 n d ( x i , y i ) . In this case, d ( x , y ) can be expressed as i j n x y ( i , j ) d ( i , j ) , where n x y ( i , j ) = n P ^ x y ( i , j ) denotes the number of occurrences of the pair ( i , j ) A 2 in ( x , y ) . Therefore, the distortion constraint regarding A 0 can be rewritten as ( i , j ) A 2 P ^ x y ( i , j ) d ( i , j ) Δ 0 . A similar formulation holds for A 1 .
Let us define
D ˜ Δ n ( P ^ y , P ) = min { P ^ x | y : E x y d ( X , Y ) Δ } D ( P ^ x P ) ,
where E x y denotes the empirical expectation, defined as
E x y d ( X , Y ) = ( i , j ) A 2 P ^ x y ( i , j ) d ( i , j )
and the minimization is carried out for a given P ^ y . Accordingly, the strategy in Equation (16) can be rewritten as
Φ * ( H 1 | y ) = exp n λ D ˜ Δ 0 n ( P ^ y P 0 ) + .
When n , D ˜ Δ n becomes (due to the density of rational numbers on the real line, the admissibility set in Equation (18) is dense in that of Equation (21); since the divergence functional is continuous, the sequence { D ˜ Δ n ( P ^ y , P ) } n 1 tends to D ˜ Δ ( P Y , P ) as n )
D ˜ Δ ( P Y , P ) = min { P X | Y : E X Y d ( X , Y ) Δ } D ( P X P ) ,
where E X Y denotes expectation with respect to P X Y .
The definition in Equation (21) can be stated for any PMF P Y in the probability simplex in R | A | . Note that the minimization problem in Equation (21) has a unique solution as it is a convex program.
The function D ˜ Δ has an important role in the remaining part of the paper, especially in the characterization of the asymptotic behavior of the games. To draw a parallelism, D ˜ Δ plays a role similar to that of the Kullback–Leibler divergence D in classical detection theory for the non-adversarial case.
The basic properties of the functional D ˜ Δ ( P Y , P ) are the following: (i) it is continuous in P Y ; and (ii) it has convex level sets, i.e., the set { P Y : D ˜ Δ ( P Y , P ) t } is convex for every t 0 . Point (ii) is a consequence of the following property, which is useful for proving some of the results in the sequel (in particular, Theorems 3, 7 and 8).
Property 1.
The function D ˜ Δ ( P Y , P ) is convex in P Y for every fixed P.
The proof follows from the convexity of the divergence functional (see Appendix A.2).
Using the above definitions, the equilibrium payoff is given by the following theorem:
Theorem 3.
Let the Neyman–Pearson detection game be as in Definition 1. Let ( Φ * , ( A Δ 0 * , A Δ 1 * ) ) be the equilibrium profile of Theorem 2. Then,
ε FN ( λ ) = lim n 1 n ln P FN ( Φ * , A Δ 1 * ) = min P Y : D ˜ Δ 0 ( P Y , P 0 ) λ D ˜ Δ 1 ( P Y , P 1 ) .
In Equation (22), we made explicit the dependence on the parameter λ in the notation of the error exponent, since this is useful in the sequel.
The proof of Theorem 3, which appears in Appendix A.3, is based on Sanov’s theorem [29,30], by exploiting the compactness of the set { P Y : D ˜ Δ 0 ( P Y , P 0 ) λ } .
From Theorem 3 it follows that ε FN ( λ ) = 0 whenever there exists a PMF P Y inside the set { P Y : D ˜ Δ 0 ( P Y , P 0 ) λ } with Δ 1 -limited expected distortion from P 1 . When this condition does not hold, P FN ( Φ * , A Δ 1 * ) 0 exponentially rapidly.
For a partially active attacker, the error exponent in Equation (22) becomes
ε FN ( λ ) = min P Y : D ( P Y , P 0 ) λ D ˜ Δ 1 ( P Y , P 1 ) .
It can be shown that the error exponent in Equation (23) is the same as the error exponent of Theorem 2 in [11] (and Theorem 2 in [31]), where deterministic strategies are considered for both the defender and the attacker. Such equivalence can be explained as follows. As already pointed, the optimal defense strategy in Equation (17) and the deterministic rule found in [11] are asymptotically equivalent (see the discussion immediately after Lemma 1). Concerning the attacker, even in the more general setup (with randomized strategies) considered here, an asymptotically optimal attack could be derived as in [11], that is, by considering the best response to the dominant defense strategy in [11]. Such attack consists of minimizing the divergence with respect to P 0 , namely D ( P ^ y | | P 0 ) , over all the admissible sequences y , and then is deterministic. Therefore, concerning the partially active case, the asymptotic behavior of the game is equivalent to the one in [11]. The main difference between the setup in [11] and the more general one addressed in this paper relies on the kind of game equilibrium, which is stronger here (namely, a dominant equilibrium) due to the existence of dominant strategies for both the defender and the attacker, rather than for the defender only.
When the distortion function d is a metric, we can state the following result, whose proof appears in Appendix A.4.
Theorem 4.
When the distortion function d is a metric, Equation (22) can be rephrased as
ε FN ( λ ) = min P Y : D ( P Y P 0 ) λ D ˜ Δ 0 + Δ 1 ( P Y , P 1 ) .
Comparing Equations (23) and (24) is insightful for understanding the difference between the fully active and partially active cases. Specifically, the FN error exponents of both cases are the same when the distortion under H 1 in the partially active case is Δ 0 + Δ 1 (instead of Δ 1 ).
When d is not a metric, Equation (24) is only an upper bound on ε FN ( λ ) , as can be seen from the proof of Theorem 4. Accordingly, in the general case (d is not a metric), applying distortion levels Δ 0 and Δ 1 to sequences from, respectively, H 0 and H 1 (in the fully active setup) is more favorable to the attacker with respect to applying a distortion Δ 0 + Δ 1 to sequences from H 0 only (in the partially active setup).

5. The Bayesian Detection Game

In this section, we study the Bayesian game (Definition 2). In contrast to the Neyman–Pearson game, in the Bayesian game, the optimal defense strategy is found by assuming that the strategy played by the attacker, namely the optimal pair of channels ( A 0 * , A 1 * ) of Theorem 1, is known to the defender, that is, by exploiting the rationalizability argument (see Section 2.1). Accordingly, the resulting optimal strategy is not dominant, thus the associated equilibrium is weaker compared to that of the Neyman–Pearson game.

5.1. Optimal Defense and Game Equilibrium

Since the payoff in Equation (3) is a special case of Equation (5) with γ = 1 and β = e a n , for any defense strategy Φ S D , the asymptotically optimal attack channels under H 0 and H 1 are given by Theorem 1, and correspond to the pair ( A Δ 0 * , A Δ 1 * ) . Then, we can determine the best defense strategy by assuming that the attacker will play ( A Δ 0 * , A Δ 1 * ) and evaluating the best response of the defender to this pair of channels.
Our solution for the Bayesian detection game is given in the following theorem, whose proof appears in Appendix B.1.
Theorem 5.
Consider the Bayesian detection game of Definition 2. Let Q 0 * ( y ) and Q 1 * ( y ) be the probability distributions induced by channels A Δ 0 * and A Δ 1 * , respectively.
Then,
Φ # ( H 1 | y ) = U 1 n log Q 1 * ( y ) Q 0 * ( y ) a
is the optimal defense strategy.
If, in addition, the distortion measure is additive, the defense strategy
Φ ( H 1 | y ) = U D ˜ Δ 0 n ( P ^ y , P 0 ) D ˜ Δ 1 n ( P ^ y , P 1 ) a
is asymptotically optimal.
It is useful to provide the asymptotically optimal strategy, Φ , in addition to the optimal one, Φ # , for the following reason: while Φ # requires the non-trivial computation of the two probabilities Q 1 ( y ) and Q 0 ( y ) , the strategy Φ , which leads to the same payoff asymptotically, is easier to implement because of its single-letter form.
According to the asymptotically optimal defense strategy defined by the above theorem, given the observed sequence y, the decision is made by comparing the value of the difference D ˜ Δ 0 n ( P ^ y | P 0 ) D ˜ Δ 1 n ( P ^ y | P 1 ) against the threshold. Then, the functions D ˜ Δ 0 n and D ˜ Δ 1 n take the role of the K-L divergence functions when the likelihood ratio test is performed in the non-adversarial case [19].
We now state the following theorem.
Theorem 6.
Consider the Bayesian game of Definition 2. Let ( A Δ 0 * , A Δ 1 * ) be the attack strategy of Theorem 1 and let Φ # and Φ be the defense strategies defined, respectively, in Equations (25) and (26). The profiles ( Φ # , ( A Δ 0 * , A Δ 1 * ) ) and ( Φ , ( A Δ 0 * , A Δ 1 * ) ) are asymptotic rationalizable equilibria of the game.
The analysis in this section can be easily generalized to any payoff function defined as in Equation (5), i.e., for any γ , β 0 .
Finally, we observe that the equilibrium found in the Bayesian case (namely, a rationalizable equilibrium) is weaker with respect to the equilibrium derived for the Neyman–Pearson game (namely, a dominant equilibrium) is a consequence of the fact that the Bayesian game is defined in a less restrictive manner than the Neyman–Pearson game. This is due to the conservative approach adopted in the latter: while in the Bayesian game the defender cares about both FP and FN probabilities and their tradeoff, in the Neymam–Pearson game the defender does not care about the value of the FP probability provided that its exponent is larger than λ , which is automatically guaranteed by restricting the set of strategies. This restriction simplifies the game so that a dominant strategy can be found for the restricted game.

5.2. Equilibrium Payoff

We now derive the equilibrium payoff of the Bayesian game. As in the Neyman–Pearson game, we assume an additive distortion measure. For simplicity, we focus on the asymptotically optimal defense strategy Φ . We have the following theorem.
Theorem 7.
Let the Bayesian detection game be as in Definition 2. Let ( Φ , ( A Δ 0 * , A Δ 1 * ) ) be the equilibrium profile of Theorem 6. The asymptotic exponential rate of the equilibrium Bayes payoff u is given by
lim n 1 n ln u ( Φ , ( A Δ 0 * , A Δ 1 * ) ) = min P Y max D ˜ Δ 1 ( P Y , P 1 ) , ( D ˜ Δ 0 ( P Y , P 0 ) a ) .
The proof appears in Appendix B.2.
According to Theorem 7, the asymptotic exponent of u is zero if there exists a PMF P Y * with Δ 1 -limited expected distortion from P 1 such that D ˜ Δ 0 ( P Y * , P 0 ) a . Therefore, when we focus on the case of zero asymptotic exponent of the payoff, the parameter a plays a role similar to λ in the Neyman–Pearson game. By further inspecting the exponent expressions of Theorems 7 and 3, we observe that, when a = λ , the exponent in Equation (27) is smaller than or equal to the one in Equation (22), where equality holds only when both Equations (22) and (27) vanish. However, comparing these two cases in the general case is difficult because of the different definition of the payoff functions and, in particular, the different role taken by the parameters λ and a. In the Neyman–Pearson game, in fact, the payoff corresponds to the FN probability and is not affected by the value of the FP probability, provided that its exponent is larger than λ ; in this way, the ratio between FP and FN error exponent at the equilibrium is generally smaller than λ (a part for the case in which the asymptotic exponent of the payoff is zero). In the Bayesian case, the payoff is a weighted combination of the two types of errors and then the term with the largest exponent is the dominating term, namely, the one which determines the asymptotic behavior; in this case, the parameter a determines the exact tradeoff between the FP and FN exponent in the equilibrium payoff.

6. Source Distinguishability

In this section, we investigate the performance of the Neyman–Pearson and Bayesian games as a function of λ and a, respectively. From the expressions of the equilibrium payoff exponents, it is clear that increase (and then the equilibrium payoffs of the games decrease) as λ and a decrease, respectively. In particular, by letting λ = 0 and a = 0 , we obtain the largest achievable exponents of both games, which correspond to the best achievable performance for the defender. Therefore, we say that two sources are distinguishable under the Neyman–Pearson (respectively, Bayesian) setting, if there exists a value of λ (respectively, α ) such that the FP and FN exponents at the equilibrium of the game are simultaneously strictly positive. When such a condition does not hold, we say that the sources are indistinguishable. Specifically, in this section, we characterize, under both the Neyman–Pearson and the Bayesian settings, the indistinguishability region, defined as the set of the alternative sources that cannot be distinguished from a given source P 0 , given the attack distortion levels Δ 0 and Δ 1 . Although each game has a different asymptotic behavior, we show that the indistinguishability regions in the Neyman–Pearson and the Bayesian settings are the same. The study of the distinguishability between the sources under adversarial conditions, performed in this section, in a way extends the Chernoff–Stein lemma [19] to the adversarial setup (see [31]).
We start by proving the following result for the Neyman–Pearson game.
Theorem 8.
Given two memoryless sources P 0 and P 1 and distortion levels Δ 0 and Δ 1 , the maximum achievable FN exponent for the Neyman–Pearson game is:
lim λ 0 ε FN ( λ ) = ε FN ( 0 ) = min { P Y | X : E X Y d ( X , Y ) Δ 0 , ( P X Y ) X = P 0 } D ˜ Δ 1 ( P Y , P 1 ) ,
where ε FN ( λ ) is as in Theorem 3.
The theorem is an immediate consequence of the continuity of ε FN ( λ ) as λ 0 + , which follows by the continuity of D ˜ Δ with respect to P Y and the density of the set { P Y : D ˜ Δ 0 ( P Y , P 0 ) λ } in { P Y : D ˜ Δ 0 ( P Y , P 0 ) = 0 } as λ 0 + (It holds true from Property 1).
We notice that, if Δ 0 = Δ 1 = 0 , there is only an admissible point in the set in Equation (28), for which P Y = P 0 ; then, ε FN ( 0 ) = D ( P 0 | | P 1 ) , which corresponds to the best achievable FN exponent known from the classical literature for the non-adversarial case (Stein lemma [19] (Theorem 11.8.3)).
Regarding the Bayesian setting, we have the following theorem, the proof of which appears in Appendix C.1.
Theorem 9.
Given two memoryless sources P 0 and P 1 and distortion levels Δ 0 and Δ 1 , the maximum achievable exponent of the equilibrium Bayes payoff is
lim a 0 lim n 1 n ln u ( Φ , ( A Δ 0 * , A Δ 1 * ) ) = min P Y max D ˜ Δ 1 ( P Y , P 1 ) , D ˜ Δ 0 ( P Y , P 0 ) ,
where the inner limit at the left hand side is as defined in Theorem 7.
Since D ˜ Δ 1 ( P Y , P 1 ) , and similarly D ˜ Δ 0 ( P Y , P 0 ) , are convex functions of P Y , and reach their minimum in P 1 , respectively P 0 (the fact that D ˜ Δ 0 ( D ˜ Δ 1 ) is 0 in a Δ 0 -limited ( Δ 1 -limited) neighborhood of P 0 ( P 1 ), and not just in P 0 ( P 1 ), does not affect the argument), the minimum over P Y of the maximum between these quantities (right-hand side of Equation (29)) is attained when D ˜ Δ 1 ( P Y * , P 1 ) = D ˜ Δ 0 ( P Y * , P 0 ) , for some PMF P Y * . This resembles the best achievable exponent in the Bayesian probability of error for the non-adversarial case, which is attained when D ( P Y * P 0 ) = D ( P Y * P 1 ) for some P Y * (see [19] (Theorem 11.9.1)). In that case, from the expression of the divergence function, such P Y * is found in a closed form and the resulting exponent is equivalent to the Chernoff information (see Section 11.9 [19]).
From Theorems 8 and 9, it follows that there is no positive λ , respectively a, for which the asymptotic exponent of the equilibrium payoff is strictly positive, if there exists a PMF P Y such that the following conditions are both satisfied:
D ˜ Δ 0 ( P Y , P 0 ) = 0 D ˜ Δ 1 ( P Y , P 1 ) = 0 .
In this case, then, P 0 and P 1 are indistinguishable under both the Neyman–Pearson and the Bayesian settings. We observe that the condition D ˜ Δ ( P Y , P X ) = 0 is equivalent to the following:
min Q X Y : ( Q X Y ) X = P X ( Q X Y ) Y = P Y E X Y d ( X , Y ) Δ ,
where the expectation E X Y is with respect to Q X Y . For ease of notation, in Equation (31), we used ( Q X Y ) Y = P Y (respectively ( Q X Y ) X = P X ) as short for x Q X Y ( x , y ) = P Y ( y ) , y A (respectively y Q X Y ( x , y ) = P X ( x ) , x A ), where Q X Y is a joint PMF and P X and P Y are its marginal PMFs.
In computer vision applications, the left-hand side of Equation (31) is known as the Earth Mover Distance (EMD) between P X and P Y , which is denoted by E M D d ( P X , P Y ) (or, by symmetry, E M D d ( P Y , P X ) ) [32]. It is also known as the ρ -bar distortion measure [33].
A brief comment concerning the analogy between the minimization in Equation (31) and optimal transport theory is worth expounding. The minimization problem in Equation (31) is known in the Operations Research literature as Hitchcock Transportation Problem (TP) [34]. Referring to the original Monge formulation of this problem [35], P X and P Y can be interpreted as two different ways of piling up a certain amount of soil; then, P X Y ( x , y ) denotes the quantity of soil shipped from location (source) x in P X to location (sink) y in P Y and d ( x , y ) is the cost for shipping a unitary amount of soil from x to y. In transport theory terminology, P X Y is referred to as transportation map. According to this perspective, evaluating the EMD corresponds to finding the minimal transportation cost of moving a pile of soil into the other. Further insights on this parallel can be found in [31].
We summarize our findings in the following corollary, which characterizes the conditions for distinguishability under both the Neyman–Pearson and the Bayesian setting.
Corollary 1 (Corollary to Theorems 8 and 9).
Given a memoryless source P 0 and distortion levels Δ 0 and Δ 1 , the set of the PMFs that cannot be distinguished from P 0 in both the Neyman–Pearson and Bayesian settings is given by
Γ = P : min P Y : E M D d ( P Y , P 0 ) Δ 0 E M D d ( P Y , P ) Δ 1 .
Set Γ is the indistinguishability region. By definition (see the beginning of this section), the PMFs inside Γ are those for which, as a consequence of the attack, the FP and FN probabilities cannot go to zero simultaneously with strictly positive exponents. Clearly, if Δ 0 = Δ 1 = 0 , that is, in the non-adversarial case, Γ = { P 0 } , as any two distinct sources are always distinguishable.
When d is a metric, for a given P Γ , the computation of the optimal P Y can be traced back to the computation of the EMD between P 0 and P, as stated by the following corollary, whose proof appears in Appendix C.2.
Corollary 2 (Corollary to Theorems 8 and 9).
When d is a metric, given the source P 0 and distortion levels Δ 0 and Δ 1 , for any fixed P, the minimum in Equation (32) is achieved when
P Y = α P 0 + ( 1 α ) P , α = 1 Δ 0 E M D ( P 0 , P ) .
Then, the set of PMFs that cannot be distinguished from P 0 in the Neyman–Pearson and Bayesian setting is given by
Γ = { P : E M D d ( P 0 , P ) Δ 0 + Δ 1 } .
According to Corollary 2, when d is a metric, the performance of the game depends only on the sum of distortions, Δ 0 + Δ 1 , and it is immaterial how this amount is distributed between the two hypotheses.
In the general case (d not a metric), the condition on the EMD stated in Equation (34) is sufficient in order for P 0 and P be indistinguishable, that is Γ { P : E M D d ( P 0 , P ) Δ 0 + Δ 1 } (see discussion in Appendix C.2, at the end of the proof of Corollary 2).
Furthermore, in the case of an L p p distortion function ( p 1 ), i.e., d ( x , y ) = i = 1 n | x i y i | p , we have the following corollary.
Corollary 3 (Corollary to Theorems 8 and 9).
When d is the L p p distortion function, for some p 1 , the set Γ can be bounded as follows
Γ { P : E M D L p p ( P 0 , P ) ( Δ 0 1 / p + Δ 1 1 / p ) p } .
Corollary 3 can be proven by exploiting the Hölder inequality [36] (see Appendix C.3).

7. Concluding Remarks

We consider the problem of binary hypothesis testing when an attacker is active under both hypotheses, and then an attack is carried out aiming at both false negative and false positive errors. By modeling the defender–attacker interaction as a game, we defined and solved two different detection games: the Neyman–Pearson and the Bayesian game. This paper extends the analysis in [11], where the attacker is active under the alternative hypothesis only. Another aspect of greater generality is that here both players are allowed to use randomized strategies. By relying on the method of types, the main result of this paper is the existence of an attack strategy, which is both dominant and universal, that is, optimal regardless of the statistics of the sources. The optimal attack strategy is also independent of the underlying hypothesis, namely fully-universal, when the distortion introduced by the attacker in the two cases is the same. From the analysis of the asymptotic behavior of the equilibrium payoff, we are able to establish conditions under which the sources can be reliably distinguished in the fully active adversarial setup. The theory developed permits to assess the security of the detection in adversarial setting and give insights on how the detector should be designed in such a way to make the attack hard.
Among the possible directions for future work, we mention the extension to continuous alphabets, which calls for an extension of the method of types to this case, or to more realistic models of finite alphabet sources, still amenable to analysis, such as Markov sources. As mentioned in the Introduction, it would be also relevant to extend the analysis to higher order statistics. In fact, the techniques introduced in this paper are very general and lend themselves to extensions that allow richer sets of sufficient statistics (as currently pursued in an ongoing work of the first two coauthors). This generalization, however, comes at the price of an increased complexity of the analysis with regard to both the expressions of the equilibrium strategies and the performance achieved by the two parties at the equilibrium. More specifically, given the set of statistics used by the defender, it turns out that the optimal attacker also uses the very same statistics (plus the empirical correlation with the source sequence, due to the distortion constraint) and the optimal attack channel is given in terms of uniform distributions across conditional types that depend on these set of statistics. Therefore, in a sequence of games, where the defender exploits more and more statistics, so would the attacker, and it would be interesting to explore the limiting behaviour of this process.
We also mention the case of unknown sources, where the source statistics are estimated from training data, possibly corrupted by the attacker. In this scenario, the detection game has been studied for a partially active case, with both uncorrupted and corrupted training data [13,14]. The extension of such analyses to the fully active scenario represents a further interesting direction for future research.
Finally, we stress that, in this paper, as well as in virtually all prior works using game theory to model the interplay between the attacker and the defender, it is assumed that the attacker is always present. This is a pessimistic or worst-case assumption, leading to the adoption of an overly conservative defense strategy (when the attacker is not present, in fact, better performance could be in principle achieved by adopting a non-adversarial detection strategy). A more far-reaching extension would require that the ubiquitous presence of the attacker is reconsidered, for instance by resorting to a two-steps analysis, where the presence or absence of the attacker is established in the first step, or by defining and studying more complex game models, e.g., games with incomplete information [37].

Author Contributions

All the authors have contributed equally to this research work and the writing of this paper.

Funding

This research received no external funding.

Acknowledgments

We thank Alessandro Agnetis of the University of Siena, for the useful discussions on optimization concepts underlying the computation of the EMD.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Neyman–Pearson Detection Game

This appendix contains the proofs of the results in Section 4.

Appendix A.1. Proof of Lemma 1

Whenever existent, the dominant defense strategy can be obtained by solving:
min Φ S D P FN ( Φ , A 1 ) ,
for any attack channel A 1 . Below, we first show that P FN ( Φ * , A 1 ) · P FN ( Φ , A 1 ) for every Φ S D and for every A 1 , that is, Φ * is asymptotically dominant. Then, by proving that max A C Δ 0 P FP ( Φ * , A ) fulfills the FP constraint, we show that Φ * is also admissible. Therefore, we can conclude that Φ * ( · | y ) asymptotically solves Equation (A1). Exploiting the memorylessness of P 0 and the permutation invariance of Φ ( H 1 | y ) and d ( x , y ) , for every y A n we have,
e λ n max A x , y P 0 ( x ) A ( y | x ) Φ ( H 1 | y ) y x P 0 ( x ) A Δ 0 * ( y | x ) Φ ( H 1 | y ) = y x : d ( x , y ) n Δ 0 P 0 ( x ) · c n ( x ) | T ( y | x ) | Φ ( H 1 | y ) ( n + 1 ) | A | · ( | A | 1 ) y x : d ( x , y ) n Δ 0 · P 0 ( x ) | T ( y | x ) | Φ ( H 1 | y ) ( a ) ( n + 1 ) | A | · ( | A | 1 ) | T ( y ) | max x : d ( x , y ) n Δ 0 | T ( x | y ) | · P 0 ( x ) | T ( y | x ) | Φ ( H 1 | y ) = ( b ) ( n + 1 ) | A | · ( | A | 1 ) Φ ( H 1 | y ) max x : d ( x , y ) n Δ 0 P 0 ( x ) · | T ( x ) | Φ ( H 1 | y ) max x : d ( x , y ) n Δ 0 e n D ( P ^ x P 0 ) ( n + 1 ) | A | 2 · ( | A | 1 ) = Φ ( H 1 | y ) exp n min x : d ( x , y ) n Δ 0 D ( P ^ x P 0 ) ( n + 1 ) | A | 2 · ( | A | 1 ) ,
where ( a ) is due to the permutation invariance of the distortion function d and ( b ) is due to the identity | T ( x ) | · | T ( y | x ) | | T ( y ) | · | T ( x | y ) | | T ( x , y ) | .
It now follows that
Φ ( H 1 | y ) · exp n λ min x : d ( x , y ) n Δ 0 D ( P ^ x P 0 ) .
Since Φ ( H 1 | y ) is a probability,
Φ ( H 1 | y ) · min 1 , exp n λ min x : d ( x , y ) n Δ 0 D ( P ^ x P 0 ) = Φ * ( H 1 | y ) .
Consequently, Φ * ( H 0 | y ) · Φ ( H 0 | y ) for every y, and so, P FN ( Φ * , A 1 ) · P FN ( Φ , A 1 ) for every A 1 . For convenience, let us denote
k n ( y ) = λ min x : d ( x , y ) n Δ 0 D ( P ^ x P 0 ) ,
so that Φ * ( H 1 | y ) = min { 1 , e n · k n ( y ) } . We now show that Φ * ( H 1 | y ) satisfies the FP constraint, up to a polynomial term in n, i.e., it satisfies the constraint asymptotically.
max A C Δ 0 P FP ( Φ * , A ) ( n + 1 ) | A | · ( | A | 1 ) P FP ( Φ * , A * ) = ( n + 1 ) | A | · ( | A | 1 ) x , y P 0 ( x ) A Δ 0 * ( y | x ) Φ * ( H 1 | y ) = ( n + 1 ) | A | · ( | A | 1 ) ( x , y ) : d ( x , y ) n Δ 0 P 0 ( x ) · c n ( x ) | T ( y | x ) | · Φ * ( H 1 | y ) ( n + 1 ) | A | · ( | A | 1 ) ( x , y ) : d ( x , y ) n Δ 0 P 0 ( x ) | T ( y | x ) | · Φ * ( H 1 | y ) ( n + 1 ) 2 | A | · ( | A | 1 ) y max x : d ( x , y ) n Δ 0 | T ( x | y ) | · P 0 ( x ) | T ( y | x ) | Φ * ( H 1 | y ) = ( n + 1 ) 2 | A | · ( | A | 1 ) P ^ y : k n ( y ) 0 e n k n ( y ) max x : d ( x , y ) n Δ 0 | T ( x ) | · P 0 ( x ) + + P ^ y : k n ( y ) < 0 max x : d ( x , y ) n Δ 0 | T ( x ) | · P 0 ( x ) ( n + 1 ) 2 | A | · ( | A | 1 ) P ^ y : k n ( y ) 0 e n λ + + P ^ y : k n ( y ) < 0 exp n min x : d ( x , y ) n Δ 0 D ( P ^ x P 0 ) ( n + 1 ) | A | ( | A | 2 + | A | 1 ) e n λ .

Appendix A.2. Proof of Property 1

We next prove that for any two PMFs P Y 1 and P Y 2 and any λ ( 0 , 1 ) ,
D ˜ Δ ( λ P Y 1 + ( 1 λ ) P Y 2 , P ) λ D ˜ Δ ( P Y 1 , P ) + ( 1 λ ) D ˜ Δ ( P Y 2 , P ) .
Let us rewrite D ˜ Δ in Equation (21) by expressing the minimization in terms of the joint PMF P X Y :
D ˜ Δ ( P Y , P ) = min { Q X Y : E X Y d ( X , Y ) Δ , ( Q X Y ) Y = P Y } D ( Q X Y ) X P ,
where we used ( Q X Y ) Y = P Y as short for x Q X Y ( x , y ) = P Y ( y ) , y , and we made explicit the dependence of D ( P X P ) on Q X Y . Accordingly:
D ˜ Δ ( λ P Y 1 + ( 1 λ ) P Y 2 , P ) = min { Q X Y : E X Y d ( X , Y ) Δ , ( Q X Y ) Y = λ P Y 1 + ( 1 λ ) P Y 2 } D ( Q X Y ) X P .
We find it convenient to rewrite the right-hand side of Equation (A8) by minimizing over pairs of PMFs ( Q X Y , Q X Y ) and considering the convex combination of these PMFs with weights λ and ( 1 λ ) , in place of Q X Y ; hence,
D ˜ Δ ( λ P Y 1 + ( 1 λ ) P Y 2 , P ) = min ( Q X Y , Q X Y ) H D λ ( Q X Y ) X + ( 1 λ ) ( Q X Y ) X P ,
where
H = ( Q X Y , Q X Y ) : λ ( Q X Y ) Y + ( 1 λ ) ( Q X Y ) Y = λ P Y 1 + ( 1 λ ) P Y 2 , λ E X Y d ( X , Y ) + ( 1 λ ) E X Y d ( X , Y ) Δ .
Let
H = Q X Y : E X Y d ( X , Y ) Δ , ( Q X Y ) Y = P Y 1 × Q X Y : E X Y d ( X , Y ) Δ , ( Q X Y ) Y = P Y 2 ;
then, H H , where the set H is separable in Q X Y and Q X Y . Accordingly, Equations (A9) and (A10) can be upper bounded by
min Q X Y : E X Y d ( X , Y ) Δ , ( Q X Y ) Y = P Y 1 min Q X Y : E X Y d ( X , Y ) Δ , ( Q X Y ) Y = P Y 2 D λ ( Q X Y ) X + ( 1 λ ) ( Q X Y ) X P .
By the convexity of D ( Q X Y ) X P with respect to Q X Y (this is a consequence of the fact that the divergence function is convex in its arguments and the operation ( · ) X is linear (see Theorem 2.7.2, [19])), it follows that
D λ ( Q X Y ) X + ( 1 λ ) ( Q X Y ) X P λ D ( Q X Y ) X P + ( 1 λ ) D ( Q X Y ) X P .
Note that the above relation is not strict since it might be that ( Q X Y ) X = ( Q X Y ) X = P . Then, an upper bound for D ˜ Δ ( λ P Y 1 + ( 1 λ ) P Y 2 , P ) is given by
min Q X Y : x Q X Y = P Y 1 , E X Y d ( X , Y ) Δ λ D ( Q X Y ) X P + min Q X Y : ( Q X Y ) Y = P Y 2 , E X Y d ( X , Y ) Δ ( 1 λ ) D ( Q X Y ) X P ,
thus proving Equation (A6).

Appendix A.3. Proof of Theorem 3

We start by proving the upper bound for the FN probability:
P FN ( Φ * , A Δ 1 * ) = x , y P 1 ( x ) A Δ 1 * ( y | x ) Φ * ( H 0 | y ) = y x : d ( x , y ) n Δ 1 P 1 ( x ) c n ( x ) | T ( y | x ) | 1 e n [ λ D ˜ Δ 0 n ( P ^ y , P 0 ) ] + y x : d ( x , y ) n Δ 1 P 1 ( x ) | T ( y | x ) | 1 e n [ λ D ˜ Δ 0 n ( P ^ y , P 0 ) ] + = y P ^ x | y : E x y d ( X , Y ) Δ 1 | T ( P ^ x | y ) | e n [ H ^ x ( X ) + D ( P ^ x P 1 ) ] | T ( P ^ y | x ) | 1 e n [ λ D ˜ Δ 0 n ( P ^ y , P 0 ) ] + = P ^ y P ^ x | y : E x y d ( X , Y ) Δ 1 | T ( P ^ x ) | e n [ H ^ x ( X ) + D ( P ^ x P 1 ) ] 1 e n [ λ D ˜ Δ 0 n ( P ^ y , P 0 ) ] + = P ^ y P ^ x | y : E x y d ( X , Y ) Δ 1 e n D ( P ^ x P 1 ) 1 e n [ λ D ˜ Δ 0 n ( P ^ y , P 0 ) ] + = P ^ y : D ˜ Δ 0 n ( P ^ y , P 0 ) < λ P ^ x | y : E x y d ( X , Y ) Δ 1 e n D ( P ^ x P 1 ) 1 e n ( λ D ˜ Δ 0 n ( P ^ y , P 0 ) ) P ^ y : D ˜ Δ 0 n ( P ^ y , P 0 ) < λ P ^ x | y : E x y d ( X , Y ) Δ 1 e n D ( P ^ x P 1 ) ( n + 1 ) 2 | A | · ( | A | 1 ) exp n min P ^ y : D ˜ Δ 0 n ( P ^ y , P 0 ) < λ min P ^ x | y : E x y d ( X , Y ) Δ 1 D ( P ^ x P 1 ) ( n + 1 ) 2 | A | · ( | A | 1 ) exp n inf P Y : D ˜ Δ 0 ( P Y , P 0 ) < λ min P X | Y : E X Y d ( X , Y ) Δ 1 D ( P X P 1 ) ( n + 1 ) 2 | A | · ( | A | 1 ) exp n min P Y : D ˜ Δ 0 ( P Y , P 0 ) λ min P X | Y : E X Y d ( X , Y ) Δ 1 D ( P X P 1 ) .
Then:
lim sup n 1 n ln P FN ( Φ * , A Δ 1 * ) min P Y : D ˜ Δ 0 ( P Y , P 0 ) λ min P X | Y : E X Y d ( X , Y ) Δ 1 D ( P X P 1 ) .
We now move on to the lower bound.
P FN ( Φ * , A Δ 1 * ) = x , y P 1 ( x ) A Δ 1 * ( y | x ) Φ * ( H 1 | y ) ( n + 1 ) | A | · ( | A | 1 ) y x : d ( x , y ) n Δ 1 P 1 ( x ) | T ( y | x ) | 1 e n [ λ D ˜ Δ 0 n ( P ^ y , P 0 ) ] + = ( n + 1 ) | A | · ( | A | 1 ) P ^ y : D ˜ Δ 0 n ( P ^ y , P 0 ) < λ P ^ x | y : E x y d ( X , Y ) Δ 1 e n D ( P ^ x P 1 ) 1 e n ( λ D ˜ Δ 0 n ( P ^ y , P 0 ) ) ( n + 1 ) | A | · ( | A | 1 ) e n D ( P ^ x P 1 ) ( 1 e n ( λ D ˜ Δ 0 n ( P ^ y , P 0 ) ) ) ,
where, for a fixed n, P ^ y is a PMF that satisfies D ˜ Δ 0 n ( P ^ y , P 0 ) λ ( ln n ) / n and P ^ x | y is such that the distortion constraint is satisfied. Since the set of rational PMFs is dense in the probability simplex, two such sequences can be chosen in such a way that ( P ^ y , P ^ x | y ) ( P Y * , P X | Y * ) (we are implicitly exploiting the fact that set { D ˜ Δ 0 n ( P ^ y , P 0 ) < λ } is dense in { D ˜ Δ 0 ( P Y , P 0 ) λ } , for every λ > 0 , which holds true from Property 1), where
( P Y * , P X | Y * ) = arg min ( P Y , P X | Y ) min P Y : D ˜ Δ 0 ( P Y , P 0 ) λ min P X | Y : E X Y d ( X , Y ) Δ 1 D ( P X P 1 ) .
Therefore, we can assert that:
lim inf n 1 n ln P FN ( Φ * , A Δ 1 * ) lim n 1 n ln e n D ( P ^ x P 1 ) 1 e n ( λ D ˜ Δ 0 n ( P ^ y , P 0 ) ) = lim n D ( P ^ x P 1 ) = D ( P X * P 1 ) = min P Y : D ˜ Δ 0 ( P Y P 0 ) λ min P X | Y : E X Y d ( X , Y ) Δ 1 D ( P X P 1 ) .
By combining the upper and lower bounds, we conclude that lim sup and lim inf coincide. Therefore, the limit of the sequence 1 / n ln P FN exists and the theorem is proven.

Appendix A.4. Proof of Theorem 4

First, observe that, by exploiting the definition of D ˜ Δ , Equation (22) can be rewritten as
ε FN = min P Y : D ˜ Δ 0 ( P Y , P 0 ) λ min P X | Y : E X Y d ( X , Y ) Δ 1 D ( P X P 1 ) = min P Z : D ( P Z P 0 ) λ min P Y | Z : E Y Z d ( Y , Z ) Δ 0 min P X | Y : E X Y d ( X , Y ) Δ 1 D ( P X P 1 ) .
To prove the theorem, we now show that Equation (A20) can be simplified as follows:
ε FN = min P Z : D ( P Z P 0 ) λ min P X | Z : E X Z d ( X , Z ) Δ 0 + Δ 1 D ( P X P 1 ) ,
which is equivalent to Equation (24) (see Section 3). The equivalence of the expressions in Equations (A20) and (A21) follows from the equivalence of the two feasible sets for the PMF P X . We first show that any feasible P X in Equation (A20) is also feasible in Equation (A21). Let P X be a feasible PMF in Equation (A20). By exploiting the properties of the triangular inequality property of the distance, we have that, regardless of the specific choice of the distributions P Y | Z and P X | Y in Equation (A20),
E X Z d ( X , Z ) E X Y Z [ d ( X , Y ) + d ( Y , Z ) ] = E X Y d ( X , Y ) + E Y Z d ( Y , Z ) Δ 0 + Δ 1 ,
and then P X is a feasible PMF in Equation (A21). To prove the opposite inclusion, we observe that, for any P Z and P X | Z such that D ( P Z P 0 ) λ and E X Z d ( X , Z ) Δ 0 + Δ 1 , it is possible to define a variable Y, and then two conditional PMFs P Y | Z and P X | Y , such that E X Y d ( X , Y ) Δ 1 and E Y Z d ( Y , Z ) Δ 0 . To do so, it is sufficient to let P Y be the convex combination of P X and P Z , that is P Y = α P X + ( 1 α ) P Z where α = Δ 0 / ( Δ 0 + Δ 1 ) . With this choice for the marginal, we can define P X | Y so that P X Y satisfies
P X Y ( i , j ) = ( 1 α ) P X Z ( i , j ) i , j i , P X Y ( i , i ) = ( 1 α ) P X Z ( i , i ) + α P X ( i ) i .
By adopting the transportation theory perspective introduced towards the end of Section 6, we can look at P X and P Y as two ways of piling up a certain amount of soil; then, P X Y can be interpreted as a map which moves P X into P Y ( P X Y ( i , j ) corresponds to the amount of soil moved from position i to j). The map in Equation (A23) is the one which leaves in place a percentage α of the mass and moves the remaining ( 1 α ) percentage to fill the pile ( 1 α ) P Z according to map ( 1 α ) P X Z .
Similarly, P Y | Z can be chosen such that P Y Z satisfies
P Y Z ( i , j ) = α P X Z ( i , j ) i , j i , P Y Z ( i , i ) = α P X Z ( i , i ) + ( 1 α ) P Z ( i ) i .
It is easy to see that, with the above choices, E X Y d ( X , Y ) = ( 1 α ) E X Z d ( X , Z ) and E Y Z d ( Y , Z ) = α E X Z d ( X , Z ) . Then, E X Y d ( X , Y ) ( 1 α ) ( Δ 0 + Δ 1 ) Δ 1 and E Y Z d ( Y , Z ) Δ 0 . Consequently, any P X belonging to the set in Equation (A21) also belongs to the one in Equation (A20).

Appendix B. Bayesian Detection Game

This appendix contains the proofs for Section 5.

Appendix B.1. Proof of Theorem 5

Given the probability distributions Q 0 ( y ) and Q 1 ( y ) induced by A Δ 0 * and A Δ 1 * , respectively, the optimal decision rule is deterministic and is given by the likelihood ratio test (LRT) [38]:
1 n ln Q 1 ( y ) Q 0 ( y ) H 1 H 0 a ,
which proves the optimality of the decision rule in Equation (25).
To prove the asymptotic optimality of the decision rule in Equation (26), let us approximate Q 0 ( y ) and Q 1 ( y ) using the method of types as follows:
Q 0 ( y ) = x P 0 ( x ) A Δ 0 * ( y | x ) = · x : d ( x , y ) n Δ 0 e n [ H ^ x ( X ) + D ( P ^ x P 0 ) ] · e n H ^ x y ( Y | X ) = · max x : d ( x , y ) n Δ 0 e n H ^ x y ( X | Y ) · e n [ H ^ x ( X ) + D ( P ^ x P 0 ) ] · e n H ^ x y ( Y | X ) = max x : d ( x , y ) n Δ 0 e n [ H ^ y ( Y ) + D ( P ^ x P 0 ) ] = ( a ) exp n H ^ y ( Y ) + + min { P ^ x | y : E x y d ( X , Y ) Δ 0 } D ( P ^ x P 0 ) = exp n [ H ^ y ( Y ) + D ˜ Δ 0 n ( P ^ y , P 0 ) ] ,
where in ( a ) we exploited the additivity of the distortion function d. Similarly,
Q 1 ( y ) = · exp n [ H ^ y ( Y ) + D ˜ Δ 1 n ( P ^ y , P 1 ) ] .
Thus, we have the following asymptotic approximation to the LRT:
D ˜ Δ 0 n ( P ^ y , P 0 ) D ˜ Δ 1 n ( P ^ y , P 1 ) H 1 H 0 a ,
which proves the second part of the theorem.

Appendix B.2. Proof of Theorem 7

To make the expression of u ( P FN ( Φ , ( A Δ 0 * , A Δ 1 * ) ) ) explicit, let us first evaluate the two error probabilities at equilibrium. Below, we derive the lower and upper bound on the probability of y under H 1 , when the attack channel is A Δ 1 * :
( n + 1 ) | A A 1 | e n [ H ^ y ( Y ) + D ˜ Δ 1 n ( P ^ y , P 1 ) ] Q 1 * ( y ) < ( n + 1 ) | A | 2 e n [ H ^ y ( Y ) + D ˜ Δ 1 n ( P ^ y , P 1 ) ] .
The same bounds hold for Q 0 * ( y ) , with D ˜ Δ 0 replacing D ˜ Δ 0 . For the FN probability, the upper bound is
P FN ( Φ , A Δ 1 * ) = y Q 1 * ( y ) · Φ ( H 0 | y ) = y : D ˜ Δ 0 n ( P ^ y , P 0 ) D ˜ Δ 1 n ( P ^ y , P 1 ) < a Q 1 * ( y ) ( n + 1 ) | A | 2 y : D ˜ Δ 0 n ( P ^ y , P 0 ) D ˜ Δ 1 n ( P ^ y , P 1 ) < a e n [ H ^ y + D ˜ Δ 1 n ( P ^ y , P 1 ) ] ( n + 1 ) | A | 2 + | A | max P ^ y : D ˜ Δ 0 n ( P ^ y , P 0 ) D ˜ Δ 1 n ( P ^ y , P 1 ) < a e n D ˜ Δ 1 n ( P ^ y , P 1 ) = ( n + 1 ) | A | 2 + | A | exp n min P ^ y : D ˜ Δ 0 n ( P ^ y , P 0 ) D ˜ Δ 1 n ( P ^ y , P 1 ) < a D ˜ Δ 1 n ( P ^ y , P 1 ) .
Then,
lim sup n 1 n ln ( P FN ( Φ , A Δ 1 * ) ) min P Y : D ˜ Δ 0 ( P Y , P 0 ) D ˜ Δ 1 ( P Y , P 1 ) a D ˜ Δ 1 ( P Y , P 1 ) .
For the lower bound,
P FN ( Φ , A Δ 1 * ) ( n + 1 ) | A A 1 | y : D ˜ Δ 0 n ( P ^ y , P 0 ) D ˜ Δ 1 n ( P ^ y , P 1 ) < a e n [ H ^ y + D ˜ Δ 1 n ( P ^ y , P 1 ) ] ( n + 1 ) | A A 1 | max P ^ y : D ˜ Δ 0 n ( P ^ y , P 0 ) D ˜ Δ 1 n ( P ^ y , P 1 ) < a e n D ˜ Δ 1 n ( P ^ y , P 1 ) = ( n + 1 ) | A A 1 | exp n min P ^ y : D ˜ Δ 0 n ( P ^ y , P 0 ) D ˜ Δ 1 n ( P ^ y , P 1 ) < a D ˜ Δ 1 n ( P ^ y , P 1 ) .
Then,
lim inf n 1 n ln ( P FN ( Φ , A Δ 1 * ) ) lim n D ˜ Δ 1 n ( P ^ y , P 1 ) = min P Y : D ˜ Δ 0 ( P Y , P 0 ) D ˜ Δ 1 ( P Y , P 1 ) a D ˜ Δ 1 ( P Y , P 1 ) ,
where P ^ y is a properly chosen PMF, belonging to the set { D ˜ Δ 0 n ( P ^ y , P 0 ) D ˜ Δ 1 n ( P ^ y , P 1 ) < a } for every n, and such that P ^ y P Y * where
P Y * = arg min P Y : D ˜ Δ 0 ( P Y , P 0 ) D ˜ Δ 1 ( P Y , P 1 ) a D ˜ Δ 1 ( P Y , P 1 ) .
By Property 1, set { D ˜ Δ 0 n ( P ^ y , P 0 ) D ˜ Δ 1 n ( P ^ y , P 1 ) < a } is dense in { P Y : D ˜ Δ 0 ( P Y , P 0 ) D ˜ Δ 1 ( P Y , P 1 ) a } and then such a sequence of PMFs can always be found.
By combining Equations (A31) and (A33), we get
ε FN = lim n 1 n ln ( P FN ( Φ , A Δ 1 * ) ) = min P Y : D ˜ Δ 0 ( P Y , P 0 ) D ˜ Δ 1 ( P Y , P 1 ) a D ˜ Δ 1 ( P Y , P 1 ) .
Therefore, from Equations (A30) and (A32) we have
P FN ( Φ , A Δ 1 * ) exp n min P ^ y : D ˜ Δ 0 n ( P ^ y , P 0 ) D ˜ Δ 1 n ( P ^ y , P 1 ) < a D ˜ Δ 1 n ( P ^ y , P 1 ) ,
and the limit of 1 n ln P FN exists and is finite.
Similar bounds can be derived for the FP probability, resulting in
P FP ( Φ * , A Δ 0 * ) exp n min P ^ y : D ˜ Δ 0 n ( P ^ y , P 0 ) D ˜ Δ 1 n ( P ^ y , P 1 ) a D ˜ Δ 0 n ( P ^ y , P 0 ) ,
and in particular
ε FP = lim n 1 n ln ( P FP ( Φ * , A Δ 0 * ) ) = min P Y : D ˜ Δ 0 ( P Y , P 0 ) D ˜ Δ 1 ( P Y , P 1 ) a D ˜ Δ 0 ( P Y , P 0 ) .
From Equation (A38), we see that, as argued, the profile ( Φ , ( A Δ 0 * , A Δ 1 * ) ) leads to a FP exponent always at least as large as a.
We are now ready to evaluate the asymptotic behavior of the payoff of the Bayesian detection game:
u = P FN ( Φ , A Δ 1 * ) + e a n P FP ( Φ , A Δ 0 * ) max { P FN ( Φ , A Δ 1 * ) , e a n P FP ( Φ , A Δ 0 * ) } exp n min min P ^ y : D ˜ Δ 0 n ( P ^ y , P 0 ) D ˜ Δ 1 n ( P ^ y , P 1 ) < a D ˜ Δ 1 n ( P ^ y , P 1 ) , min P ^ y : D ˜ Δ 0 n ( P ^ y , P 0 ) D ˜ Δ 1 n ( P ^ y , P 1 ) a ( D ˜ Δ 0 n ( P ^ y , P 0 ) a ) = exp n min P y max D ˜ Δ 1 n ( P ^ y , P 1 ) , ( D ˜ Δ 0 n ( P ^ y , P 0 ) a ) exp n min P Y max D ˜ Δ 1 ( P Y , P 1 ) , ( D ˜ Δ 0 ( P Y , P 0 ) a ) ,
where the asymptotic equality in the last line follows from the density of the set of empirical probability distributions of n-length sequences in the probability simplex and from the continuity of the to-be-minimized expression in round brackets as a function of P Y .

Appendix C. Source Distinguishability

This appendix contains the proofs for Section 6.

Appendix C.1. Proof of Theorem 9

The theorem directly follows from Theorem 7. In fact, by letting
e a ( P Y ) = max D ˜ Δ 1 ( P Y , P 1 ) , D ˜ Δ 0 ( P Y , P 0 ) a ,
a 0 , the limit in Equation (29) can be derived as follows:
lim a 0 min P Y e a ( P Y ) = min P Y lim a 0 e a ( P Y ) = min P Y max D ˜ Δ 1 ( P Y , P 1 ) , D ˜ Δ 0 ( P Y , P 0 ) ,
where the order of limit and minimum can be exchanged because of the uniform convergence of e a ( P Y ) to e 0 ( P Y ) as a tends to 0.

Appendix C.2. Proof of Corollary 2

The corollary can be proven by exploiting the fact that, when d is a metric, the EMD is a metric and then E M D d ( P 0 , P ) satisfies the triangular inequality. In this case, it is easy to argue that the P Y achieving the minimum in Equation (32) is the one for which the triangular relation holds at the equality, which corresponds to the convex combination of P 0 and P (i.e., the PMF lying on the straight line between P 0 and P) with combination coefficient α such that E M D d ( P 0 , P Y ) (or equivalently, by symmetry, E M D d ( P Y , P 0 ) ) is exactly equal to Δ 0 .
Formally, let X P 0 and Z P . We want to find the PMF P Y which solves
min P Y : E M D d ( P Y , P 0 ) Δ 0 E M D d ( P Y , P ) .
For any Y P Y and any choice of P X Y and P Y Z (that is, P Y | X and P Z | Y ), by exploiting the triangular inequality property of the distance, we can write
E X Z d ( X , Z ) E X Y d ( X , Y ) + E Y Z d ( Y , Z ) ,
where P X Z can be any joint distribution with marginals P 0 and P. Then,
E M D ( P 0 , P ) E X Y d ( X , Y ) + E Y Z d ( Y , Z ) .
From the arbitrariness of the choice of P X Y and P Y Z , if we let P X Y * and P Y Z * be the joint distributions achieving the EMD between X and Y, and Y and Z, we get
E M D ( P 0 , P ) E M D ( P 0 , P Y ) + E M D ( P Y , P ) .
From the above relation, we can derive the following lower bound for the to-be-minimized quantity in Equation (A42):
(A46) E M D ( P Y , P ) E M D ( P 0 , P ) E M D ( P 0 , P Y ) (A47) E M D ( P 0 , P ) Δ 0 .
We now show that P Y defined as in Equation (33) achieves the above lower bound while satisfying the constraint E M D ( P 0 , P Y ) Δ 0 , and then get the minimum value in Equation (A42).
Let P X Z * be the joint distribution achieving the EMD between X and Z. Then, E X Z * d ( X , Z ) = E M D ( P 0 , P ) (where the star on the apex indicates that the expectation is taken under P X Z * ). Given the marginal P Y = α P 0 + ( 1 α ) P , we can define P X Y and P Y Z , starting from P X Z * , as in the proof of Theorem 4 (Equations (A23) and (A24)). With this choice, E X Y d ( X , Y ) = ( 1 α ) E M D ( P 0 , P ) and E Y Z d ( Y , Z ) = α E M D ( P 0 , P ) . Then, for the value of α in Equation (33) we have that E X Y d ( X , Y ) = Δ 0 and
E Y Z d ( Y , Z ) = E M D ( P 0 , P ) Δ 0 .
By combining Equations (A47) and (A48), we argue that E M D ( P Y , P ) = E M D ( P 0 , P ) Δ 0 (we also argue that the choice made for P Y Z minimizes the expected distortion between Y and Z, i.e., it yields E Y Z d ( Y , Z ) = E M D ( P Y , P ) . Furthermore, being E X Y d ( X , Y ) = Δ 0 , it holds E M D ( P Y , P ) = E M D ( P 0 , P ) E X Y d ( X , Y ) and then, from the triangular inequality in Equation (A45), it follows that E M D ( P 0 , P Y ) = E X Y d ( X , Y ) = Δ 0 ) . Therefore, P Y in Equation (33) solves Equation (A42).
To prove the second part of the corollary, we just need to observe that a PMF P belongs to the indistinguishability set in Equation (32) if and only if
E M D ( P Y , P ) = E M D ( P 0 , P ) Δ 0 Δ 1 ,
that is E M D ( P 0 , P ) Δ 0 + Δ 1 .
From the above proof, we notice that, for any P in the set in Equation (34), i.e., such that E M D d ( P 0 , P ) Δ 0 + Δ 1 , the PMF P Y = α P 0 + ( 1 α ) P with α as in Equation (33) satisfies EMD ( P Y , P 0 ) = Δ 0 and EMD ( P Y , P 1 ) = Δ 1 for any choice of d. Then, when d is not a metric, the region in Equation (34) is contained in the indistinguishability region.

Appendix C.3. Proof of Corollary 3

By inspecting the minimization in Equation (32), we see that, for any source P that cannot be distinguished from P 0 , it is possible to find a source P Y such that E M D d ( P Y , P ) Δ 1 and E M D d ( P Y , P 0 ) Δ 0 . To prove the corollary, we need to show that such P lies inside the set defined in Equation (35).
We give the following definition. Given two random variables X and Y, the Hölder inequality applied to the expectation function [36] reads:
E X Y | X Y | E X [ | X | r ] 1 / r E Y [ | Y | q ] 1 / q ,
where r 1 and q = r / ( r 1 ) , namely, the Hölder conjugate of r.
We use the notation E X Y * for the expectation of the pair ( X , Y ) when the probability map is the one achieving the E M D ( P X , P Y ) , namely P X Z * . Then, we can write:
E M D L p p ( P 0 , P ) = E X Z * [ | | X Z | | p ] ( a ) E X Y Z * [ ( | | X Y | | + | | Y Z | | ) p ] ( b ) E X Y Z | | X Y | | p + | | Y Z | | p + p · | | X Y | | p 1 | | Y Z | | + + p ( p 1 ) / 2 · | | X Y | | p 2 | | Y Z | | 2 + + p · | | X Y | | | | Y Z | | p 1 = E X Y Z [ | | X Y | | p ] + E X Y Z [ | | Y Z | | p ] + p · E X Y Z [ | | X Y | | p 1 | | Y Z | | ] + + p ( p 1 ) / 2 · E X Y Z [ | | X Y | | p 2 | | Y Z | | 2 ] + . + p · E X Y Z [ | | X Y | | | | Y Z | | p 1 ] ( c ) E X Y Z [ | | X Y | | p ] + E X Y Z [ | | Y Z | | p ] + p · ( E X Y Z [ | | X Y | | p ] ) p 1 p ( E X Y Z [ | | Y Z | | p ] ) 1 p + p ( p 1 ) / 2 · ( E X Y Z [ | | X Y | | p ) p 2 p ( E X Y Z [ | | Y Z | | p ] ) 2 p + + p · ( E X Y Z [ | | X Y | | p ] ) 1 p ( E X Y Z [ | | Y Z | | p ] ) p 1 p = E X Y [ | | X Y | | p ] + E Y Z [ | | Y Z | | p ] + p · ( E X Y [ | | X Y | | p ] ) p 1 p ( E Y Z [ | | Y Z | | p ] ) 1 p + p ( p 1 ) / 2 · ( E X Y [ | | X Y | | p ] ) p 2 p ( E Y Z [ | | Y Z | | p ] ) 2 p + + p · ( E X Y [ | | X Y | | p ] ) 1 p ( E Y Z [ | | Y Z | | p ] ) p 1 p = ( E X Y Z [ | | X Y | | p ] ) 1 / p + ( E X Y Z [ | | Y Z | | p ] ) 1 / p p Δ 0 1 / p + Δ 1 1 / p p ,
where in ( a ) we considered the joint distribution P X Y Z such that Z P X Y Z = P X Y * , X P X Y Z = P Y Z * (and, consequently, Y P X Y Z = P X Z * ) and in ( b ) we developed the p-power of the binomial (binomial theorem). Finally, in ( c ) , we applied the Hölder’s inequality to the various terms of Newton’s binomial: specifically, for each term E X Y Z [ | | X Y | | p t | | Y Z | | t ] , with t = 1 , , p 1 , the Hölder inequality is applied with r = p / ( p t ) (and q = r / ( r 1 ) ).

References

  1. Barni, M.; Pérez-González, F. Coping with the enemy: Advances in adversary-aware signal processing. In Proceedings of the 2013 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Vancouver, BC, Canada, 26–31 May 2013; pp. 8682–8686. [Google Scholar]
  2. Ker, A.D. Batch steganography and the threshold game. In Security, Steganography, and Watermarking of Multimedia Contents; SPIE: Bellingham, WA, USA, 2007; p. 650504. [Google Scholar]
  3. Schöttle, P.; Böhme, R. Game Theory and Adaptive Steganography. IEEE Trans. Inf. Forensics Secur. 2016, 11, 760–773. [Google Scholar] [CrossRef]
  4. Cohen, A.; Lapidoth, A. The Gaussian watermarking game. IEEE Trans. Inf. Theory 2002, 48, 1639–1667. [Google Scholar] [CrossRef] [Green Version]
  5. Wu, Y.; Wang, B.; Liu, K.R.; Clancy, T.C. Anti-jamming games in multi-channel cognitive radio networks. IEEE J. Sel. Areas Commun. 2012, 30, 4–15. [Google Scholar] [CrossRef]
  6. Dalvi, N.; Domingos, P.; Mausam, P.; Sanghai, S.; Verma, D. Adversarial classification. In Proceedings of the 10th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Seattle, WA, USA, 22–25 August 2004; pp. 99–108. [Google Scholar]
  7. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. In Advances in Neural Information Processing Systems; MIT Press: Cambridge, UK, 2014; pp. 2672–2680. [Google Scholar]
  8. O’Sullivan, J.; Moulin, P.; Ettinger, J. Information theoretic analysis of steganography. In Proceedings of the IEEE International Symposium on Information Theory, Cambridge, MA, USA, 16–21 August 1998; p. 297. [Google Scholar] [CrossRef]
  9. Moulin, P.; O’Sullivan, J. Information-theoretic analysis of information hiding. IEEE Trans. Inf. Theory 2003, 49, 563–593. [Google Scholar] [CrossRef] [Green Version]
  10. Somekh-Baruch, A.; Merhav, N. On the capacity game of public watermarking systems. IEEE Trans. Inf. Theory 2004, 50, 511–524. [Google Scholar] [CrossRef]
  11. Barni, M.; Tondi, B. The source identification game: An information-theoretic perspective. IEEE Trans. Inf. Forensics Secur. 2013, 8, 450–463. [Google Scholar] [CrossRef]
  12. Merhav, N.; Sabbag, E. Optimal watermark embedding and detection strategies under limited detection resources. IEEE Trans. Inf. Theory 2008, 54, 255–274. [Google Scholar] [CrossRef]
  13. Barni, M.; Tondi, B. Binary Hypothesis Testing Game With Training Data. IEEE Trans. Inf. Theory 2014, 60, 4848–4866. [Google Scholar] [CrossRef] [Green Version]
  14. Barni, M.; Tondi, B. Adversarial Source Identification Game With Corrupted Training. IEEE Trans. Inf. Theory 2018, 64, 3894–3915. [Google Scholar] [CrossRef] [Green Version]
  15. Moulin, P.; Ivanovic, A. Game-theoretic analysis of watermark detection. In Proceedings of the International Conference on Image Processing, Thessaloniki, Greece, 7–10 October 2001; Volume 3, pp. 975–978. [Google Scholar]
  16. Vigna, G.; Robertson, W.; Balzarotti, D. Testing network-based intrusion detection signatures using mutant exploits. In Proceedings of the 11th ACM Conference on Computer and Communications Security, Washington, DC, USA, 25–29 October 2004; pp. 21–30. [Google Scholar]
  17. Patton, S.; Yurcik, W.; Doss, D. An Achilles’ heel in signature-based IDS: Squealing false positives in SNORT. In Proceedings of the RAID, Davis, CA, USA, 10–12 October 2001; Volume 2001. [Google Scholar]
  18. Mutz, D.; Kruegel, C.; Robertson, W.; Vigna, G.; Kemmerer, R.A. Reverse engineering of network signatures. In Proceedings of the Auscert Asia Pacific Information Technology Security Conference, Gold Coast, Australia, 22–26 May 2005. [Google Scholar]
  19. Cover, T.M.; Thomas, J.A. Elements of Information Theory, 2nd ed.; Wiley: Hoboken, NJ, USA, 2006. [Google Scholar]
  20. Cao, G.; Zhao, Y.; Ni, R.; Li, X. Contrast enhancement-based forensics in digital images. IEEE Trans. Inf. Forensics Secur. 2014, 9, 515–525. [Google Scholar] [CrossRef]
  21. Pan, X.; Zhang, X.; Lyu, S. Exposing image splicing with inconsistent local noise variances. In Proceedings of the 2012 IEEE International Conference on Computational Photography (ICCP), Seattle, WA, USA, 28–29 April 2012; pp. 1–10. [Google Scholar]
  22. Popescu, A.C.; Farid, H. Statistical Tools for Digital Forensics. In Proceedings of the 6th International Conference on Information Hiding, IH’04, Toronto, ON, Canada, 23–25 May 2004; Springer: Berlin/Heidelberg, Germany, 2004; pp. 128–147. [Google Scholar] [CrossRef]
  23. Tondi, B.; Barni, M.; Merhav, N. Detection games with a fully active attacker. In Proceedings of the 2015 IEEE International Workshop on Information Forensics and Security (WIFS), Hong Kong, China, 11–13 December 2015; pp. 1–6. [Google Scholar] [CrossRef]
  24. Osborne, M.J.; Rubinstein, A. A Course in Game Theory; MIT Press: Cambridge, UK, 1994. [Google Scholar]
  25. Nash, J. Equilibrium points in n-person games. Proc. Natl. Acad. Sci. USA 1950, 36, 48–49. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  26. Chen, Y.C.; Van Long, N.; Luo, X. Iterated strict dominance in general games. Games Econ. Behav. 2007, 61, 299–315. [Google Scholar] [CrossRef] [Green Version]
  27. Bernheim, D. Rationalizable Strategic Behavior. Econometrica 1984, 52, 1007–1028. [Google Scholar] [CrossRef]
  28. Hoeffding, W. Asymptotically optimal tests for multinomial distributions. Ann. Math. Stat. 1965, 36, 369–401. [Google Scholar] [CrossRef]
  29. Sanov, I.N. On the probability of large deviations of random variables. Math. Sb. 1957, 42, 11–44. [Google Scholar]
  30. Csiszár, I.; Shields, P.C. Information theory and statistics: A tutorial. Found. Trends Commun. Inf. Theory 2004, 1, 417–528. [Google Scholar] [CrossRef]
  31. Barni, M.; Tondi, B. Source Distinguishability Under Distortion-Limited Attack: An Optimal Transport Perspective. IEEE Trans. Inf. Forensics Secur. 2016, 11, 2145–2159. [Google Scholar] [CrossRef]
  32. Rubner, Y.; Tomasi, C.; Guibas, L.J. The Earth Mover’s Distance as a metric for image retrieval. Int. J. Comput. Vis. 2000, 40, 99–121. [Google Scholar] [CrossRef]
  33. Gray, R.M. Entropy and Information Theory; Springer Science & Business Media: Berlin, Germany, 2011. [Google Scholar]
  34. Hitchcock, F.L. The distribution of a product from several sources to numerous localities. J. Math. Phys. 1941, 20, 224–230. [Google Scholar] [CrossRef]
  35. Monge, G. Mémoire sur la Théorie des Déblais et des Remblais; De l’Imprimerie Royale: Paris, France, 1781. [Google Scholar]
  36. Karr, A.F. Probability; Springer: New York, NY, USA, 1993. [Google Scholar]
  37. Osborne, M.J. An Introduction to Game Theory; Oxford University Press: New York, NY, USA, 2004; Volume 3. [Google Scholar]
  38. Van Trees, H.L. Detection, Estimation and Modulation Theory. vol. 2., Nonlinear Modulation Theory; John Wiley and Sons: New York, NY, USA, 1971. [Google Scholar]
Figure 1. Schematic representation of the adversarial setup considered in this paper. In the case of partially active attacker, channel A 0 corresponds to the identity channel.
Figure 1. Schematic representation of the adversarial setup considered in this paper. In the case of partially active attacker, channel A 0 corresponds to the identity channel.
Entropy 21 00023 g001
Figure 2. Graphical interpretation of the behavior of the attack channel A Δ * . The number of admissible conditional type classes is polynomial in n, that is { y : d ( x , y ) n Δ } = i p ( n ) T ( y i | x ) where p ( n ) is a polynomial of n.
Figure 2. Graphical interpretation of the behavior of the attack channel A Δ * . The number of admissible conditional type classes is polynomial in n, that is { y : d ( x , y ) n Δ } = i p ( n ) T ( y i | x ) where p ( n ) is a polynomial of n.
Entropy 21 00023 g002

Share and Cite

MDPI and ACS Style

Tondi, B.; Merhav, N.; Barni, M. Detection Games under Fully Active Adversaries. Entropy 2019, 21, 23. https://doi.org/10.3390/e21010023

AMA Style

Tondi B, Merhav N, Barni M. Detection Games under Fully Active Adversaries. Entropy. 2019; 21(1):23. https://doi.org/10.3390/e21010023

Chicago/Turabian Style

Tondi, Benedetta, Neri Merhav, and Mauro Barni. 2019. "Detection Games under Fully Active Adversaries" Entropy 21, no. 1: 23. https://doi.org/10.3390/e21010023

APA Style

Tondi, B., Merhav, N., & Barni, M. (2019). Detection Games under Fully Active Adversaries. Entropy, 21(1), 23. https://doi.org/10.3390/e21010023

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop