Next Article in Journal
Probabilistic Resumable Quantum Teleportation of a Two-Qubit Entangled State
Previous Article in Journal
Making the Environment an Informative Place: A Conceptual Analysis of Epistemic Policies and Sensorimotor Coordination
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Bayesian Input Design for Linear Dynamical Model Discrimination

Department of Automatic Control and Robotics, AGH University of Science and Technology, Al. A. Mickiewicza 30, 30-059 Krakow, Poland
Entropy 2019, 21(4), 351; https://doi.org/10.3390/e21040351
Submission received: 16 January 2019 / Revised: 12 March 2019 / Accepted: 27 March 2019 / Published: 30 March 2019
(This article belongs to the Section Signal and Data Analysis)

Abstract

:
A Bayesian design of the input signal for linear dynamical model discrimination has been proposed. The discrimination task is formulated as an estimation problem, where the estimated parameter indexes particular models. As the mutual information between the parameter and model output is difficult to calculate, its lower bound has been used as a utility function. The lower bound is then maximized under the signal energy constraint. Selection between two models and the small energy limit are analyzed first. The solution of these tasks is given by the eigenvector of a certain Hermitian matrix. Next, the large energy limit is discussed. It is proved that almost all (in the sense of the Lebesgue measure) high energy signals generate the maximum available information, provided that the impulse responses of the models are different. The first illustrative example shows that the optimal signal can significantly reduce error probability, compared to the commonly-used step or square signals. In the second example, Bayesian design is compared with classical average D-optimal design. It is shown that the Bayesian design is superior to D-optimal design, at least in this example. Some extensions of the method beyond linear and Gaussian models are briefly discussed.

1. Introduction

Discrimination of various dynamical models of the same process has a wide area of applications, especially in multiple-model fault detection and isolation [1,2,3,4] and, in many other estimation and control problems [5,6,7], it is necessary to choose the most likely dynamical model from a finite set. The discrimination task can be formulated as an estimation problem, where the estimated parameter θ indexes particular models. The problem can also be considered as a finite-dimensional approximation of more general identification tasks [8]. As the error probability or variance of the estimator of θ usually depends on the input signal, it is important to select a signal that minimizes error probability or maximizes a utility function that encodes the purpose of the experiment. Selection of an input signal that maximizes a utility function is strongly related to optimal experimental design [9].
Experimental design methods can be divided into classical and Bayesian. The classical methods, also called optimal experimental design, typically use various functionals of the Fisher information matrix as a utility function. These methods are widely described in the literature and work well if the model is linear in its parameters (see [8,10,11,12] and the review article [9]). Unfortunately, in typical identification tasks, the solution of model equation and the covariance depends non-linearly on θ , even if the model equation is linear. This implies that the information matrix and the utility function depend on the parameter θ to be estimated. Therefore, only locally-optimal design can be obtained [13]. To obtain more robust methods, an averaging over the prior parameter distribution or minimax design [8,14] (Section 6.1) are commonly used, but these methods are not fully Bayesian.
Bayesian optimal design uses the utility function, a functional of the posterior distribution (see [15,16] and the review articles [13,17,18]). The most commonly used utility functions are mutual information between parameters and model output, Kullback-Leibler divergence between the prior and posterior distributions, and the determinant of the posterior covariance matrix [13,16]. In contrast to classical methods, in Bayesian design the utility function does not depend on the parameters to be estimated. Hence, the method can cope with non-linear problems. The utility function, which is suitable for model discrimination, is the error probability of the MAP estimator of θ [19]. Such a utility function is generally difficult to calculate (see [19]), but the result of Feder & Merhav [20] implies that the error probability of the MAP estimator is upper-bounded by some decreasing function of mutual information between θ and the output of the system. Hence, the maximization of mutual information creates the possibility of reducing the error probability, provided that appropriate estimator is used. However, the most serious problem that inhibits the development of this idea is great computational complexity in calculating the mutual information.
The main contribution of this article is a fully-Bayesian (in the terminology of [13]) method for finding an input signal that maximizes the mutual information between θ and the system output. Maximization of information or, equivalently, maximization of the output entropy has been proposed by many authors (see, e.g., [13,15,17,18,21,22,23]), but the mutual information is very hard to compute and the problem is often intractable. To overcome this serious difficulty, instead of mutual information the lower bound, given by Kolchinsky & Tracey [24], has been used. This is a pairwise-distance based entropy estimator and it it useful here, since it is differentiable, tight, and asymptotically reaches the maximum possible information (see [24] (Sections 3.2, 4, 6)). Maximization of such a lower bound, under the signal energy (i.e., the square of the signal norm) constraints, is much simpler, gives satisfactory solutions, and allows for practical implementation of the idea of maximizing information. This is illustrated with examples. Moreover, it is shown that, for certain cases, this problem reduces to a solution of a certain eigenproblem.
The article is organized as follows. In Section 2, the estimation task is formulated and the upper bound of the error probability and the lower bound of the mutual information are given. In Section 2.1, a selection between two models is discussed and an exact solution is given. Design of input signals with small energy, which is required in some applications, is described in Section 2.2. In Section 2.3, the large energy limit is discussed. An application to linear dynamical systems with unknown parameters is given in Section 3. An example of finding the most likely model among three stochastic models with different structures is given in Section 4. Comparison with classical D-optimal design is performed in Section 5. The article ends with conclusions and references.

2. Maximization of Mutual Information between the System Output and Parameter

Let us consider a family of linear models
Y = F θ U + Z ,
where θ 1 , 2 , . . . , r , Y , Z R n Y , and U R n U . The matrices F θ are bounded. The parameter θ is unknown. The prior distribution of θ is given by
P ( θ = i ) = p 0 , i , i = 1 , . . . , r .
The random variable Z is conditionally normal (i.e., p ( Z | θ ) = N ( Z , 0 , S θ ) ), where the covariance matrices S θ are given a priori and S θ > 0 , for all θ . The variable U is called the input signal. In all formulas below, the input signal U is a deterministic variable. The set of admissible signals is given by
S ϱ = { U R n U ; U T U ϱ } .
Under these assumptions, and after applying Bayes rule:
p ( Y | U ) = θ = 1 r p 0 , θ N ( Y , F θ U , S θ ) ,
p ( Y | θ , U ) = N ( Y , F θ U , S θ ) ,
p ( θ | Y , U ) = p 0 , θ N ( Y , F θ U , S θ ) j = 1 r p 0 , j N ( Y , F j U , S j ) .
The entropies of Y and θ and the conditional entropies are defined as
H ( θ ) = θ = 1 r p 0 , θ ln p 0 , θ ,
H ( θ | Y , U ) = p ( Y | U ) θ = 1 r p ( θ | Y , U ) ln p ( θ | Y , U ) d Y ,
H ( Y | U ) = p ( Y | U ) ln p ( Y | U ) d Y ,
H ( Y | θ ) = 1 2 θ = 1 r p 0 , θ ln ( 2 π e ) n y | S θ | .
The mutual information between θ and Y is defined as (see [25] (pp. 19, 250))
I ( Y ; θ | U ) = H ( θ | U ) H ( θ | Y , U ) = H ( Y | U ) H ( Y | θ , U ) .
As H ( θ | U ) = H ( θ ) and H ( Y | θ , U ) = H ( Y | θ ) , then I ( Y ; θ | U ) is given by
I ( Y ; θ | U ) = H ( θ ) H ( θ | Y , U ) = H ( Y | U ) H ( Y | θ ) .
The MAP estimator of θ is defined as
θ ^ ( Y , U ) = arg max θ { 1 , . . . , r } p ( θ | Y , U ) .
The error probability of θ ^ is given by (see [20])
P e ( U ) = 1 max θ { 1 , . . . , r } p ( θ | Y , U ) p ( Y | U ) d Y .
It follows from Fano’s inequality ([25] (p. 38)), that P e is lower bounded by an increasing function in H ( θ | Y , U ) . Feder & Merhav [20] proved that 2 P e ( U ) H ( θ | Y , U ) log 2 e . As H ( θ | Y , U ) = H ( θ ) I ( Y ; θ | U ) and H ( θ ) does not depend on U, then the maximization of I ( Y ; θ | U ) creates the possibility of reducing P e , and the optimal signal is given by
U * ( ϱ ) = arg max U S ϱ I ( Y ; θ | U ) .
To overcome the problems associated with the calculation of I ( Y ; θ | U ) , we will use its lower bound.
Lemma 1. 
(Information bounds). For all U R n U ,
I l ( U ) I ( Y ; θ | U ) H ( θ ) ,
where
I l ( U ) = i = 1 r p 0 , i ln j = 1 r p 0 , j e D i , j ( U ) ,
D i , j ( U ) = 1 4 U T Q i , j U + 1 2 ln | 1 2 ( S i + S j ) | 1 4 ln | S i | | S j | , and
Q i , j = ( F i F j ) T ( S i + S j ) 1 ( F i F j ) .
Proof. 
According to (4), p ( Y | U ) is finite Gaussian mixture. For such mixtures, the information bounds are known. A detailed proof, based on Chernoff α -divergence, is given in [24] (Section 4). □
Lemma 2.
Let θ ^ ( Y , U ) = arg max θ { 1 , . . . , r } p ( θ | Y , U ) be the MAP estimator of θ, and let P e ( U ) denote its error probability. There exists a continuous, increasing, and concave function f : [ 0 , H ( θ ) ] [ 0 , 1 r 1 ] , such that
P e ( U ) f ( H ( θ ) I l ( U ) ) 1 2 ( H ( θ ) I l ( U ) ) log 2 e .
Proof. 
Feder & Merhav [20] (see Theorem 1 and Equation (14)) proved that there exists an increasing, continuous, and convex function ϕ : [ 0 , 1 r 1 ] [ 0 , H ( θ ) log 2 e ] , such that
2 P e ( U ) ϕ ( P e ( U ) ) H ( θ | Y , U ) log 2 e .
As H ( θ | Y , U ) = H ( θ ) I ( Y ; θ | U ) and I l ( U ) I ( Y ; θ | U ) , then ϕ ( P e ( U ) ) ( H ( θ ) I l ( U ) ) log 2 e . The function g = ϕ 1 is increasing, continuous, concave, and it follows from (18) that 2 g ( η ) η . Hence, P e ( U ) ϕ 1 ( ( H ( θ ) I l ( U ) ) log 2 e ) = g ( ( H ( θ ) I l ( U ) ) log 2 e ) 1 2 ( H ( θ ) I l ( U ) ) log 2 e . Taking f ( η ) = g ( η log 2 e ) we obtain the result. □
Now, the approximate solution of (12) is given by
U * ( ϱ ) = arg max U S ϱ I l ( U ) .
As I l is smooth and S ϱ is compact, (19) is well-defined.

2.1. Selection between Two Models

Suppose that θ takes only two values, 1 and 2, with prior probabilities p 0 , 1 and p 0 , 2 = 1 p 0 , 1 , respectively. It’s easy to check, by direct calculation, that
e I l ( U ) = ( p 0 , 1 + p 0 , 2 e D 1 , 2 ( U ) ) p 0 , 1 ( p 0 , 1 e D 1 , 2 ( U ) + p 0 , 2 ) p 0 , 2 .
Equation (20) implies that the maximization of I l is equivalent to the maximization of D 1 , 2 . On the basis of (15), we have the following optimization: task
max U T U ϱ U T Q 1 , 2 U .
The solution of (21) is the eigenvector of Q 1 , 2 corresponding to its largest eigenvalue; that is,
Q 1 , 2 U * = λ m a x ( Q 1 , 2 ) U * , | | U * | | 2 = ϱ .

2.2. Small Energy Limit

In many practical applications, the energy of an excitation signal must be small. The second order Taylor expansion of (14) gives
I l ( U ) = 1 4 U T Q U i = 1 r p 0 , i ln j = 1 r α i , j + o ( | | U | | 2 ) ,
where
Q = i = 1 r p 0 , i j = 1 r α i , j Q i , j j = 1 r α i , j , and
α i , j = p 0 , j e D i , j ( 0 ) .
If the value of ϱ (see (3)) is small, then the last term in (23) can be omitted. As the second term does not depend on U, we have the following optimization task:
max U T U ϱ U T Q U .
The solution of (26) is the eigenvector of Q corresponding to its largest eigenvalue.

2.3. Large Energy Limit

We will investigate asymptotic behaviour of I ( Y ; θ | U ) when | | U | | . On the basis of Lemma 1, the condition
min i j U T Q i , j U > 0
guarantees that lim ϱ I ( Y ; θ | ϱ U ) = H ( θ ) . It is also possible that lim ϱ I ( Y ; θ | ϱ U ) < H ( θ ) , for some U. Such signals are weakly informative and they cannot generate the maximum information, even if their amplitude tends to infinity. Let S 1 denote the unit ball in R n U and let μ be the Lebesgue measure on S 1 . The set of weakly informative signals is defined as
Ω = { U S 1 : lim ϱ I ( Y ; θ | ϱ U ) < H ( θ ) } .
Theorem 1.
μ ( Ω ) = 0 if and only if F i F j , for all i j .
Proof. 
: On the basis of Lemma 1 and (28), Ω = i j Ω i , j , where Ω i , j = { ξ S 1 : ξ T Q i , j ξ = 0 } . Since S i + S j is positive-definite and F i F j then, on the basis of (16), the matrix Q i , j has at least one positive eigenvalue. Hence, μ ( Ω i , j ) = 0 and μ ( Ω ) = i j μ ( Ω i , j ) = 0 . : The condition μ ( Ω ) = 0 implies that μ ( Ω i , j ) = 0 . Hence, Q i , j has at least one positive eigenvalue, which is possible only if F i F j . □
As a conclusion, we have the following result.
Theorem 2.
Let θ ^ ( Y , U ) = arg max θ { 1 , . . . , r } p ( θ | Y , U ) , be the MAP estimator of θ and let P e ( U ) denote its error probability. If F i F j for all i j , then, for any ϵ > 0 , there exists a number ϱ > 0 and a signal U S ϱ , such that P e ( U ) < ϵ .
Proof. 
By the assumption, and from Theorem 1, the set S 1 Ω is non-empty. If U S 1 Ω , then min i j U T Q i , j U > 0 and, from Lemma 1, we get lim ϱ I l ( ϱ U ) = lim ϱ I ( Y ; θ | ϱ U ) = H ( θ ) . Now, Lemma 2 implies that 2 P e ( U ) ( H ( θ ) I l ( ϱ U ) ) log 2 e < ϵ , for sufficiently large ϱ . □

3. Application to Linear Dynamical Systems

Consider, now, the family of linear systems
x k + 1 = A θ x k + B θ u k + G θ w k , k = 0 , 1 , 2 , . . . , N 1 ,
y k = C θ x k + D θ v k , k = 1 , 2 , . . . , N ,
where the prior distribution of θ is given by (2) and x k R n , y k R m , w k R n w , v k R m , w k N ( 0 , I n w ) , and v k N ( 0 , I m ) . The variables w 0 , . . . , w N 1 , v 1 , . . . , v N are mutually independent. The initial condition is zero. The solution of (29) with initial condition x 0 = 0 has the form
x k = i = 0 k 1 A θ k i 1 B θ u i + i = 0 k 1 A θ k i 1 G θ w i .
If we denote X = col ( x 1 , . . . , x N ) , Y = col ( y 1 , . . . , y N ) , U = col ( u 0 , . . . , u N 1 ) , W = col ( w 0 , . . . , w N 1 ) , and V = col ( v 1 , . . . , v N ) , then (31) and (30) can be rewritten as
X = B θ U + G θ W , and
Y = C θ X + D θ V ,
where the matrices B θ , G θ , C θ , and D θ follow forms (30) and (31). The variables W and V are independent, where W N ( 0 , I N n w ) and V N ( 0 , I N m ) . Substituting (32) into (33) we get Equation (1), where F θ = C θ B θ , Z = C θ G θ W + D θ V . The conditional density of Z has the form p ( Z | θ ) = N ( Z , 0 , S θ ) , where the covariance matrix is given by
S θ = D θ D θ T + C θ G θ G θ T C θ T .
Hence, the results of Section 2 can be applied to the dynamical system (29) and (30).

4. Example

In some fault detection and isolation problems [3,4], there is a need to determine which of the known models of the process is the most adequate. It is, therefore, important to find a signal that emphasizes the differences between these various models. As an example of this type of problem, let us consider three stochastic continuous-time models:
d x = ( A θ x + B θ u ) d t + G θ d w ,
where θ { 1 , 2 , 3 } , x ( t ) R θ , x ( 0 ) = 0 , u ( t ) , w ( t ) R , w is a standard Wiener process, and
A 1 = 1 , B 1 = 1 , G 1 = 0 . 05 ,
A 2 = 0 1 3 2 . 5 , B 2 = 0 3 , G 2 = 0 0 . 05 ,
A 3 = 0 1 0 3 3 . 5 1 0 0 10 , B 3 = 0 0 30 , G 3 = 0 0 0 . 05 .
The step responses of these models are similar and they are difficult to experimentally distinguish from each other if the noise level is significant. The observation equation has the form
y k = x 1 ( t k ) + 0 . 05 v k , k = 1 , 2 , . . . , N ,
where v k N ( 0 , 1 ) , t k = k T 0 , T 0 = 0 . 1 , and x 1 is the first component of x ( t ) . If x k = x ( t k ) and u ( t ) = u k , t [ t k 1 , t k ) , then, after discretization, the state x k and the output y k are described by (29) and (30), with appropriate matrices A θ , B θ , C θ , G θ , and D θ . The matrices F θ and S θ are calculated by using (31)–(34). Let us observe that, although the orders of the systems are different, the size of both F θ and S θ is always N × N . We are interested in the maximization of I l ( U ) . The solutions of (19) and (26) with a uniform prior, ϱ = N , and N = 200 steps, are shown in the upper part of Figure 1. The step responses and the optimal responses are shown in the bottom part of Figure 1.
Let us observe that, in contrast to the step signal, the optimal signal clearly distinguishes the systems—although the energy of all input signals was the same and equal to N.
Let U s t , U s q S 1 denote the normalized step and square signal with period of three, respectively, and let U * ( ϱ ) denote the optimal signal. To check the validity of the results, the error probabilities P e ( ϱ U s t ) , P e ( ϱ U s q ) , and P e ( U * ( ϱ ) ) were estimated by Monte Carlo simulation with 10 6 trials and N = 50 steps. The results are shown in Figure 2. It was observed that the optimal signal gives an error probability several thousand times smaller than the step or square signal with the same energy. The second observation is that P e ( ϱ U s q ) initially increased with ϱ . To explain this, let us note that Inequality (17) does not guarantee that P e is decreasing function of ϱ . Hence, it is possible that P e increases in certain directions, although Theorem 2 guarantees that P e tends to zero, provided that signal norm tends to infinity.

5. Comparison with the Average D-Optimal Design

Classical methods of signal design for parameter identification use various functionals of the Fisher information matrix as a utility function. One of the most popular is D-optimal design, which consists of finding a signal that maximizes the determinant of the information matrix (see [10,11,12] and the review article [9]). These methods are well-suited to models that are linear in their parameters. Unfortunately, in typical identification and discrimination tasks, the output is a non-linear function of the parameters and the information matrix depends on unknown parameters to be identified. One of the possibilities for avoiding this problem is the averaging of the utility function over the prior parameter distribution. This method is called average D-optimal design (see [14,26] and [9] (Sections 5.3.5 and 6), for details). The Bayesian design, described in the previous sections, will be compared with the average D-optimal design. To that end, let us consider a finite family of linear models (see also [12] (pp. 91–93))
y k = b θ z 1 1 a θ z 1 u k + σ v v k ,
where θ { 1 , 2 , 3 , 4 } , a θ = 0.6 + 0.1 ( θ 1 ) , b θ = 1 a θ , σ v = 0.1 , and v k N ( 0 , 1 ) . The prior distribution of θ is uniform (i.e., p 0 , θ = 0.25 ). The state space representation of (40) has the form
x k + 1 = a θ x k + b θ u k ,
y k = x k + σ v v k ,
which is consistent with (29) and (30). The Fisher information matrix is given by
M F ( θ , U ) = 1 N σ v 2 k = 1 N d k d k T ,
where d k = ( ξ k , η k ) T and ξ k = y k a θ , η k = y k b θ , denote the sensitivity of the output y k to changes in parameters a and b, respectively. The derivatives ξ k and η k fulfil the sensitivity equations
ξ k = 2 a θ ξ k 1 a θ 2 ξ k 2 + b θ u k 2 ,
η k = a θ η k 1 + u k 1 , k = 1 , 2 , . . . , N ,
with zero initial conditions. The average D-optimal design consists in finding a signal U that maximizes the expectation of the determinant of the information matrix (see [9] (Sections 5.3.5 and 6), [11,14] (Chapter 6), and [12] for details). Hence, the utility function to be maximized has the form
J ( U ) = θ = 1 4 p 0 , θ | M F ( θ , U ) | ,
with the energy constraints given by (3). Maximization of the utility function (46) has been performed for various signal energies and the error probability of the MAP estimator was estimated by Monte Carlo with 10 5 trials. The same procedure was repeated using Bayesian design for (41) and (42). The results are shown in Figure 3. The error rate of Bayesian method is significantly smaller when compared to D-optimal design, at least in this example. In particular, the signal shown in the upper-right part of Figure 3 gives an error probability approximately three times smaller than that of D-optimal signal, although the energy of both signals was the same.

6. Possible Extensions of the Results

In this section, we will briefly discuss some possible extensions of the results to an infinite set of parameters and beyond linear and Gaussian models.

6.1. Non-Linear Models

Although the article refers to linear models, it is possible to extend the results to non-linear models of the form
Y = F θ ( U ) + Z ,
where the conditional density of variable Z is given by
p ( Z | θ , U ) = N ( Z , 0 , S θ ( U ) )
and S θ ( U ) > 0 , for all U U a d , θ { 1 , . . . , r } . Under these assumptions, the density of Y still remains a Gaussian mixture and the information lower bound takes the form
I l ( U ) = i = 1 r p 0 , i ln j = 1 r p 0 , j e D i , j ( U ) ,
where
D i , j ( U ) = 1 4 ( F i ( U ) F j ( U ) ) T ( S i ( U ) + S j ( U ) ) 1 ( F i ( U ) F j ( U ) ) + + 1 2 ln | 1 2 ( S i ( U ) + S j ( U ) ) | 1 4 ln | S i ( U ) | | S j ( U ) | .

6.2. Non-Gaussian Models

If p ( Z | θ , U ) is non-Gaussian distribution, then it is possible, on the basis of Equation (10) in [24], to construct an information lower bound of the form
I l ( U ) = i = 1 r p 0 , i ln j = 1 r p 0 , j e C α ( p i | | p j ) ,
where
C α ( p i | | p j ) = ln p ( Z | i , U ) α p ( Z | j , U ) 1 α d Z
is the Chernoff α -divergence and α [ 0 , 1 ] . Unfortunately, calculation of (52) is difficult if n Y is large.

6.3. Infinite Set of Parameters

Let us consider following model:
Y = F ( θ ) U + Z ,
where p ( Z | θ ) = N ( Z , 0 , S ( θ ) ) and θ R p . If we assume that prior density of θ is Gaussian; that is,
p 0 ( θ ) = N ( θ , m θ , S θ ) , S θ > 0 ,
then
p ( Y ) = p 0 ( θ ) N ( Y , F ( θ ) U , S ( θ ) ) ,
and p ( Y ) can be approximated by a finite Gaussian mixture
p ( Y | U ) = p 0 ( θ ) N ( Y , F ( θ ) U , S ( θ ) ) d θ j = 1 N a p 0 , j N ( Y , F ( θ j ) U , S ( θ j ) ) ,
where p 0 , j 0 and j = 1 N a p 0 , j = 1 . It’s possible to calculate weights and nodes in (56) by using multidimensional quadrature. If n θ is large, then an appropriate sparse grid should be used. To illustrate the method, we will show only a simple, second-order quadrature with 2 n θ points.
Lemma 3.
The approximate value of the integral J ( f ) = N ( θ , m θ , S θ ) f ( θ ) d θ is given by
J ( f ) 1 2 n θ j = 1 2 n θ f ( θ j ) ,
where
θ 2 i 1 = m θ S θ 0 . 5 e i , θ 2 i = m θ + S θ 0 . 5 e i , i = 1 , . . . , n θ
and e i is i t h basis vector. If f ( θ ) = 1 2 θ T A θ + b T θ + c , then equality holds in (57).
Proof. 
Direct calculation. □
Application of Lemma 3 to (56) gives p 0 , j = p 0 = ( 2 n θ ) 1 , N a = 2 n θ . Now, since (56) is a Gaussian mixture, the results of Section 2 can be utilized and the information lower bound takes the form
I l ( U ) = p 0 i = 1 r ln j = 1 r e D i , j ( U ) ln p 0 ,
where D i , j and θ j are given by (15), (16), and (58), respectively, and F j = F ( θ j ) , S j = S ( θ j ) . The approximate solution of (12) can be found by maximization of (59) with the constraints (3).

7. Discussion and Conclusions

An effective Bayesian design method for linear dynamical model discrimination has been developed. The discrimination task is formulated as an estimation problem, where the estimated parameter θ indexes particular models. To overcome the computational complexity, instead of I ( Y ; θ | U ) , its lower bound ( I l ( U ) ), proposed by Kolchinsky & Tracey [24], was used as a utility function. This bound is especially useful, as it is differentiable, tight, and reaches the maximum available information H ( θ ) . It has been proved, on the basis of the results of Feder & Merhav [20], that the error probability of the MAP estimator (see Lemma 2) is upper bounded by 1 2 ( H ( θ ) I l ( U ) ) log 2 e . The maximization of I l ( U ) has been considered under the signal energy constraint, but other kinds of constraints can also be easily implemented. It was shown that the maximization of I l ( U ) , in the case of two parameters, is equivalent to maximization of a quadratic form on the sphere (see also [3]). Next, the small energy limit was analyzed. It was proved that the solution is given by an eigenvector corresponding to maximal eigenvalue of some Hermitian matrix. This result can serve as starting point for numerical maximization of I l ( U ) . If the energy of the signal tends to infinity, then almost all (in the sense of the Lebesgue measure) signals generate maximum information, provided that the impulse responses of the models are pairwise different. Under these conditions, it was proved that P e of the MAP estimator tends to zero.
An example of discrimination of three stochastic models with different structures was given. It is easy to observe, from Figure 1, that, in contrast to the step signal, the optimal signal clearly distinguished the systems, although the energy of both signals was the same. The P e of the MAP estimator was calculated by Monte Carlo simulation. It was observed that the square signal gave an error probability several thousand times greater than the optimal signal with the same energy. Hence, we conclude that the error probability and the accuracy of MAP estimator depends very strongly on the excitation signal. Although Theorem 2 implies that lim ϱ P e ( ϱ U ) = 0 for almost all U, there exist signals such that P e ( ϱ U ) is locally increasing. This is the case of a high-frequency square signal, as illustrated in Figure 2.
It was shown, in Section 5 (see Figure 3), that P e of the MAP estimator corresponding to Bayesian design was a few times smaller than P e generated by D-optimal design, at least in the analyzed example. This result suggests that Bayesian design can be applied to non-linear problems and that it is superior to classical D-optimal design.
Some extensions of the results to the infinite set of parameters and beyond linear and Gaussian assumptions were briefly discussed in Section 6. Extension to non-linear models seems to be easy, but the non-Gaussian case is difficult and a deeper analysis is required. The case of an infinite set of parameters was discussed in Section 6.3. It was shown that the measurement density can be approximated by a finite Gaussian mixture, after which the results of Section 2 could be directly applied. The general conclusion of this analysis is that the information bounds can be easily constructed, as long as the measurement density is a mixture of Gaussian distributions.
An analytical gradient formula must be provided for the effective numerical maximization of I l ( U ) . The matrix inversions and the determinants in (15) and (16) should be calculated by SVD. To reduce the computational complexity and required memory resources, the symmetries appearing in (15) and (16), and the fact that D i , i = 0 , should be utilized. The determinants in (15) and the matrices F i , S i can be calculated off-line, but the matrices Q i , j may require too much memory if N and r are large. Therefore, Q i , j was calculated on-line.
Applications of the presented methods in dual control problems [5] are expected as part of our future work. An additional area in applications is the issue of automated testing of dynamical systems.

Funding

This research was financed from the statutory subsidy of the AGH University of Science and Technology, No. 16.16.120.773.

Acknowledgments

I would like to thank Jerzy Baranowski for discussions and comments.

Conflicts of Interest

The author declares no conflict of interest.

Abbreviations

ξ N ( m , S ) means that ξ has normal distribution with mean m and covariance S. The density of a normally-distributed variable is denoted by N ( x , m , S ) = 2 π n 2 | S | 1 2 exp ( 0.5 ( x m ) T S 1 ( x m ) ) . The symbol col ( a 1 , a 2 , . . . , a n ) denotes a column vector. The set of symmetric, positive-definite matrices of dimension n is denoted by S + ( n ) .

References

  1. Bania, P.; Baranowski, J. Bayesian estimator of a faulty state: Logarithmic odds approach. In Proceedings of the 22nd International Conference on Methods and Models in Automation and Robotics (MMAR), Miedzyzdroje, Poland, 28–31 August 2017; pp. 253–257. [Google Scholar]
  2. Baranowski, J.; Bania, P.; Prasad, I.; Cong, T. Bayesian fault detection and isolation using Field Kalman Filter. EURASIP J. Adv. Signal Process. 2017, 79. [Google Scholar] [CrossRef]
  3. Blackmore, L.; Williams, B. Finite Horizon Control Design for Optimal Model Discrimination. In Proceedings of the 44th IEEE Conference on Decision and Control, Seville, Spain, 12–15 December 2005. [Google Scholar] [CrossRef]
  4. Pouliezos, A.; Stavrakakis, G. Real Time Fault Monitoring of Industrial Processes; Kluwer Academic: Boston, MA, USA, 1994. [Google Scholar]
  5. Bania, P. Example for equivalence of dual and information based optimal control. Int. J. Control 2018. [Google Scholar] [CrossRef]
  6. Lorenz, S.; Diederichs, E.; Telgmann, R.; Schütte, C. Discrimination of dynamical system models for biological and chemical processes. J. Comput. Chem. 2007, 28, 1384–1399. [Google Scholar] [CrossRef] [PubMed]
  7. Ucinski, D.; Bogacka, B. T-optimum designs for discrimination between two multiresponse dynamic models. J. R. Stat. Soc. B 2005, 67, 3–18. [Google Scholar] [CrossRef]
  8. Walter, E.; Pronzato, L. Identification of Parametric Models from Experimental Data. In Series: Communications and Control Engineering; Springer: Berlin/Heidelberg, Germany, 1997. [Google Scholar]
  9. Pronzato, L. Optimal experimental design and some related control problems. Automatica 2008, 44, 303–325. [Google Scholar] [CrossRef]
  10. Atkinson, A.C.; Donev, A.N. Optimum Experimental Design; Oxford University Press: Oxford, UK, 1992. [Google Scholar]
  11. Goodwin, G.C.; Payne, R.L. Dynamic System Identification: Experiment Design and Data Analysis; Academic Press: New York, NY, USA, 1977. [Google Scholar]
  12. Payne, R.L. Optimal Experiment Design for Dynamic System Identification. Ph.D. Thesis, Department of Computing and Control, Imperial College of Science and Technology, University of London, London, UK, February 1974. [Google Scholar]
  13. Ryan, E.G.; Drovandi, C.C.; McGree, J.M.; Pettitt, A.N. A Review of Modern Computational Algorithms for Bayesian Optimal Design. Int. Stat. Rev. 2016, 84, 128–154. [Google Scholar] [CrossRef]
  14. Fedorov, V.V. Convex design theory. Math. Operationsforsch. Stat. Ser. Stat. 1980, 1, 403–413. [Google Scholar]
  15. Lindley, D.V. Bayesian Statistics—A Review; Society for Industrial and Applied Mathematics (SIAM): Philadelphia, PA, USA, 1972. [Google Scholar]
  16. Ryan, E.G.; Drovandi, C.C.; Pettitt, A.N. Fully Bayesian Experimental Design for Pharmacokinetic Studies. Entropy 2015, 17, 1063–1089. [Google Scholar] [CrossRef]
  17. Chaloner, K.; Verdinelli, I. Bayesian Experimental Design: A Review. Stat. Sci. 1995, 10, 273–304. [Google Scholar] [CrossRef]
  18. DasGupta, A. Review of Optimal Bayes Designs; Technical Report; Purdue University: West Lafayette, IN, USA, 1995. [Google Scholar]
  19. Routtenberg, T.; Tabrikian, J. A general class of lower bounds on the probability of error in multiple hypothesis testing. In Proceedings of the 25th IEEE Convention of Electrical and Electronics Engineers in Israel, Eilat, Israel, 3–5 December 2008; pp. 750–754. [Google Scholar]
  20. Feder, M.; Merhav, N. Relations between entropy and error probability. IEEE Trans. Inf. Theory 1994, 40, 259–266. [Google Scholar] [CrossRef]
  21. Arimoto, S.; Kimura, H. Optimum input test signals for system identification—An information-theoretical approach. Int. J. Syst. Sci. 1971, 1, 279–290. [Google Scholar] [CrossRef]
  22. Fujimoto, Y.; Sugie, T. Informative input design for Kernel-Based system identification. Procedings of the 2016 IEEE 55th Conference on Decision and Control (CDC), Las Vegas, NV, USA, 12–14 December 2016; pp. 4636–4639. [Google Scholar]
  23. Hatanaka, T.; Uosaki, K. Optimal Input Design for Discrimination of Linear Stochastic Models Based on Kullback-Leibler Discrimination Information Measure. In Proceedings of the 8th IFAC/IFOORS Symposium on Identification and System Parameter Estimation 1988, Beijing, China, 27–31 August 1988; pp. 571–575. [Google Scholar]
  24. Kolchinsky, A.; Tracey, B.D. Estimating Mixture Entropy with Pairwise Distances. Entropy 2017, 19, 361. [Google Scholar] [CrossRef]
  25. Cover, T.M.; Thomas, J.A. Elements of Information Theory, 2nd ed.; John Wiley & Sons, Inc.: Hoboken, NJ, USA, 2006. [Google Scholar]
  26. Atkinson, A.C.; Fedorov, V.V. Optimal design: Experiments for discriminating between several models. Biometrika 1975, 62, 289–303. [Google Scholar]
Figure 1. (Top) Numerical solution of (19) and the small energy approximation (26), for ϱ = N = 200 . (Bottom) Step responses and optimal responses of all systems.
Figure 1. (Top) Numerical solution of (19) and the small energy approximation (26), for ϱ = N = 200 . (Bottom) Step responses and optimal responses of all systems.
Entropy 21 00351 g001
Figure 2. Error probability of the MAP estimator for the optimal signal (.), step signal (+), and square (*) signal with period of three. The number of steps is N = 50 . The error probability has been estimated by a Monte Carlo method with 10 6 trials. Standard error bars were multiplied by factor of 10 for better visibility.
Figure 2. Error probability of the MAP estimator for the optimal signal (.), step signal (+), and square (*) signal with period of three. The number of steps is N = 50 . The error probability has been estimated by a Monte Carlo method with 10 6 trials. Standard error bars were multiplied by factor of 10 for better visibility.
Entropy 21 00351 g002
Figure 3. Error probability of theMAP estimator (see Section 2), as a function of signal norm and the exemplary input signals (top right) generated by D-optimal and Bayesian methods. The error probability was calculated by a Monte Carlo method with 10 5 trials. Standard error bars were multiplied by a factor of 10 for better visibility.
Figure 3. Error probability of theMAP estimator (see Section 2), as a function of signal norm and the exemplary input signals (top right) generated by D-optimal and Bayesian methods. The error probability was calculated by a Monte Carlo method with 10 5 trials. Standard error bars were multiplied by a factor of 10 for better visibility.
Entropy 21 00351 g003

Share and Cite

MDPI and ACS Style

Bania, P. Bayesian Input Design for Linear Dynamical Model Discrimination. Entropy 2019, 21, 351. https://doi.org/10.3390/e21040351

AMA Style

Bania P. Bayesian Input Design for Linear Dynamical Model Discrimination. Entropy. 2019; 21(4):351. https://doi.org/10.3390/e21040351

Chicago/Turabian Style

Bania, Piotr. 2019. "Bayesian Input Design for Linear Dynamical Model Discrimination" Entropy 21, no. 4: 351. https://doi.org/10.3390/e21040351

APA Style

Bania, P. (2019). Bayesian Input Design for Linear Dynamical Model Discrimination. Entropy, 21(4), 351. https://doi.org/10.3390/e21040351

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop