Next Article in Journal
Nitrogen Treatment by a Dry Detention Basin with Stormwater Wetland Characteristics
Previous Article in Journal
Hydrological Behavior of Karst Systems Identified by Statistical Analyses of Stable Isotope Monitoring Results
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Geometric Analysis of Conditional Bias-Informed Kalman Filters

1
Len Technologies, Oak Hill, VA 20171, USA
2
Department of Civil Engineering, University of Texas at Arlington, Arlington, TX 76019, USA
*
Author to whom correspondence should be addressed.
Hydrology 2022, 9(5), 84; https://doi.org/10.3390/hydrology9050084
Submission received: 19 February 2022 / Revised: 6 May 2022 / Accepted: 7 May 2022 / Published: 11 May 2022

Abstract

:
This paper presents a comparative geometric analysis of the conditional bias (CB)-informed Kalman filter (KF) with the Kalman filter (KF) in the Euclidean space. The CB-informed KFs considered include the CB-penalized KF (CBPKF) and its ensemble extension, the CB-penalized Ensemble KF (CBEnKF). The geometric illustration for the CBPKF is given for the bi-state model, composed of an observable state and an unobservable state. The CBPKF co-minimizes the error variance and the variance of the Type-II error. As such, CBPKF-updated state error vectors are larger than the KF-updated, the latter of which is based on minimizing the error variance only. Different error vectors in the Euclidean space imply different eigenvectors and covariance ellipses in the state space. To characterize the differences in geometric attributes between the two filters, numerical experiments were carried out using the Lorenz 63 model. The results show that the CBEnKF yields more accurate confidence regions for encompassing the truth, smaller errors in the ensemble mean, and larger norms for Kalman gain and error covariance matrices than the EnKF, particularly when assimilating highly uncertain observations.

1. Introduction

With highly uncertain observations and model dynamics, KF [1] and EnKF [2] estimates tend to be conditionally biased. The conditional bias (CB)-informed Kalman filter (KF) is designed to improve the estimation of extreme states in geophysical data assimilation, particularly at the onset of extreme phenomena. The CB-informed KFs developed to date include the CB-penalized Kalman filter (CBPKF; [3,4]) and its ensemble extension, the CB-penalized Ensemble Kalman filter (CBEnKF; [5]), and aim at addressing CB [3,4,5] to improve estimation of extreme states. With extreme precipitation, droughts, and floods occurring more frequently in many parts of the globe, the estimation and prediction of extremes is an increasingly important topic in hydrology. With the increasing availability of diverse sources of observations, hydrologic data assimilation has a large role to play in improving state estimation. The CB-informed KF is an effort to help address both.
There are two types of CB, Type I and Type II. The Type-I CB is defined as E[X|X* = x*] − x* where X, X*, and x* represent the unknown truth, the estimate, and the realization of X*, respectively. The Type-II CB is defined as E[X*|X = x] − x where x represents the realization of X. The Type-I CB is associated with false alarms which can be reduced by calibration. The Type-II CB is associated with failure to detect an event and cannot be reduced by calibration. The CBPKF minimizes a linearly weighted sum of the error covariance and the expectation of the Type-II error squared. With skillful specification of the weight for the latter, the CBPKF improves estimation of extremes over the KF while slightly increasing the unconditional MSE [3,4].
Though extensively studied in statistics, econometrics, meteorology, and hydrology [6,7,8,9,10,11,12,13,14,15,16], CB has gained interest only recently in data assimilation. Relevant studies to date include the CBPKF [3,4] and the CBEnKF [5]—the CBEnKF application to flood forecasting [5], adaptive filtering for the CBPKF [17] and the CBEnKF [18], and the variance-inflated KF-based approximation of the CBPKF for reduced computation [17,18]. Improving understanding of how the CB-informed KF compares with conventional data assimilation techniques, such as the KF, is crucial for interpreting and analyzing its application results, as well as identifying areas for further development. To that end, this study provides a geometric analysis of the CB-informed KF and compares performance with the EnKF via numerical experiments.
Stochastic variables may be expressed as vectors in the Euclidean domain, which can be used to geometrically illustrate vectors of innovation, observation errors, and forecast and analysis errors in the Euclidean domain [19]. The geometric interpretation of state estimation thus helps understand how Kalman gain shape to analysis error vectors and assimilation solutions for observable as well as unobservable states [19]. Error covariance matrices may also be analyzed geometrically in the state space and characterized with eigenvalues, eigenvectors, and associated confidence regions.
The purpose of this work is to gain additional insight into intuitive understanding of how the CBPKF solution differs from the KF, by casting them in error and state spaces using a bi-state model, and to advance understanding of the comparative performance of the CBEnKF with the EnKF by identifying and characterizing the representative geometric attributes in the state space via eigenvalue analysis. The new and significant contributions of this work are: comparative geometric analysis of the CBPKF solution with the KF, development of a set of geometry-based relationships for improved understanding of the CBPKF solution, and geometric characterization of ensemble analysis for improved understanding of the comparative performance of the CBEnKF with the EnKF. The comparative performance is based on numerical experiments with the Lorenz 63 model, chosen in this study for familiarity and simplicity. For real-world flood forecasting applications of the CBEnKF, the interested reader is referred to Lee et al. [5] and Shen et al. [18].
This paper is organized as follows. Section 2 describes the state updating problem, and the two solution approaches, the KF and the CBPKF. Section 3 presents the geometric analysis of the CB-informed KF in relation to the KF and the EnKF. First, Section 3.1 describes how to use vectors to represent stochastic variables in the Euclidean domain, and Section 3.2 expresses the filter equations in terms of state and observation error terms. Section 3.3 geometrically illustrates the filter equations for a low-order model based on the visualization of error vectors and their relations to Kalman gain and covariance matrices in the Euclidean space [19]. Section 3.4 describes geometric analyses in the state-space with eigenvalues and eigenvectors of error covariance matrices and associated confidence regions. In Section 3.5, an example is given for geometric analysis in the state space in which the CBEnKF, the EnKF, and Open Loop (OL) are compared for the Lorenz 63 model via numerical experiments. Finally, Section 4 describes conclusions.

2. Methodology

2.1. State Updating Problem

The nonlinear dynamical model is written as Equation (1).
X k = M ( X k 1 ) + W k 1
In the above, X k denotes the ( n c × 1 ) model state, or control vector, where n c denotes the number of variables in the control vector, M( ) denotes the dynamical model for state variables, and Wk−1 denotes the dynamical model error at a time step k − 1 with a mean of zero and a covariance of Q k . The nonlinear observational model is written as:
Z k = H k ( X k ) + V k
In the above, Zk denotes the (n × 1) observation vector, where n denotes the total number of observations. Vk denotes the (n × 1) observation error vector at a time step k with a mean of zero and a covariance of R k . H k () denotes the nonlinear observation operator that maps X k to Z k . This study solves the linear observation model which renders H k ( X k ) in Equation (2) linear, i.e., H k X k :
Z k = H k X k + V k
In the case of a nonlinear observation operator, e.g., soil moisture–streamflow transformation, one can still render Equation (2) linear via state augmentation [5,20,21], which is beyond the scope of this study. Interested readers are referred to Lee et al. [5], where state augmentation is used to solve a highly nonlinear flood forecasting problem.
Equation (4) shows the state updating equation where X k | k , X k | k 1 , and K k represent updated state at a time step k, state forecast from k − 1 to k, and Kalman gain, respectively.
X k | k = X k | k 1 + K k [ Z k H k X k | k 1 ]
To find K k , the KF and the EnKF minimize the error variance ( Σ E V ) in Equation (5). The CBPKF and the CBEnKF minimize the weighted sum of Σ E V and the expectation of the Type-II CB squared ( Σ C B ) in Equation (6), or Σ E V + α Σ C B where α is the weight given to the CB penalty term. The α can be estimated using an iterative method that yields the error covariance within theoretically expected bounds (see [5] for details). Since the iterative method is computationally expensive, the adaptive filtering method has been developed, and the results are reported in Shen et al. [17].
Σ E V = E [ ( X k X k | k ) ( X k X k | k ) T ]
Σ C B = E [ ( X k E X k | k [ X k | k | X k ] ) ( X k E X k | k [ X k | k | X k ] ) T ]
The following sections present expressions for Kalman gain ( K k ) and covariance ( P k | k ) matrices in the case of the KF and the CBPKF.

2.2. Kalman Filter, KF

Equation (4) is rewritten as Equation (7) but with the superscript K to denote the KF.
X k | k K = X k | k 1 + K k K [ Z k H k X k | k 1 ]
Minimizing Σ E V in Equation (5) results in Kalman gain K k K in Equation (8) and the covariance analysis matrix P k | k in Equation (9) [1].
K k K = P k | k 1 H T [ H P k | k 1 H T + R ] 1
P k | k = [ 1 K k H ] P k | k 1 = P k | k 1 K k [ H P k | k 1 H T + R k ] K k T

2.3. Conditional Bias-Penalized Kalman Filter, CBPKF

State updating equation (Equation (4)) can be rewritten into Equation (10) with the superscript C denoting the CBPKF.
X k | k C = X k | k 1 + K k C [ Z k H k X k | k 1 ]
Minimizing Σ E V + α Σ C B produces the CBPKF gain K k C in Equation (11) and the covariance matrix Σ k | k in Equation (12) [3,4,17].
K k C = [ ϖ 1 , k H k + ϖ 2 , k ] 1 ϖ 1 , k
Σ k | k = [ ϖ 1 , k H k + ϖ 2 , k ] 1 ( ϖ 1 , k R k ϖ 1 , k T + ϖ 2 , k Σ k | k 1 ϖ 2 , k T ) [ ϖ 1 , k H k + ϖ 2 , k ] 1
In Equations (11) and (12), the (m × n) and (m × m) weight matrices for the observations and model prediction, ϖ 1 , k and ϖ 2 , k , respectively, are given by:
ϖ 1 , k = H ^ k T Γ 11 , k + Γ 21 , k
ϖ 2 , k = H ^ k T Γ 12 , k + Γ 22 , k
In Equations (13) and (14), the (m × n) modified observation matrix, H ^ k T , and the (n × n), (n × m), and (m × m) matrices, Γ 11 , k , Γ 12 , k   ( = Γ 21 , k T ) , and Γ 22 , k , respectively, are given by:
H ^ k T = H k T + α C k T
Γ 22 , k 1 = Λ 22 , k Λ 21 , k Λ 11 , k 1 Λ 12 , k
Γ 11 , k = Λ 11 , k 1 + Λ 11 , k 1 Λ 12 , k Γ 22 , k Λ 21 , k Λ 11 , k 1
Γ 12 , k = Λ 11 , k 1 Λ 12 , k Γ 22 , k
In Equations (15)–(18), the (n × m) CB gain matrix for the observation vector, C k , and the (n × n), (n × m), and (m × m) modified error covariance matrices, Λ 11 , k , Λ 12 , k   ( = Λ 12 , k T ) , and Λ 22 , k , respectively, are given by:
C k = [ H k Σ k | k 1 G 2 , k 1 + R k H k ] [ G 2 , k 1 Σ k | k 1 G 2 , k 1 + 2 ( H k T R k H k + Σ k | k 1 ) ] 1 G 2 , k 1
Λ 11 , k = R k + α ( 1 α ) C k Σ k | k 1 C k T + Λ 12 , k H k T + H k Λ 21 , k
Λ 12 , k = α C k Σ k | k 1
Λ 22 , k = Σ k | k 1
In Equations (19)–(22), R k and Σ k | k 1 denote the (n × n) observation error covariance matrix, C o v [ V k , V k ] , and the (m × m) forecast error covariance matrix, respectively, and G 2 , k 1 = H k T H k + I .
If α = 0 , the CBPKF solution is reduced to the KF, i.e., Σ k | k = [ H k T R k 1 H k + Σ k | k 1 1 ] 1 , and K k = Σ k | k H k T R k 1 [5].

3. Geometric Analysis of the CB-Informed KF

3.1. Error Representation of Filter Equations

Since the error covariance of stochastic variables plays the key role in geometric analyses in the Euclidean space, this section defines errors in states and a residual in measurement in order to rewrite state updating and observation equations with error terms.
ε k C = X k X k | k C
ε k K = X k X k | k K
ε k = X k X k | k 1
y k = Z k H k X k | k 1
In the above, ε k C , ε k K , and ε k denote the CBPKF analysis error, the KF analysis error, and the forecast error, respectively; y k represents innovation; X k , X k | k C , X k | k K , and X k | k 1 represent the truth, the CBPKF-updated state, the KF-updated state, and the state forecast, respectively.
State update equations for the KF and the CBPKF are rewritten into Equations (27) and (28), assuming the same a priori states used in both filters.
ε k K = ε k K k K y k
ε k C = ε k K k C y k
where K k K and K k C are Kalman gains from the KF and the CBPKF, respectively.
Updated states from both filters have the following mathematical relationship.
ε k C = ε k K ( K k C K k K ) y k
The observation model in Equation (3) can be rewritten as follows:
y k = H k ε k + V k
From Equation (3), the covariance matrix of y k , or D k , can be written as
D k = H k P k H k T + R k
where P k is the covariance matrix of the state forecast.

3.2. KF and CBPKF Solutions for a Bi-State Model

A bi-state model ( X k = [ x 1 , k   x 2 , k ] T ) is used to illustrate the geometric relation of errors in states and observations from the CBPKF and the KF. Observation z 1 , k exists for a state x 1 , k but not for x 2 , k to investigate how state updating can be illustrated in the two-dimensional (2-D) Euclidean space for observable as well as unobservable states, i.e., Z k = [ z 1 , k ] ;   H = [ 1   0 ] is used for simplicity. Kronhamn [19] used the same X k ,   Z k and H matrices as those described in this Section for the geometric illustration of the KF, whereas we focus on the CBPKF in reference to the KF. For the bi-state model with a forecast error covariance P f , k = [ σ 1 , k 2 σ 12 , k σ 12 , k σ 2 , k 2 ] and the observation error R k = [ σ z , k 2 ] , matrices of Kalman gain (Equations (32) and (33)), analysis error covariance (Equations (34) and (35)), and correlation (Equations (36)–(38)) for the CBPKF and the KF are evaluated in the equations below where σ 12 , k   =   ρ 12 , k σ 1 , k σ 2 , k with ρ 12 , k representing the correlation between x 1 , k and x 2 , k .
K k C = [ K 1 , k C K 2 , k C ] = K k K + α [ 2 σ 1 , k 2 σ z , k 2 ( σ 1 , k 2 + σ z , k 2 ) ( ( 1 + 2 α ) σ 1 , k 2 + σ z , k 2 ) 2 σ 12 , k σ z , k 2 ( σ 1 , k 2 + σ z , k 2 ) ( ( 1 + 2 α ) σ 1 , k 2 + σ z , k 2 ) ]
where
K k K = [ K 1 , k K K 2 , k K ] = [ σ 1 , k 2 σ 1 , k 2 + σ z , k 2 σ 12 , k σ 1 , k 2 + σ z , k 2 ]
Equation (32) indicates that for an observable state x 1 , k , K k C is always larger than K k K , but for an unobservable state x 2 , k , the sign of K k C K k K depends on the sign of σ 12 , k . The CBPKF covariance matrix P a , k C and its relation to the KF-equivalent P a , k K are given in Equation (34).
P a , k C = P a , k K + α [ α ( 2 σ 1 , k 2 σ z , k 2 ) 2 ( σ 1 , k 2 + σ z , k 2 ) ( ( 1 + 2 α ) σ 1 , k 2 + σ z , k 2 ) 2         α σ 12 , k ( 2 σ 1 , k σ z , k 2 ) 2 ( σ 1 , k 2 + σ z , k 2 ) ( ( 1 + 2 α ) σ 1 , k 2 + σ z , k 2 ) 2 α σ 12 , k ( 2 σ 1 , k σ z , k 2 ) 2 ( σ 1 , k 2 + σ z , k 2 ) ( ( 1 + 2 α ) σ 1 , k 2 + σ z , k 2 ) 2         α ( 2 σ 12 , k σ z , k 2 ) 2 ( σ 1 , k 2 + σ z , k 2 ) ( ( 1 + 2 α ) σ 1 , k 2 + σ z , k 2 ) 2 ]
where
P a , k K = [ σ 1 , k 2 σ z , k 2 σ 1 , k 2 + σ z , k 2                               σ 12 , k σ z , k 2 σ 1 , k 2 + σ z , k 2 σ 21 , k σ z , k 2 σ 1 , k 2 + σ z , k 2                   σ 2 , k 2 σ 12 , k 2 σ 1 , k 2 + σ z , k 2 ]
In Equation (34), variances of both updated states x 1 , k and x 2 , k are larger than KF-equivalents, and the sign of covariance terms of P a , k C P a , k K depend on the sign of σ 12 , k . Equations (34) and (35) imply that the CBPKF-updated state ensembles have larger spreads in the state space than the KF-updated, and that the matrix norm of P a , k C is larger than that of P a , k K . From covariance matrices, Pearson product-moment correlation matrix C a , k C can be computed by Equation (36).
C a , k C = ( d i a g ( P a , k C ) ) 1 2 P a , k C ( d i a g ( P a , k C ) ) 1 2 = [ 1                       C a , 12 , k C C a , 12 , k C                       1 ]
where diag( ) denotes a diagonal matrix. Equation (37) shows that the correlation coefficient of the CBPKF-updated states C a , 12 , k C can be either larger or smaller than the KF-equivalent C a , 12 , k K .
C a , 12 , k C = σ 12 , k σ z , k ( ( 1 + 2 α ) 2 σ 1 , k 2 + σ z , k 2 ) σ 1 , k ( σ 1 , k 2 + σ z , k 2 ) ( σ 2 , k 2 ( σ 1 , k 2 + σ z , k 2 ) σ 12 , k 2 ) + 4 α ( 1 + α ) ( σ 1 , k 2 σ z , k 2 ) ( σ 1 , k 2 σ 2 , k 2 σ 12 , k 2 ) = C a , 12 , k K + σ 12 , k σ z , k A 1 A 2 σ 1 , k σ 2 , k 2 ( σ 1 , k 2 + σ z , k 2 ) σ 12 , k 2 ( σ 2 , k 2 ( σ 1 , k 2 + σ z , k 2 ) σ 12 , k 2 ) ( σ 1 , k 2 + σ z , k 2 ) + 4 α ( 1 + α ) ( σ 1 , k 2 σ z , k 2 ) ( σ 1 , k 2 σ 2 , k 2 σ 12 , k 2 )
where
A 1 = ( σ 2 , k 2 ( σ 1 , k 2 + σ z , k 2 ) σ 12 , k 2 ) ( ( 1 + 2 α ) 2 σ 1 , k 2 + σ z , k 2 )
A 2 = ( σ 2 , k 2 ( σ 1 , k 2 + σ z , k 2 ) σ 12 , k 2 ) ( σ 1 , k 2 + σ z , k 2 ) + 4 α ( 1 + α ) ( σ 1 , k 2 σ z , k 2 ) ( σ 1 , k 2 σ 2 , k 2 σ 12 , k 2 )
C a , 12 , k K = σ 12 , k σ z , k σ 1 , k σ 2 , k 2 ( σ 1 , k 2 + σ z , k 2 ) σ 12 , k 2
From Equations (32)–(38), if α = 0 , then K k C = K k K , P a , k C = P a , k K , and C a , k C = C a , k K .

3.3. Geometric Representation of KF and CBPKF Solutions

This Section begins with describing a relation between stochastic variables and vectors in the Euclidean domain at Equations (39)–(41), and then describes the geometric representation of KF and CBPKF equations. The covariance of stochastic variables x s and y s can be used to compute the scalar product of two vectors x E and y E in the Euclidean domain (Equation (39)). The vector norm | | x E | | corresponds to the standard deviation of x s (Equation (40)). The angle of x E and y E can be computed from the correlation of x s and y s (Equation (41)) [19,22]. Figure 1 shows the vector representation of stochastic variables in the Euclidean space.
x E · y E = | | x E | |   | | y E | | cos θ = C o v [ x s , y s ] = σ x s y s
| | x E | | = x E · x E = C o v [ x s , x s ] = σ x s
θ = cos 1 ( x E · y E | | x E | |   | | y E | | ) = cos 1 ( σ x s y s σ x s σ y s ) = cos 1 ( ρ x s y s )
Figure 2 shows the error vectors from (a) the KF and (b) the CBPKF for the observable state x 1 , k of the bi-state model, where | | ε 1 , k | | = σ 1 , k , | | v 1 , k | | = σ z , k , and | | y 1 , k | | 2 = ( σ 1 , k 2 + σ z , k 2 ) , where the last equality is from Equation (30). In Figure 2a, the state forecast error vector ε 1 , k is orthogonal to the observation error vector v 1 , k owing to the independence assumption. Being a minimum variance solution,   ε 1 , k K is orthogonal to y 1 , k [19,23]. Figure 2a also shows that the forecast error is a vector sum of the gain-weighted innovation and the analysis error, i.e., ε 1 , k = K 1 , k K y 1 , k + ε 1 , k K as expected from Equation (27). The KF analysis error variance may be obtained in Figure 2a via the Pythagorean theorem as | | ε 1 , k K | | 2 = | | ε 1 , k | | 2 | | K 1 , k K y 1 , k | | 2 = σ 1 , k 2 ( σ 1 , k 2 σ 1 , k 2 + σ z , k 2 ) 2 ( σ 1 , k 2 + σ z , k 2 ) = σ 1 , k 2 σ z , k 2 σ 1 , k 2 + σ z , k 2 where K 1 , k K = σ 1 , k 2 σ 1 , k 2 + σ z , k 2 as expected from Equation (35). Since ε 1 , k C = ε 1 , k K ( K 1 , k C K 1 , k K ) y 1 , k , we may write via the Pythagorean theorem | | ε 1 , k C | | = | | ε 1 , k K | | 2 + | | ( K 1 , k C K 1 , k K ) y 1 , k | | 2 = ( ( 1 + 2 α ) 2 σ 1 , k 2 + σ z , k 2 ) σ 1 , k 2 σ z , k 2 ( ( 1 + 2 α ) σ 1 , k 2 + σ z , k 2 ) 2 which is expected from Equation (34). In Figure 2, the inequality, | | K 1 , k K y 1 , k | | < | | K 1 , k C y 1 , k | | , arises due to the fact that the CBPKF solution minimizes not the error variance but a weighted sum of the error variance and the variance of the Type-II CB.
Figure 3 is the same as Figure 2 but for the unobservable state x 2 , k where | | ε 2 , k | | = σ 2 , k ; | | ε 2 p , k | | = | | ε 2 , k | | cos θ = σ 12 , k σ 1 , k ; | | ε 2 o , k | | = | | ε 2 , k | | 2 | | ε 2 p , k | | 2 = σ 2 , k 2 σ 12 , k 2 σ 1 , k 2 ; and | | ε 2 c , k | | = | | ε 2 p , k | | 2 | | K 2 , k K y 1 , k | | 2 = σ 12 , k 2 σ z , k 2 σ 1 , k 2 ( σ 1 , k 2 + σ z , k 2 ) . In Figure 3a, the proportionality, | | ε 2 p , k | | : | | K 2 , k K y 1 , k | | = | | ε 1 , k | | : | | K 1 , k K y 1 , k | | , gives K 2 , k K = K 1 , k K | | ε 2 p , k | | | | ε 1 , k | | = K 1 , k K | | ε 2 , k | | cos θ | | ε 1 , k | | = K 1 , k K σ 2 , k σ 1 , k σ 12 , k σ 1 , k σ 2 , k = K 1 , k K σ 12 , k σ 1 , k 2 as expected from Equation (33). In Figure 3a, the orthogonality, ε 2 , k K y 1 , k , and the equality ε 2 , k K = ε 2 , k K 2 , k K y 1 , k , yield | | ε 2 , k K | | 2 = | | ε 2 , k | | 2 | | K 2 , k K y 1 , k | | 2 = σ 2 , k 2 σ 12 , k 2 σ 1 , k 2 + σ z , k 2 in agreement with Equation (35). In Figure 3b, | | ε 2 , k C | | 2 may be written via the Pythagorean theorem as:
| | ε 2 , k C | | 2 = | | ε 2 , k K | | 2 + | | ( K 2 , k C K 2 , k K ) y 1 , k | | 2
Using Equations (35), (32) and (31) for | | ε 2 , k K | | 2 , K 2 , k C K 2 , k K , and | | y 1 , k | | 2 , respectively, we may rewrite Equation (42), in agreement with Equation (34), as:
| | ε 2 , k C | | 2 = σ 2 , k 2 σ 12 , k 2 σ 1 , k 2 + σ z , k 2 + ( 2 α σ 12 , k σ z , k 2 ) 2 ( σ 1 , k 2 + σ z , k 2 ) ( ( 1 + 2 α ) σ 1 , k 2 + σ z , k 2 ) 2
Figure 2 and Figure 3 show that the KF- and the CBPKF-updated state error vectors point to different directions in the state space. Figure 4 shows the updated state error vectors in Figure 2 and Figure 3 to visually compare the differences in the angle, the magnitude, and the direction. The angles of the two-state error vectors for the CBPKF ( θ C in Equation (44)) and the KF ( θ K in Equation (45)) can be computed from the correlation C a , 12 , k C and C a , 12 , k K in Equations (37) and (38), respectively:
θ C = cos 1 C a , 12 , k C = cos 1 ( ε 1 , k C · ε 2 , k C | | ε 1 , k C | |   | | ε 2 , k C | | ) = cos 1 ( σ x 1 , k | k C x 2 , k | k C σ x 1 , k | k C σ x 2 , k | k C )
θ K = cos 1 C a , 12 , k K = cos 1 ( ε 1 , k K · ε 2 , k K | | ε 1 , k K | |   | | ε 2 , k K | | ) = cos 1 ( σ x 1 , k K x 2 , k K σ x 1 , k | k K σ x 2 , k | k K )
Below, we develop a set of geometric expressions in the 2-D state space for the analysis error covariance via eigenvalue decomposition (EVD, [24]).

3.4. Geometric Analysis in the State Space

Geometric characteristics of state ensembles in the 2-D state space can be quantified by confidence regions (CR), eigenvectors, eigenvalues, and the angle between the eigenvector and the basis vector of the x-axis. Assuming normal distributions for state ensembles in a 2-D state space, a CR, or so-called a covariance ellipse, can be constructed based on the EVD of a covariance matrix as well as the Chi-Square probability table. The presence of the CB results in different variances (eigenvalues) and directions (eigenvectors) of updated state ensembles in a 2-D state space. The major and minor axis lengths of the covariance ellipse are 2 s λ 1 and 2 s λ 2 where λ 1 > λ 2 and the value of s is from the Chi-Square probability table for a given confidence region, e.g., s   = 4.605 for a 90% confidence region given the Chi-Square probability P ( s < 4.605 ) = 0.9 in the case of degrees of freedom of 2. In a 2-D state space, the error in the orientation of the covariance ellipse with respect to the truth can be estimated by the angle, θ, between the largest eigenvector u 1 and the vector a connecting the truth and the ensemble mean, i.e., θ ( u 1 , a ) = cos 1 ( u 1 · a | | u 1 | |   | | a | | ) .
The EVD of the CBPKF analysis covariance P a , k C may be written as:
P a , k C = U E U T
where
U = [ cos θ       sin θ sin θ     cos θ ] = [ u 1     u 2 ] ,   E = [ λ 1 C     0 0     λ 2 C ] ,
In the above, U is the eigenvector matrix which rotates the white data (W), or uncorrelated standard normal variates by θ . The eigenvalue matrix E explains the variance along the principal error direction, or the direction of the eigenvector. In a 2-D state space, E is a scale factor applied to W. The dataset (D) resulting from scaling W by E and rotating by U , i.e., D = U E W , has the covariance matrix of P a , k C = U E ( U E ) T . Below we apply EVD to the CBPKF analysis error covariance, P a , k C , from the bi-state model in Section 3.2.
With λ 1 C > λ 2 C > 0 , and P a , k K = [ p 1 , k 2   p 12 , k p 12 , k   p 2 , k 2 ] , P a , k C in Equation (34) may be rewritten as:
P a , k C = P a , k K + ζ [ σ 1 , k 4                   σ 12 , k σ 1 , k 2 σ 12 , k σ 1 , k 2             σ 12 , k 2         ]   where   ζ = ( 2 α σ z , k 2 ) 2 ( σ 1 , k 2 + σ z , k 2 ) ( ( 1 + 2 α ) σ 1 , k 2 + σ z , k 2 ) 2
Appendix A describes how eigenvalues and eigenvectors may be evaluated for a 2 × 2 covariance matrix. Using Equation (A11) in Appendix A, we have for λ 1 C :
λ 1 C = 0.5 ( ( p 1 , k 2 + ζ σ 1 , k 4 ) 2 + ( p 2 , k 2 + ζ σ 12 , k 2 ) 2 ) + 0.5 ( p 1 , k 2 p 2 , k 2 + ζ σ 1 , k 4 ζ σ 12 , k 2 ) 2 + 4 ( p 12 , k + ζ σ 12 , k σ 1 , k 2 ) 2
The difference in the largest eigenvalue between the KF and the CBPKF analysis error covariance is given by:
λ 1 C λ 1 K = 0.5 ( 2 p 1 , k 2 ζ σ 1 , k 4 + 2 p 2 , k 2 ζ σ 12 , k 2 + ζ 2 σ 1 , k 8 + ζ 2 σ 12 , k 4 ) + 0.5 ( p 1 , k 2 p 2 , k 2 + ζ σ 1 , k 4 ζ σ 12 , k 2 ) 2 + 4 ( p 12 , k + ζ σ 12 , k σ 1 , k 2 ) 2 0.5 ( p 1 , k 2 p 2 , k 2 ) 2 + 4 ( p 12 , k ) 2
where λ 1 K denotes the largest eigenvalue of P a , k K . Using Equation (A12), we may write λ 2 C and λ 2 C λ 2 K :
λ 2 C = 0.5 ( ( p 1 , k 2 + ζ σ 1 , k 4 ) 2 + ( p 2 , k 2 + ζ σ 12 , k 2 ) 2 ) 0.5 ( p 1 , k 2 p 2 , k 2 + ζ σ 1 , k 4 ζ σ 12 , k 2 ) 2 + 4 ( p 12 , k + ζ σ 12 , k σ 1 , k 2 ) 2
λ 2 C λ 2 K = 0.5 ( 2 p 1 , k 2 ζ σ 1 , k 4 + 2 p 2 , k 2 ζ σ 12 , k 2 + ζ 2 σ 1 , k 8 + ζ 2 σ 12 , k 4 ) 0.5 ( p 1 , k 2 p 2 , k 2 + ζ σ 1 , k 4 ζ σ 12 , k 2 ) 2 + 4 ( p 12 , k + ζ σ 12 , k σ 1 , k 2 ) 2 + 0.5 ( p 1 , k 2 p 2 , k 2 ) 2 + 4 ( p 12 , k ) 2
Similarly, using Equation (A14), we may write θ a , k C ( u 1 , i ) and θ a , k C ( u 1 , i ) θ a , k K ( u 1 , i ) as:
θ a , k C ( u 1 , i ) = 1 2 tan 1 ( 2 ( p 12 , k + ζ σ 12 , k σ 1 , k 2 ) p 1 , k 2 + ζ σ 1 , k 4 p 2 , k 2 ζ σ 12 , k 2 )
θ a , k C ( u 1 , i ) θ a , k K ( u 1 , i ) = 1 2 tan 1 ( 2 ( p 12 , k + ζ σ 12 , k σ 1 , k 2 ) p 1 , k 2 + ζ σ 1 , k 4 p 2 , k 2 ζ σ 12 , k 2 ) 1 2 tan 1 ( 2 p 12 , k p 1 , k 2 p 2 , k 2 )
where the vector i is the basis vector of the x-axis and θ a , k C ( u 1 , i ) θ a , k K ( u 1 , i ) is equal to θ a , k C ( u 1 , a ) θ a , k K ( u 1 , a ) . With the geometric attributes established above, we now carry out the comparative geometric analysis of the KF and the CBPKF analysis results using the Lorenz 63 model [25].
With the EVD of P a , k C and the ensemble mean ( x 1 ¯ , x 2 ¯ ) , the minimum percentage confidence C R M I N to contain the verifying truth ( x 1 , T , x 2 , T ) within the confidence region can be computed by C R M I N = 100 × P ( s < s ˜ ) , where P ( s < s ˜ ) is the Chi-Square probability with degrees of freedom of 2; s ˜ satisfies Equation (55).
{ ( x 1 , T x 1 ¯ ) cos θ a , k C + ( x 2 , T x 2 ¯ ) sin θ a , k C } 2 s ˜ λ 1 C + { ( x 1 , T x 1 ¯ ) sin θ a , k C ( x 2 , T x 2 ¯ ) cos θ a , k C } 2 s ˜ λ 2 C = 1
Figure 5 shows an example of computing a , θ ( u 1 , a ) , C R M I N , and eigenvalues and eigenvectors of a covariance matrix.

3.5. Numerical Experiment with the Lorenz 63 Model

In the sections above, a linear two-state model was used for theoretical simplicity. In this section, we use the three-state Lorenz 63 model to illustrate the differences between the EnKF and the CBEnKF solutions in terms of the geometric attributes introduced above. In this experiment, synthetically generated observations of all three states in the Lorenz 63 model were assimilated at every time step using the EnKF and the CBEnKF [26]. Preliminary experiments suggested observation error variances ( σ z 2 ) of 10 or 400 can be used for the cases of assimilating less uncertain or largely uncertain observations, respectively, based on the ensemble spread. To render the assimilation problem more challenging, σ z 2 = 400 is used to compare the performance of the CBEnKF to that of the EnKF in Figure 6, Figure 7, Figure 8, Figure 9, Figure 10 and Figure 11, where the ensemble size ( n S ) used is 2000 to minimize filter performance degradation owing to a small ensemble size. Figure 12 and Figure 13 compare assimilation results from the two filters in the case of σ z 2 = 10 or 400 as a function of an ensemble size, where n S values used include 10, 20, 30, 50, 70, 100, 200, 300, 500, 700, 1000, and 2000.
Figure 6 presents the results for states x 1 and x 2 of the Lorenz 63 model in a 2-D state space. The attributes shown include | a h | , | a v | , C R M I N , λ 1 , θ , ρ , and λ 1 u 1 ; | a h | , and | a v | reflect errors in the ensemble mean measured along horizontal or vertical directions, respectively; large dots overlaid in the scatter plots represent mean values of the samples in each of the ten bins equally dividing the entire state space. Figure 7 and Figure 8 are the same as Figure 6 but for the state spaces of ( x 1 , x 3 ) and ( x 2 , x 3 ) , respectively. The following summarizes general observations from Figure 6, Figure 7 and Figure 8. The spread of | a h | and | a v | in the scatter plots shows that the CBEnKF reduces CBs more effectively than the EnKF or the OL, particularly in the extremes. In some cases, however, their mean values appear similar, e.g., | a h | in the state spaces of ( x 1 , x 3 ) or ( x 2 , x 3 ) . The OL results for | a h | and | a v | show that their patterns appearing in the state space are similar to those of the state space plot in the top left of Figure 6, Figure 7 and Figure 8; this indicates the notable dependency of the amount of ensemble mean errors on model dynamics, e.g., the larger ensemble mean errors at the extreme—this is also seen in EnKF solutions but less so in the CBEnKF. This may be explained by the CBEnKF with a larger weight to observations than the EnKF in the case of largely uncertain observations ( σ z 2 = 400 ), which reduces the reliance of CBEnKF solutions on the model dynamics. Based on λ 1 , CBEnKF covariances are generally larger than the EnKF at all state spaces. Larger λ 1 and smaller | a h | and | a v | of the CBEnKF than the EnKF yield consistently smaller C R M I N than the EnKF at both extremes and the median of all three variables. This signifies the benefit of using the CB-informed KF for the estimation of extremes given that the EnKF’s C R M I N quickly increases towards extremes, i.e., the EnKF is less confident in estimating extremes than the CBEnKF. For example, 3% confidence regions for selected extreme values presented in the bottom right plots show the truth (green dots) contained within the CBEnKF’s confidence regions (red ellipses) but not within the EnKF’s (blue ellipses). At the plots, arrows represent λ 1 u 1 . θ and ρ do not clearly indicate differences between the two filters.
Figure 9 shows C R M I N , | a | , λ 1 , θ , and ρ of Figure 6, Figure 7 and Figure 8 but as a function of exceedance probabilities to highlight CB-informed KF performances at extremes. At extremes with low exceedance probabilities, differences between the CBEnKF and the EnKF are vivid in the case of | a | and C R M I N . On the other hand, λ 1 of the CBEnKF is consistently larger than those of the EnKF and the OL across exceedance probabilities. As exceedance probabilities increase, the EnKF’s | a | becomes similar to the CBEnKF’s, implying unconditionally less biased. The CBEnKF keeps consistently low C R M I N at all exceedance probabilities owing to small | a | and large λ 1 , compared to the EnKF or the OL. Since the EnKF seeks orthogonal solutions to minimize analysis covariances, its λ 1 is always smaller than the OL’s as well as the CBEnKF’s. On the other hand, the CBEnKF increases λ 1 to address CBs which helps keep C R M I N low to contain the truth. Both θ and ρ show no consistent patterns across different state spaces as well as exceedance probabilities.
In Figure 10, the ensemble mean error time series indicates that among the three variables, CBEnKF’s improvement is the largest for x 1 . On the other hand, for the state x 3 , the CBEnKF mainly remedies the underestimation of x 3 compared to the EnKF. In the case of x 2 , the CBEnKF slightly outperforms the EnKF. These observations may imply the different amounts of CBs present in different states, hence the need of applying a separate weight α to the CB penalty for the individual state, which warrants a future effort. To compare P a , k and K k from the two filters, the time series of Frobenius norm of P a , k and K k is computed by Equations (56) and (57), respectively. Compared to the EnKF, the CBEnKF yields | | K k | | F and | | P a , k | | F consistently larger at all assimilation cycles, and the mean values of | | K k | | F and | | P a , k | | F are five and three times larger, respectively.
| | P a , k | | F = i = 1 3 λ i 2
| | K k | | F = i = 1 3 λ i 2
Figure 11 shows mean | | K k | | F , and | | P a , k | | F as a function of exceedance probabilities. At extremes, both the CBEnKF and the EnKF show that mean | | K k | | F and | | P a , k | | F are larger than those at high exceedance probabilities, and that large differences in mean | | K k | | F and | | P a , k | | F between the CBEnKF and the EnKF are consistent across exceedance probabilities.
Figure 6, Figure 7, Figure 8, Figure 9, Figure 10 and Figure 11 are based on the case of uncertain observations ( σ z 2 = 400 ) where the CBEnKF may supposedly outperform the EnKF. To explore the CBEnKF performance with less uncertain observations ( σ z 2 = 10 ) and also to see the sensitivity to the ensemble size ( n S ), Figure 12 presents results from the combination of n S = 10, 20, 30, 50, 70, 100, 200, 300, 500, 700, 1000, and 2000, and σ z 2 = 10 and 400. In Figure 12, | a | plots indicate that with σ z 2 = 10 , the accuracy of the ensemble mean continuously increases with an increase of n S at both cases of extremes (an exceedance probability of 0.1; red and blue dots for the CBEnKF and the EnKF, respectively) and all data (red and blue lines for the CBEnKF and the EnKF, respectively). When σ z 2 = 10 , the EnKF’s | a | is slightly smaller than the CBEnKF’s, but the CBEnKF’s λ 1 is slightly larger than the EnKF’s. The resulting C R M I N from both filters are very similar. This implies when observations are less uncertain, the EnKF solutions are as accurate and as confident as the CBEnKF solutions at extremes as well as the whole range. When n S 200 and σ z 2 = 10 , mean C R M I N maintains ~1%. When n S < 200 and σ z 2 = 10 , C R M I N quickly increases with a decrease of n S because of inaccurate error covariance estimates with an insufficient ensemble size. When observations are largely uncertain ( σ z 2 = 2000 ) , the CBEnKF clearly shows more accurate ensemble means (smaller | a | ) and higher confidence in covariance estimates (smaller C R M I N ) than the EnKF, particularly at extremes. Compared to σ z 2 = 10 , assimilating largely uncertain observations ( σ z 2 = 2000 ) reduces accuracies in covariance estimates, resulting in larger λ 1 in both filters, although the CBEnKF’s λ 1 addressing the CB is larger than the EnKF’s. When σ z 2 = 2000 , | a | and C R M I N tend to be less sensitive to n S than the case of σ z 2 = 10 . Both θ and ρ show neither any consistent patterns nor sensitivities to n S , but are included in Figure 12 for completeness.
Finally, Figure 13 presents mean | | K k | | F and | | P a , k | | F as a function of n S . Compared to the results from σ z 2 = 2000 , σ z 2 = 10 results in larger | | K k | | F in both filters due to bigger weights to the observations. When σ z 2 = 2000 , the CBEnKF maintains relatively large | | K k | | F to account for the CB; however, the EnKF’s | | K k | | F is conspicuously small. Both | | K k | | F and | | P a , k | | F tend to be little sensitive to the ensemble size n S , except the all data case of the CBEnKF with σ z 2 = 2000 (pink line). With uncertain observations ( σ z 2 = 2000 ), the CBEnKF’s | | P a , k | | F becomes large at extremes (pink dots) as well as all data (pink line) at all n S values used to reflect CBs in all states.

4. Conclusions

Error covariance and gain matrices of two CB-informed KFs, i.e., the CBPKF and the CBEnKF, are geometrically illustrated and compared with the KF equivalents [19] for a bi-state model using error vectors in the Euclidean space. Geometric illustration and analysis offer an intuitive understanding of the relationship between the two filters. Unlike the KF, the CBPKF solution is not orthogonal to its error, which renders its error covariances and gains to be larger than the KF’s. The above differences result in different confidence regions and principal error directions in the state space. Synthetic sensitivity experiments with the Lorenz 63 model showed that the CBEnKF solutions have generally smaller errors in the ensemble mean, larger eigenvalues in the error covariance matrix, more accurate confidence regions for encompassing the truth, and larger Frobenius norms of the error covariance and gain matrices than the KF. The above differences are particularly pronounced when the observations are highly uncertain.
Future research recommendations include applying the CBPKF and the CBEnKF to diverse geophysical problems of estimating and predicting extremes, e.g., extreme precipitation or floods. The bi-state model was used in this work for a comparative geometric analysis of the CBPKF and the KF. Possible extension to an arbitrary number of states poses an interesting research topic.

Author Contributions

Conceptualization, H.L., H.S. and D.-J.S.; methodology, H.L. and D.-J.S.; software, H.L. and D.-J.S.; validation, H.L., H.S. and D.-J.S.; formal analysis, H.L.; investigation, H.L., H.S. and D.-J.S.; writing—original draft preparation, H.L.; writing—review and editing, H.L., H.S. and D.-J.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Nomenclature

SymbolDescription
Superscript
CConditional bias-penalized Kalman filter (CBPKF)
KKalman filter (KF)
TTranspose of a matrix
Forecast
Subscript
aAnalysis
EVariable in the Euclidean domain
hHorizontal
kTime step
sStochastic variable
TTruth
vVertical
a Vector connecting the ensemble mean and the truth
| a | Magnitude of the vector a
CPearson product-moment correlation matrix
CRConfidence region
CRMINMinimum percentage confidence required to contain the verifying truth
EExpected value
HObservation operator that maps model states X to the observation vector Z
i Basis vector of the x-axis
KKalman gain
| | K | | F Frobenius norm of K
MThe dynamical model
nThe total number of observations
ncThe number of variables in the control vector
nsEnsemble size
PCovariance matrix of state estimate
PCovariance matrix of state forecast
P(s < p)Chi-Square probability in the case of s smaller than p
QThe covariance of the dynamical model error W
RThe covariance of the observation error vector V
sThe value in the Chi-Square probability table for a given confidence region
u Eigenvector of the covariance matrix P
VThe observation error vector
WThe dynamical model error
xIndividual model state
| | x | | Norm of the vector x
x*Realization of an estimated state X*
XModel state vector
X*The vector of the estimated states
yInnovation
ZThe observation vector
αThe weight given to the CB penalty term in ΣEV + αΣCB
ɛError in a state
λEigenvalue
ρijCorrelation between variables i and j
σ i j Covariance of variables i and j
σ i 2 Variance of a variable i
σ Z 2 Observation error variance
ΣEVThe error variance
ΣCBThe expectation of the Type-II CB squared

Appendix A. Eigenvalue Decomposition of a Two-by-Two Covariance Matrix

The eigenvalue decomposition (EVD) of the covariance matrix P a , k = [ s 1 , k 2 s 12 , k s 12 , k s 2 , k 2 ] can be written as
P a , k = [ s 1 , k 2 s 12 , k s 12 , k s 2 , k 2 ] = U E U T
U = [ cos θ       sin θ sin θ     cos θ ] ,   E = [ λ 1     0 0     λ 2 ] ,  
U is the eigenvector matrix that can be interpreted as the rotation matrix applied to white data, or a standard normal and uncorrelated data. E is the eigenvalue matrix that explains the variance of the principal error direction, or the direction of the eigenvector. In the case of two-dimensional data, E can be interpreted as a scale factor applied to white data. In other words, the resulting dataset (D) from scaling white data (W) by E and then rotating it by U , i.e., D = U E W , will have the covariance matrix of P a , k = U E ( U E ) T .
Let λ1 > λ2 > 0, or Pa,k is positive definite
[ s 1 , k 2 s 12 , k s 12 , k s 2 , k 2 ] = [ cos θ       sin θ sin θ     cos θ ] [ λ 1     0 0     λ 2 ] [ cos θ sin θ sin θ             cos θ ]                                                                           = [ cos θ       sin θ sin θ     cos θ ] [ λ 1 cos θ   λ 1 sin θ λ 2 sin θ               λ 2 cos θ ]                                                                           = [ λ 1 cos 2 θ + λ 2 sin 2 θ           λ 1 sin θ cos θ + λ 2 sin θ cos θ λ 1 sin θ cos θ + λ 2 sin θ cos θ               λ 1 sin 2 θ + λ 2 cos 2 θ ]
s 1 , k 2 = λ 1 cos 2 θ + λ 2 sin 2 θ
s 12 , k = ( λ 2 λ 1 ) sin θ cos θ
s 2 , k 2 = λ 1 sin 2 θ + λ 2 cos 2 θ
A(3) + A(5), A(3) − A(5), and 2 × (A4) yield (A6), (A7), and (A8), respectively.
s 1 , k 2 + s 2 , k 2 = λ 1 + λ 2
s 1 , k 2 s 2 , k 2 = λ 1 ( cos 2 θ sin 2 θ ) λ 2 ( cos 2 θ sin 2 θ ) = ( λ 1 λ 2 ) cos 2 θ
2 s 12 , k = ( λ 2 λ 1 ) 2 sin θ cos θ = ( λ 2 λ 1 ) sin 2 θ
From (A7) and (A8)
( λ 1 λ 2 ) 2 = ( s 1 , k 2 s 2 , k 2 ) 2 + 4 s 12 , k 2
Since λ1 > λ2,
λ 1 λ 2 = ( s 1 , k 2 s 2 , k 2 ) 2 + 4 s 12 , k 2
(A6) + (A10)
λ 1 = 1 2 ( s 1 , k 2 + s 2 , k 2 + ( s 1 , k 2 s 2 , k 2 ) 2 + 4 s 12 , k 2 )
λ 2 = 1 2 ( s 1 , k 2 + s 2 , k 2 ( s 1 , k 2 s 2 , k 2 ) 2 + 4 s 12 , k 2 )
From (A7) and (A8)
tan 2 θ = 2 s 12 , k s 1 , k 2 s 2 , k 2
θ = 1 2 tan 1 ( 2 s 12 , k s 1 , k 2 s 2 , k 2 )

References

  1. Kalman, R.E. A new approach to linear filtering and prediction problems. J. Basic Eng. 1960, 82, 35–45. [Google Scholar] [CrossRef] [Green Version]
  2. Evensen, G. The Ensemble Kalman Filter: Theoretical formulation and practical implementation. Ocean Dyn. 2003, 53, 343–367. [Google Scholar] [CrossRef]
  3. Seo, D.-J.; Saifuddin, M.M.; Lee, H. Conditional bias-penalized Kalman filter for improved estimation and prediction of extremes. Stoch. Environ. Res. Risk Assess. 2018, 32, 183–201. [Google Scholar] [CrossRef]
  4. Seo, D.-J.; Saifuddin, M.M.; Lee, H. Correction to: Conditional bias-penalized Kalman filter for improved estimation and prediction of extremes. Stoch. Environ. Res. Risk Assess. 2018, 32, 3561–3562. [Google Scholar] [CrossRef] [Green Version]
  5. Lee, H.; Shen, H.; Noh, S.J.; Kim, S.; Seo, D.-J.; Zhang, Y. Improving flood forecasting using conditional bias-penalized ensemble Kalman filter. J. Hydrol. 2019, 575, 596–611. [Google Scholar] [CrossRef]
  6. Ciach, G.J.; Morrissey, M.L.; Krajewski, W.F. Conditional bias in radar rainfall estimation. J. Appl. Meteorol. 2000, 39, 1941–1946. [Google Scholar] [CrossRef]
  7. Fuller, W.A. Measurement Error Models; Wiley: Chichester, UK, 1987. [Google Scholar]
  8. Seber, G.A.F.; Wild, C.J. Nonlinear Regression; John Wiley & Sons: New York, NY, USA, 1989. [Google Scholar]
  9. Hausman, J.A. Mismeasured variables in econometric analysis: Problems from the right and problems from the left. J. Econ. Perspect. 2001, 15, 57–67. [Google Scholar] [CrossRef] [Green Version]
  10. Hughes, M.D. Regression dilution in the proportional hazards model. Biometrics 1993, 49, 1056–1066. [Google Scholar] [CrossRef] [PubMed]
  11. Frost, C.; Thompson, S. Correcting for regression dilution bias: Comparison of methods for a single predictor variable. J. R. Stat. Soc. Ser. A 2000, 163, 173–190. [Google Scholar] [CrossRef]
  12. Joliffe, I.T.; Stephenson, D.B. Forecast Verification: A Practitioner’s Guide in Atmospheric Science; Wiley: New York, NY, USA, 2003. [Google Scholar]
  13. Kim, B.; Seo, D.-J.; Noh, S.; Prat, O.P.; Nelson, B.R. Improving multisensor estimation of heavy-to-extreme precipitation via conditional bias-penalized optimal estimation. J. Hydrol. 2018, 556, 1096–1109. [Google Scholar] [CrossRef]
  14. Wilks, D.S. Statistical Methods in the Atmospheric Sciences, 2nd ed.; Academic Press: London, UK, 2006; p. 648. [Google Scholar]
  15. Brown, J.D.; Seo, D.-J. Evaluation of a nonparametric post-processor for bias correction and uncertainty estimation of hydrologic predictions. Hydrol. Processes 2013, 27, 83–105. [Google Scholar] [CrossRef]
  16. Seo, D.-J.; Siddique, R.; Zhang, Y.; Kim, D. Improving real-time estimation of heavy-to-extreme precipitation using rain gauge data via conditional bias-penalized optimal estimation. J. Hydrol. 2014, 519 Pt B, 1824–1835. [Google Scholar] [CrossRef]
  17. Shen, H.; Lee, H.; Seo, D.-J. Adaptive conditional bias-penalized Kalman filter for improved estimation of extremes and its approximation for reduced computation. Hydrology 2022, 9, 35. [Google Scholar] [CrossRef]
  18. Shen, H.; Seo, D.-J.; Lee, H.; Liu, Y.; Noh, S. Improving flood forecasting using conditional bias-aware assimilation of streamflow observations and dynamic assessment of flow-dependent information content. J. Hydrol. 2022, 605, 127247. [Google Scholar] [CrossRef]
  19. Kronhamn, T.R. Geometric illustration of the Kalman filter gain and covariance update algorithms. IEEE Control Syst. Mag. 1985, 5, 41–43. [Google Scholar] [CrossRef]
  20. Lorentzen, R.J.; Nævdal, G. An iterative ensemble Kalman filter. IEEE Trans. Autom. Control 2011, 56, 5–8. [Google Scholar] [CrossRef]
  21. Houtekamer, P.L.; Zhang, F. Review of the ensemble Kalman filter for atmospheric data assimilation. Mon. Wea. Rev. 2016, 144, 4489–4532. [Google Scholar] [CrossRef]
  22. Astrom, K.J. Introduction to Stochastic Control Theory; Academic Press: New York, NY, USA, 1970. [Google Scholar]
  23. Schweppe, F.C. Uncertain Dynamic Systems; Prentice-Hall: Englewood Cliffs, NJ, USA, 1973; p. 563. [Google Scholar]
  24. Kreyszig, E. Advanced Engineering Mathematics, 10th ed.; Wiley: New York, NY, USA, 2011. [Google Scholar]
  25. Lorenz, E.N. Deterministic nonperiodic flow. J. Atmos. Sci. 1963, 20, 130–141. [Google Scholar] [CrossRef] [Green Version]
  26. Shen, H.; Lee, H.; Noh, S.; Kim, S.; Seo, D.-J.; Zhang, Y. Conditional bias-penalized Kalman filter for improved state estimation over the tails of distribution. In Proceedings of the AGU Fall Meeting, Washington, DC, USA, 10–14 December 2018. [Google Scholar]
Figure 1. Vector representation of stochastic variables in the Euclidean space, where xE and yE are vectors in the Euclidean space and xS and yS are stochastic variables.
Figure 1. Vector representation of stochastic variables in the Euclidean space, where xE and yE are vectors in the Euclidean space and xS and yS are stochastic variables.
Hydrology 09 00084 g001
Figure 2. Geometric illustration of assimilation solutions for the observable state x1,k for (a) the KF (; after Kronhamn [19]), and (b) the CBPKF ().
Figure 2. Geometric illustration of assimilation solutions for the observable state x1,k for (a) the KF (; after Kronhamn [19]), and (b) the CBPKF ().
Hydrology 09 00084 g002
Figure 3. Geometric illustration of assimilation solutions for the unobservable state x2,k for (a) the KF (; after Kronhamn [19]), and (b) the CBPKF ().
Figure 3. Geometric illustration of assimilation solutions for the unobservable state x2,k for (a) the KF (; after Kronhamn [19]), and (b) the CBPKF ().
Hydrology 09 00084 g003
Figure 4. Vector representation of updated states from (a) the KF and (b) the CBPKF in Figure 3, where thick blue and red vectors indicate the principal error direction of the KF and the CBPKF in the Euclidean space, respectively.
Figure 4. Vector representation of updated states from (a) the KF and (b) the CBPKF in Figure 3, where thick blue and red vectors indicate the principal error direction of the KF and the CBPKF in the Euclidean space, respectively.
Hydrology 09 00084 g004
Figure 5. Illustration of geometric characteristics in a 2-D state space. Pink dots represent 2000 ensemble pairs of states ( x 1 , x 2 ) . The red and green dots denote the ensemble mean pair and the truth, respectively. a is the vector connecting the ensemble mean and the truth. u 1 is the eigenvector of the covariance matrix P that corresponds to the largest eigenvalue λ 1 of P. λ 1 u 1 represents the vector along the major axis of covariance ellipses, where λ 1 represents a scale factor applied to the white data. θ is the angle between a and u 1 . Assuming normal distributions of state ensembles, the major axis length of the covariance ellipse, or the confidence region (CR), is 2 s λ 1 , where s = 1.39, 4.605 or 9.21 for 50, 90, or 99% CRs, respectively.
Figure 5. Illustration of geometric characteristics in a 2-D state space. Pink dots represent 2000 ensemble pairs of states ( x 1 , x 2 ) . The red and green dots denote the ensemble mean pair and the truth, respectively. a is the vector connecting the ensemble mean and the truth. u 1 is the eigenvector of the covariance matrix P that corresponds to the largest eigenvalue λ 1 of P. λ 1 u 1 represents the vector along the major axis of covariance ellipses, where λ 1 represents a scale factor applied to the white data. θ is the angle between a and u 1 . Assuming normal distributions of state ensembles, the major axis length of the covariance ellipse, or the confidence region (CR), is 2 s λ 1 , where s = 1.39, 4.605 or 9.21 for 50, 90, or 99% CRs, respectively.
Hydrology 09 00084 g005
Figure 6. State space plot of x1 and x2 of the Lorenz 63 model from the CBEnKF, the EnKF, the Open Loop and the truth. Note that at the bottom right plot, u 1 is plotted based on EVD results; however, u 1 (not shown) should also be considered to interpret the principal error vector pointing towards the truth—this is reflected in the shape and the size of covariance ellipses at the plot.
Figure 6. State space plot of x1 and x2 of the Lorenz 63 model from the CBEnKF, the EnKF, the Open Loop and the truth. Note that at the bottom right plot, u 1 is plotted based on EVD results; however, u 1 (not shown) should also be considered to interpret the principal error vector pointing towards the truth—this is reflected in the shape and the size of covariance ellipses at the plot.
Hydrology 09 00084 g006
Figure 7. Same as Figure 6 but for x1 and x3.
Figure 7. Same as Figure 6 but for x1 and x3.
Hydrology 09 00084 g007
Figure 8. Same as Figure 6 but for x2 and x3.
Figure 8. Same as Figure 6 but for x2 and x3.
Hydrology 09 00084 g008
Figure 9. Geometric statistics of 2-D state-space from the CBEnKF and the EnKF in the case of ns = 2000, and σ Z 2   = 400 as a function of an exceedance probability.
Figure 9. Geometric statistics of 2-D state-space from the CBEnKF and the EnKF in the case of ns = 2000, and σ Z 2   = 400 as a function of an exceedance probability.
Hydrology 09 00084 g009
Figure 10. Time series of state ensemble mean errors, | | K k | | F , and | | P a , k | | F . At the top three plots, the true state (green line) time series is overlaid for comparison. Inverse Hyperbolic Sine (IHS) transformation is applied to the top three plots to properly view ensemble mean errors, denoted by blue and red lines for the EnKF and the CBEnKF, respectively; x ¯ denotes the ensemble mean of x; the superscript + represents the updated state; the subscript T denotes the truth; | | K k , i C | | F ¯ = 0.38 ; | | K k , i K | | F ¯ = 0.08 ;   1 N i = 1 N | | K k , i C | | F | | K k , i K | | F = 5.1 ;   | | P k , i C | | F ¯ = 92 ; | | P k , i K | | F ¯ = 31 ,   and 1 N i = 1 N | | P k , i C | | F | | P k , i K | | F = 3.1 .
Figure 10. Time series of state ensemble mean errors, | | K k | | F , and | | P a , k | | F . At the top three plots, the true state (green line) time series is overlaid for comparison. Inverse Hyperbolic Sine (IHS) transformation is applied to the top three plots to properly view ensemble mean errors, denoted by blue and red lines for the EnKF and the CBEnKF, respectively; x ¯ denotes the ensemble mean of x; the superscript + represents the updated state; the subscript T denotes the truth; | | K k , i C | | F ¯ = 0.38 ; | | K k , i K | | F ¯ = 0.08 ;   1 N i = 1 N | | K k , i C | | F | | K k , i K | | F = 5.1 ;   | | P k , i C | | F ¯ = 92 ; | | P k , i K | | F ¯ = 31 ,   and 1 N i = 1 N | | P k , i C | | F | | P k , i K | | F = 3.1 .
Hydrology 09 00084 g010
Figure 11. Frobenius norm of K k and P a , k , or | | K k | | F and | | P a , k | | F , respectively, as a function of exceedance probabilities, where ns = 2000, and σ Z 2 = 400 are used for both the CBEnKF (red line) and the EnKF (blue line).
Figure 11. Frobenius norm of K k and P a , k , or | | K k | | F and | | P a , k | | F , respectively, as a function of exceedance probabilities, where ns = 2000, and σ Z 2 = 400 are used for both the CBEnKF (red line) and the EnKF (blue line).
Hydrology 09 00084 g011
Figure 12. Geometric statistics of the 2-D state space from the CBEnKF and the EnKF in the case of ns = 10, 20, 30, 50, 70, 100, 200, 300, 500, 700, 1000, and 2000, and σ Z 2 = 10 and 400. Dots (extreme 1%) denote statistics based on samples with the exceedance probability of 0.01.
Figure 12. Geometric statistics of the 2-D state space from the CBEnKF and the EnKF in the case of ns = 10, 20, 30, 50, 70, 100, 200, 300, 500, 700, 1000, and 2000, and σ Z 2 = 10 and 400. Dots (extreme 1%) denote statistics based on samples with the exceedance probability of 0.01.
Hydrology 09 00084 g012
Figure 13. Kalman gain norm and analysis covariance norm of the CBEnKF and the EnKF in the case of ns = 10, 20, 30, 50, 70, 100, 200, 300, 500, 700, 1000, and 2000, and σ Z 2 = 10 and 400.
Figure 13. Kalman gain norm and analysis covariance norm of the CBEnKF and the EnKF in the case of ns = 10, 20, 30, 50, 70, 100, 200, 300, 500, 700, 1000, and 2000, and σ Z 2 = 10 and 400.
Hydrology 09 00084 g013
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Lee, H.; Shen, H.; Seo, D.-J. Geometric Analysis of Conditional Bias-Informed Kalman Filters. Hydrology 2022, 9, 84. https://doi.org/10.3390/hydrology9050084

AMA Style

Lee H, Shen H, Seo D-J. Geometric Analysis of Conditional Bias-Informed Kalman Filters. Hydrology. 2022; 9(5):84. https://doi.org/10.3390/hydrology9050084

Chicago/Turabian Style

Lee, Haksu, Haojing Shen, and Dong-Jun Seo. 2022. "Geometric Analysis of Conditional Bias-Informed Kalman Filters" Hydrology 9, no. 5: 84. https://doi.org/10.3390/hydrology9050084

APA Style

Lee, H., Shen, H., & Seo, D. -J. (2022). Geometric Analysis of Conditional Bias-Informed Kalman Filters. Hydrology, 9(5), 84. https://doi.org/10.3390/hydrology9050084

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop