3.1. Error Representation of Filter Equations
Since the error covariance of stochastic variables plays the key role in geometric analyses in the Euclidean space, this section defines errors in states and a residual in measurement in order to rewrite state updating and observation equations with error terms.
In the above, , , and denote the CBPKF analysis error, the KF analysis error, and the forecast error, respectively; represents innovation; , , , and represent the truth, the CBPKF-updated state, the KF-updated state, and the state forecast, respectively.
State update equations for the KF and the CBPKF are rewritten into Equations (27) and (28), assuming the same a priori states used in both filters.
where
and
are Kalman gains from the KF and the CBPKF, respectively.
Updated states from both filters have the following mathematical relationship.
The observation model in Equation (3) can be rewritten as follows:
From Equation (3), the covariance matrix of
, or
, can be written as
where
is the covariance matrix of the state forecast.
3.2. KF and CBPKF Solutions for a Bi-State Model
A bi-state model (
) is used to illustrate the geometric relation of errors in states and observations from the CBPKF and the KF. Observation
exists for a state
but not for
to investigate how state updating can be illustrated in the two-dimensional (2-D) Euclidean space for observable as well as unobservable states, i.e.,
is used for simplicity. Kronhamn [
19] used the same
and
matrices as those described in this Section for the geometric illustration of the KF, whereas we focus on the CBPKF in reference to the KF. For the bi-state model with a forecast error covariance
and the observation error
, matrices of Kalman gain (Equations (32) and (33)), analysis error covariance (Equations (34) and (35)), and correlation (Equations (36)–(38)) for the CBPKF and the KF are evaluated in the equations below where
=
with
representing the correlation between
and
.
where
Equation (32) indicates that for an observable state
,
is always larger than
but for an unobservable state
, the sign of
depends on the sign of
. The CBPKF covariance matrix
and its relation to the KF-equivalent
are given in Equation (34).
where
In Equation (34), variances of both updated states
and
are larger than KF-equivalents, and the sign of covariance terms of
depend on the sign of
. Equations (34) and (35) imply that the CBPKF-updated state ensembles have larger spreads in the state space than the KF-updated, and that the matrix norm of
is larger than that of
. From covariance matrices, Pearson product-moment correlation matrix
can be computed by Equation (36).
where
diag( ) denotes a diagonal matrix. Equation (37) shows that the correlation coefficient of the CBPKF-updated states
can be either larger or smaller than the KF-equivalent
.
where
From Equations (32)–(38), if , then , , and .
3.3. Geometric Representation of KF and CBPKF Solutions
This Section begins with describing a relation between stochastic variables and vectors in the Euclidean domain at Equations (39)–(41), and then describes the geometric representation of KF and CBPKF equations. The covariance of stochastic variables
and
can be used to compute the scalar product of two vectors
and
in the Euclidean domain (Equation (39)). The vector norm
corresponds to the standard deviation of
(Equation (40)). The angle of
and
can be computed from the correlation of
and
(Equation (41)) [
19,
22].
Figure 1 shows the vector representation of stochastic variables in the Euclidean space.
Figure 2 shows the error vectors from (a) the KF and (b) the CBPKF for the observable state
of the bi-state model, where
,
, and
where the last equality is from Equation (30). In
Figure 2a, the state forecast error vector
is orthogonal to the observation error vector
owing to the independence assumption. Being a minimum variance solution,
is orthogonal to
[
19,
23].
Figure 2a also shows that the forecast error is a vector sum of the gain-weighted innovation and the analysis error, i.e.,
as expected from Equation (27). The KF analysis error variance may be obtained in
Figure 2a via the Pythagorean theorem as
where
as expected from Equation (35). Since
=
, we may write via the Pythagorean theorem
which is expected from Equation (34). In
Figure 2, the inequality,
, arises due to the fact that the CBPKF solution minimizes not the error variance but a weighted sum of the error variance and the variance of the Type-II CB.
Figure 3 is the same as
Figure 2 but for the unobservable state
where
;
;
; and
. In
Figure 3a, the proportionality,
, gives
as expected from Equation (33). In
Figure 3a, the orthogonality,
, and the equality
, yield
in agreement with Equation (35). In
Figure 3b,
may be written via the Pythagorean theorem as:
Using Equations (35), (32) and (31) for
,
, and
, respectively, we may rewrite Equation (42), in agreement with Equation (34), as:
Figure 2 and
Figure 3 show that the KF- and the CBPKF-updated state error vectors point to different directions in the state space.
Figure 4 shows the updated state error vectors in
Figure 2 and
Figure 3 to visually compare the differences in the angle, the magnitude, and the direction. The angles of the two-state error vectors for the CBPKF (
in Equation (44)) and the KF (
in Equation (45)) can be computed from the correlation
and
in Equations (37) and (38), respectively:
Below, we develop a set of geometric expressions in the 2-D state space for the analysis error covariance via eigenvalue decomposition (EVD, [
24]).
3.4. Geometric Analysis in the State Space
Geometric characteristics of state ensembles in the 2-D state space can be quantified by confidence regions (CR), eigenvectors, eigenvalues, and the angle between the eigenvector and the basis vector of the x-axis. Assuming normal distributions for state ensembles in a 2-D state space, a CR, or so-called a covariance ellipse, can be constructed based on the EVD of a covariance matrix as well as the Chi-Square probability table. The presence of the CB results in different variances (eigenvalues) and directions (eigenvectors) of updated state ensembles in a 2-D state space. The major and minor axis lengths of the covariance ellipse are 2 and 2 where and the value of is from the Chi-Square probability table for a given confidence region, e.g., = 4.605 for a 90% confidence region given the Chi-Square probability in the case of degrees of freedom of 2. In a 2-D state space, the error in the orientation of the covariance ellipse with respect to the truth can be estimated by the angle, θ, between the largest eigenvector and the vector connecting the truth and the ensemble mean, i.e., .
The EVD of the CBPKF analysis covariance
may be written as:
where
In the above,
U is the eigenvector matrix which rotates the white data (W), or uncorrelated standard normal variates by
. The eigenvalue matrix E explains the variance along the principal error direction, or the direction of the eigenvector. In a 2-D state space,
is a scale factor applied to W. The dataset (D) resulting from scaling W by
and rotating by
, i.e.,
, has the covariance matrix of
. Below we apply EVD to the CBPKF analysis error covariance,
, from the bi-state model in
Section 3.2.
With
, and
,
in Equation (34) may be rewritten as:
Appendix A describes how eigenvalues and eigenvectors may be evaluated for a
covariance matrix. Using Equation (A11) in
Appendix A, we have for
:
The difference in the largest eigenvalue between the KF and the CBPKF analysis error covariance is given by:
where
denotes the largest eigenvalue of
. Using Equation (A12), we may write
and
:
Similarly, using Equation (A14), we may write
and
as:
where the vector
is the basis vector of the x-axis and
is equal to
. With the geometric attributes established above, we now carry out the comparative geometric analysis of the KF and the CBPKF analysis results using the Lorenz 63 model [
25].
With the EVD of
and the ensemble mean
, the minimum percentage confidence
to contain the verifying truth
within the confidence region can be computed by
where
is the Chi-Square probability with degrees of freedom of 2;
satisfies Equation (55).
Figure 5 shows an example of computing
,
,
, and eigenvalues and eigenvectors of a covariance matrix.
3.5. Numerical Experiment with the Lorenz 63 Model
In the sections above, a linear two-state model was used for theoretical simplicity. In this section, we use the three-state Lorenz 63 model to illustrate the differences between the EnKF and the CBEnKF solutions in terms of the geometric attributes introduced above. In this experiment, synthetically generated observations of all three states in the Lorenz 63 model were assimilated at every time step using the EnKF and the CBEnKF [
26]. Preliminary experiments suggested observation error variances (
) of 10 or 400 can be used for the cases of assimilating less uncertain or largely uncertain observations, respectively, based on the ensemble spread. To render the assimilation problem more challenging,
is used to compare the performance of the CBEnKF to that of the EnKF in
Figure 6,
Figure 7,
Figure 8,
Figure 9,
Figure 10 and
Figure 11, where the ensemble size (
) used is 2000 to minimize filter performance degradation owing to a small ensemble size.
Figure 12 and
Figure 13 compare assimilation results from the two filters in the case of
or 400 as a function of an ensemble size, where
values used include 10, 20, 30, 50, 70, 100, 200, 300, 500, 700, 1000, and 2000.
Figure 6 presents the results for states
and
of the Lorenz 63 model in a 2-D state space. The attributes shown include
,
,
,
,
,
, and
;
and
reflect errors in the ensemble mean measured along horizontal or vertical directions, respectively; large dots overlaid in the scatter plots represent mean values of the samples in each of the ten bins equally dividing the entire state space.
Figure 7 and
Figure 8 are the same as
Figure 6 but for the state spaces of
and
, respectively. The following summarizes general observations from
Figure 6,
Figure 7 and
Figure 8. The spread of
and
in the scatter plots shows that the CBEnKF reduces CBs more effectively than the EnKF or the OL, particularly in the extremes. In some cases, however, their mean values appear similar, e.g.,
in the state spaces of
or
. The OL results for
and
show that their patterns appearing in the state space are similar to those of the state space plot in the top left of
Figure 6,
Figure 7 and
Figure 8; this indicates the notable dependency of the amount of ensemble mean errors on model dynamics, e.g., the larger ensemble mean errors at the extreme—this is also seen in EnKF solutions but less so in the CBEnKF. This may be explained by the CBEnKF with a larger weight to observations than the EnKF in the case of largely uncertain observations (
), which reduces the reliance of CBEnKF solutions on the model dynamics. Based on
, CBEnKF covariances are generally larger than the EnKF at all state spaces. Larger
and smaller
and
of the CBEnKF than the EnKF yield consistently smaller
than the EnKF at both extremes and the median of all three variables. This signifies the benefit of using the CB-informed KF for the estimation of extremes given that the EnKF’s
quickly increases towards extremes, i.e., the EnKF is less confident in estimating extremes than the CBEnKF. For example, 3% confidence regions for selected extreme values presented in the bottom right plots show the truth (green dots) contained within the CBEnKF’s confidence regions (red ellipses) but not within the EnKF’s (blue ellipses). At the plots, arrows represent
.
and
do not clearly indicate differences between the two filters.
Figure 9 shows
,
,
,
, and
of
Figure 6,
Figure 7 and
Figure 8 but as a function of exceedance probabilities to highlight CB-informed KF performances at extremes. At extremes with low exceedance probabilities, differences between the CBEnKF and the EnKF are vivid in the case of
and
. On the other hand,
of the CBEnKF is consistently larger than those of the EnKF and the OL across exceedance probabilities. As exceedance probabilities increase, the EnKF’s
becomes similar to the CBEnKF’s, implying unconditionally less biased. The CBEnKF keeps consistently low
at all exceedance probabilities owing to small
and large
, compared to the EnKF or the OL. Since the EnKF seeks orthogonal solutions to minimize analysis covariances, its
is always smaller than the OL’s as well as the CBEnKF’s. On the other hand, the CBEnKF increases
to address CBs which helps keep
low to contain the truth. Both
and
show no consistent patterns across different state spaces as well as exceedance probabilities.
In
Figure 10, the ensemble mean error time series indicates that among the three variables, CBEnKF’s improvement is the largest for
. On the other hand, for the state
, the CBEnKF mainly remedies the underestimation of
compared to the EnKF. In the case of
, the CBEnKF slightly outperforms the EnKF. These observations may imply the different amounts of CBs present in different states, hence the need of applying a separate weight
to the CB penalty for the individual state, which warrants a future effort. To compare
and
from the two filters, the time series of Frobenius norm of
and
is computed by Equations (56) and (57), respectively. Compared to the EnKF, the CBEnKF yields
and
consistently larger at all assimilation cycles, and the mean values of
and
are five and three times larger, respectively.
Figure 11 shows mean
, and
as a function of exceedance probabilities. At extremes, both the CBEnKF and the EnKF show that mean
and
are larger than those at high exceedance probabilities, and that large differences in mean
and
between the CBEnKF and the EnKF are consistent across exceedance probabilities.
Figure 6,
Figure 7,
Figure 8,
Figure 9,
Figure 10 and
Figure 11 are based on the case of uncertain observations (
) where the CBEnKF may supposedly outperform the EnKF. To explore the CBEnKF performance with less uncertain observations (
) and also to see the sensitivity to the ensemble size (
),
Figure 12 presents results from the combination of
= 10, 20, 30, 50, 70, 100, 200, 300, 500, 700, 1000, and 2000, and
and 400. In
Figure 12,
plots indicate that with
, the accuracy of the ensemble mean continuously increases with an increase of
at both cases of extremes (an exceedance probability of 0.1; red and blue dots for the CBEnKF and the EnKF, respectively) and all data (red and blue lines for the CBEnKF and the EnKF, respectively). When
, the EnKF’s
is slightly smaller than the CBEnKF’s, but the CBEnKF’s
is slightly larger than the EnKF’s. The resulting
from both filters are very similar. This implies when observations are less uncertain, the EnKF solutions are as accurate and as confident as the CBEnKF solutions at extremes as well as the whole range. When
200 and
, mean
maintains ~1%. When
200 and
,
quickly increases with a decrease of
because of inaccurate error covariance estimates with an insufficient ensemble size. When observations are largely uncertain (
, the CBEnKF clearly shows more accurate ensemble means (smaller
) and higher confidence in covariance estimates (smaller
) than the EnKF, particularly at extremes. Compared to
, assimilating largely uncertain observations (
) reduces accuracies in covariance estimates, resulting in larger
in both filters, although the CBEnKF’s
addressing the CB is larger than the EnKF’s. When
,
and
tend to be less sensitive to
than the case of
. Both
and
show neither any consistent patterns nor sensitivities to
, but are included in
Figure 12 for completeness.
Finally,
Figure 13 presents mean
and
as a function of
. Compared to the results from
,
results in larger
in both filters due to bigger weights to the observations. When
, the CBEnKF maintains relatively large
to account for the CB; however, the EnKF’s
is conspicuously small. Both
and
tend to be little sensitive to the ensemble size
, except the all data case of the CBEnKF with
(pink line). With uncertain observations (
), the CBEnKF’s
becomes large at extremes (pink dots) as well as all data (pink line) at all
values used to reflect CBs in all states.