Next Article in Journal
Variations in Aerosol Optical Properties over East Asian Dust Storm Source Regions and Their Climatic Factors during 2000–2021
Previous Article in Journal
Spatiotemporal Variation in Air Pollution Characteristics and Influencing Factors in Ulaanbaatar from 2016 to 2019
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Analytical Four-Dimensional Ensemble Variational Data Assimilation for Joint State and Parameter Estimation

1
School of Marine Science and Technology, Tianjin University, Tianjin 300072, China
2
Tianjin Key Laboratory for Oceanic Meteorology, Tianjin 300074, China
*
Authors to whom correspondence should be addressed.
Atmosphere 2022, 13(6), 993; https://doi.org/10.3390/atmos13060993
Submission received: 11 May 2022 / Revised: 16 June 2022 / Accepted: 18 June 2022 / Published: 20 June 2022
(This article belongs to the Section Atmospheric Techniques, Instruments, and Modeling)

Abstract

:
The joint state and parameter estimation problem is an important issue in data assimilation. An adjoint free data assimilation method, namely analytical four-dimensional ensemble variational (A-4DEnVar) data assimilation method, was developed to provide a solution for the joint estimation problem. In the algorithm, to estimate the adjoint model reasonably, the ensemble initial conditions and parameters are generated by Gaussian noise whose covariance is constructed by multiplying a very small factor by their background error covariance. The ensemble perturbations are calculated with respect to background states rather than the ensemble mean. Next, the usage of temporal cross covariances makes it possible to avoid the adjoint model and estimate the gradient in 4DVar. Furthermore, we update the solution iteratively with a linear search process to improve the stability and ensure the convergence of the algorithm. The method is tested using the three-variable Lorenz model (Lorenz-1963) to illustrate its efficiency. It is shown that A-4DEnVar results in similar performance with 4DVar. Sensitivity experiments show that A-4DEnVar is able to assimilate observations successfully with different settings. The proposed method is able to work as well as 4DVar and avoid adjoint models for the joint state and parameter estimation.

1. Introduction

As a chaotic system, the weather prediction models, which suffer from uncertainty in initial conditions and parameters, are very sensitive to its initial conditions [1]. With small errors in initial conditions, even the short-term state variables result in a big gap between forecasting and truth. Besides, the impact of small-scale processes on large-scale processes must be parameterized in practice due to the limitation of computational recourses [2,3]. Though some of the parameters have definite physical meaning in NWP systems and can be deduced from observations empirically or based on physical laws, there are lots of underlying parameters which cannot be easily inferred from observations [4]. For long-term predictions, the parameterizations influence the behavior much more than in short-term forecasting [5].
An effective mean to alleviate the uncertainty of initial conditions and parameters is using the information extracted from realistic observations. Data assimilation algorithms, such as variational methods [6,7] and ensemble Kalman filter methods (EnKF) [8], are proposed to involve the information from observations into state variables. Under similar statistical assumptions that the error is multi-dimensional Gaussian distributed, variational data assimilation schemes and EnKF data assimilation schemes are similar [9]. Constructing covariances to describe the evolution of errors is the key to both variational methods and EnKF methods.
The variational data assimilation methods use a static and flow-independent background error covariance. Various studies developed spatially inhomogeneous and anisotropic background error covariance [10,11,12,13,14,15]. As a non-sequential data assimilation scheme, if not combined with a specific estimation method, like the method proposed by the National Meteorological Centre (NMC) [16], and re-estimated with the covariance at the beginning of a data assimilation cycle, four-dimensional variational (4DVar) data assimilation scheme itself cannot provide an explicit flow-dependent error covariance at the next cycle window. By the great superiority of assimilating all observations spread over a data assimilation window simultaneously and taking the model constraint into account, 4DVar has been widely used in operational systems [17]. However, the cost of this advantage is high. Coding and maintaining the tangent linear model and adjoint model of a specific operational system often costs many human resources but cannot be easily transplanted to other models.
In contrast to variational methods, EnKF uses the flow-dependent covariance explicitly. Improved from Kalman filter theory and extended Kalman filter (EKF) [18], EnKF estimates the error covariance by ensemble members. Once the observation is introduced, the algorithm updates the forecast states straightforwardly. In these loops, the initial conditions are not updated so that the adjoint model is avoided [19]. Hence, EnKF and its varieties [20,21] have been also widely used recently.
Although the variational methods and EnKF achieve similar performances in NWP models, both of them have merits and demerits [22,23]. As many studies mentioned, an attractive issue is to combine both sides to improve performance. A typical development is introducing ensemble-based covariance into the variational framework to make the background error covariance is flow-dependent [24,25,26,27]. When combined with 4DVar, since the integrations of the tangent linear model and adjoint model are also necessary, these kinds of algorithms are named as ensemble four-dimensional variational methods (En4DVar) [28]. The second typical development, so-called four-dimensional ensemble variational method (4DEnVar), goes even further. Introduced by Liu et al. [29], the 4DEnVar algorithm not only introduces ensemble-based covariance but also avoids the use of tangent linear model and the adjoint model. In their studies, experiments based on the one-dimensional shallow water model were conducted to prove its efficiency. Initially, Liu et al. [29] called the proposed 4DEnVar as En4DVar method. With the continuous emergence of hybrid method, Lorenc et al. [28] renamed it as 4DEnVar method at the World Meteorological Organization (WMO) meeting to highlight its essence as a variational method without adjoint models.
Because of the good properties of 4DEnVar mentioned above, this method has attracted much attention and developed rapidly in theory and application. The POD-E4DVar method proposed by Tian et al. [30] uses the POD method (proper orthogonal decomposition) to decompose the set members into orthogonal state variables and represents the analysis field as a linear combination of these orthogonal state variables, which effectively avoids the use of adjoint mode. Although, in the study of Tian et al. [30], POD-E4DVar relies on 4DVar framework, this method is still enlightening and pioneering. Liu et al. [31] proposed an applicable localization scheme to apply this method to operational analysis and prediction. In 2013, Liu and Xiao [32] used 4DEnVar to assimilate data from radiosonde ships to study a cyclone process in the Ross Sea in the South Pacific. Buehner et al. [25] compared the assimilation effects of 3DVar, 4DVar, EnKF, and 4DEnVar. Lorenc et al. [33] combined the flow dependent background error covariance matrix estimated by the ensemble members with the static background error covariance matrix constructed by the climate state in the Met Office operational system, proposed the hybrid 4DEnVar and compared it with En4DVar. Goodliff et al. [34] used Lorenz-63 model to compare the differences of assimilation algorithms including 4DEnVar under different observation frequencies and different assimilation window lengths. Tian et al. [35] and Zhang and Tian [36] proposed a series of nonlinear least squares En4DVar (NLS- En4DVar) and carried out experimental verification based on Lorenz-96 model. Although named as En4DVar, it is in fact a 4DEnVar method that takes the analysis field as the linear combination of orthogonal state variables. Tian et al. [37,38] further constructed ensemble members from historical data and improved the performance of existing algorithms by introducing constraints in the cost function [39] and improving its optimization iterations.
In another paper [40], the 4DEnVar proposed by Liu et al. [29] was improved by our team to capture the model nonlinearity. A new adjoint free data assimilation method called the analytical four-dimensional ensemble variational data assimilation method (A-4DEnVar) was proposed. As mentioned above, the 4DEnVar is motivated by containing the flow-dependent information into background error covariance and also avoids using the adjoint model in data assimilation. However, if the ensemble perturbations are not small enough, the gradients 4DEnVar provided are not the same as that from 4DVar (even they use the same background error covariance and cost function) under strong nonlinear situations. This is mainly because 4DEnVar maintains the idea of EnKF, in which the covariance of the ensemble members is the same as background or analysis error covariance. As the ensemble members are not close enough to the prior estimation, the high order terms in the Taylor expansion influence the estimation accuracy of the adjoint model (see Section 2 below for more details). Thus, the gradients are not well calculated in 4DEnVar. To better obtain the adjoint model and gradients, the high order terms in Taylor expansion should be as small as possible. To this end, A-4DEnVar constructs the covariance of ensemble members by multiplying the background covariance by a small factor (denoted as μ in our formula derivation). Furthermore, to keep consistent with 4DVar, the ensemble perturbations in A-4DEnVar are calculated through centralizing the ensemble members to the evolution of a prior field rather than the mean of these ensemble members as 4DEnVar does. In addition, it is well known that the value of the adjoint model is updated iteratively in 4DVar to capture the nonlinearity. A similar update of the prior initial states in A-4DEnVar is also included to achieve the same effect. We also notice that some variants of 4DEnVar, like NLS-En4DVar [36], also introduced an iterative loop to greatly improved its processing ability of nonlinearity. With these improvements, the A-4DEnVar completely inherits the properties of the conventional 4DVar scheme. Especially, the background error covariance used in A-4DEnVar and 4DVar could be the same. The covariance is not flow-dependent at the beginning of the first data assimilation window. However, if the data assimilation is proceeded in many cycle windows, i.e., the end of one window is the beginning of the next, it is also possible to construct a flow-dependent covariance in A-4DEnVar as 4DEnVar does. This issue will be discussed in our future work.
Many studies focused on optimizing the initial conditions using 4DEnVar algorithms. Following the development of its theory, Liu et al. [31,32] incorporated 4DEnVar into the Advanced Research Weather Research and Forecasting (ARW-WRF) model. Buehner et al. [25] and Lorenc et al. [33] compared these hybrid variational and ensemble methods in global NWP systems. Liu et al. [41] also discussed the relationships of several 4DEnVar algorithms. Liu et al. [42] used a developed 4DEnVar algorithm, i.e., dimension-reduced projection four-dimensional variational data assimilation scheme (DRP-4DVar), to optimize initial conditions in the Lorenz-96 [43] model. In Liang et al. [40], we optimized initial conditions in the Lorenz-63 model. However, as mentioned above, beyond the initial conditions, the state-of-the-art NWP systems consist of uncertainty caused by parameterizations as well. It is widely acknowledged that the conventional data assimilation methods also provide efficient ways to solve the combined state and parameter estimation problem. There are many such studies. To name a few, see [44,45] for 4DVar and see [46,47,48,49,50] for EnKF.
It is illustrated that the A-4DEnVar works as well as conventional 4DVar, even with a small ensemble size and a relatively long data assimilation window with sparse observations. However, the joint–estimation problem has not been well discussed in A-4DEnVar. It is expected to incorporate the simultaneous estimation method under the framework of A-4DEnVar. Therefore, this issue is to be addressed in the paper. In this paper, we developed A-4DEnVar to estimate the initial conditions and parameters simultaneously. At the beginning of the paper, we consider the evaluations of both state variables and parameter perturbations in the dynamic model. The adjoint models are related to the temporal-spatial covariances of these perturbations. After introducing an approximation of the true initial conditions, the analytical solution of the minimizations of cost function is represented using adjoint models. Then, replacing the adjoint models with temporal-spatial covariances, the adjoint models can be removed. Experiments are conducted to show the equivalence of proposed A-4DEnVar with conventional 4DVar in the joint estimation problem.
This paper is organized as follows. Section 2 shows the proposed estimation algorithm and its derivation. Section 3 validates the performance of the A-4DEnVar by the Lorenz model with three parameters. Summary and conclusions are given in Section 4.

2. Methodology

2.1. Relating the Adjoint Models to Second-Order Statistics

Suppose that the dynamic system is defined as
x i = M i , i 1 ( x i 1 , λ , F i 1 )
where M denotes the model operator evolves from time t i 1 to time t i , and x i 1 , λ and F i 1 are the state variables, model parameters, and forcing terms, respectively. To begin the data assimilation, we assume that the prior estimations of the initial condition and parameters are known and denoted as x 0 * and λ * , respectively. Using the same dynamic model in the equations above, under a strong constraint framework, the estimation satisfies
x i * = M i , i 1 ( x i 1 * , λ * , F i 1 )
Suppose any other approximation of x i is composed of the a priori estimation and a perturbation (here, marked with a tilde), i.e.,
x i = x i * + x ˜ i ,   λ = λ * + λ ˜
In contrast to our earlier study, the parameters are also included in the combined estimation problem. Substituting these equations into the dynamical model and expressing through Taylor series expansion gives
x i = M i , i 1 ( x i 1 * + x ˜ i 1 , λ * + λ ˜ , F i 1 ) M i , i 1 ( x i 1 * , λ * , F i 1 ) + M i , i 1 x | x i 1 * , λ * , F i 1 ( x ˜ i 1 ) + M i , i 1 λ | x i 1 * , λ * , F i 1 ( λ ˜ )
where M i , i 1 x | x i 1 * , λ * , F i 1 and M i , i 1 λ | x i 1 * , λ * , F i 1 represent the tangent linear expansion of the dynamic model M i , i 1 with respect to the state variables and parameters, respectively. We do not optimize the forcing terms in this paper, so that the tangent linear model about the forcing terms is omitted without any misunderstanding. Comparing Equation (4) with Equation (2) leads to a one-step evolution of the perturbations based on the tangent linear model, i.e.,
x ˜ i M i , i 1 x | x i 1 * , λ * ( x ˜ i 1 ) + M i , i 1 λ | x i 1 * , λ * ( λ ˜ )
We emphasize here that, although Equation (5) was also mentioned in previous studies like Liu et al. [29], it is often overlooked that the perturbations must be particularly small. The background error covariances are usually too large to be used directly, so without any modification they are not a reasonable choice for x ˜ 0 or λ ˜ . It is easy to generalize the result to the multiple steps evolution corresponding to the initial perturbations
x ˜ i M i , 0 | x 0 * , λ * x ˜ 0 + D i , 0 | x 0 * , λ * λ ˜
where M i , 0 | x 0 * , λ * M i , i 1 x | x i 1 * , λ * M 1 , 0 x | x 0 * , λ * represents the tangent linear model from time step t 0 to time step t i , and
D i , 0 | x 0 * , λ M i , 1 | x 0 * , λ * M 1 , 0 λ | x 0 * , λ * + + M i , i 1 | x 0 * , λ * M i 1 , i 2 λ | x 0 * , λ * + M i , i 1 λ | x 0 * , λ *
demonstrates the compound tangent linear models with respect to state variables and parameters. Note that the above derivations hold for any perturbations that are small enough so that the high order terms in Taylor expansion can be neglected. Different from conventional 4DEnVar and EnKF schemes, the perturbations in A-4DEnVar are used to estimate the adjoint models as shown below.
If there are N perturbations ( x ˜ i k , λ ˜ i k ) ( k = 1 , 2 , , N ) substituting them into Equation (6) yields
( x ˜ i 1 x ˜ i 2 x ˜ i N ) M i , 0 | x 0 * , λ * ( x ˜ 0 1 x ˜ 0 2 x ˜ 0 N ) + D i , 0 | x 0 * , λ * ( λ ˜ 1 λ ˜ 2 λ ˜ N )
Multiplying 1 N ( x ˜ i 1 x ˜ i 2 x ˜ i N ) by both sides of the equation, it yields
1 N k = 1 N ( x ˜ i k x ˜ 0 k T ) M i , 0 | x 0 * , λ * ( 1 N k = 1 N x ˜ 0 k x ˜ 0 k T ) + 1 N k = 1 N ( D i , 0 | x 0 * , λ * λ ˜ k x ˜ 0 k T )
Again, multiplying 1 N ( λ ˜ 1 λ ˜ 2 λ ˜ N ) by both sides of the equation, it yields
1 N k = 1 N ( x ˜ i k λ ˜ k T ) M i , 0 | x 0 * , λ * ( 1 N k = 1 N x ˜ 0 k λ ˜ k T ) + 1 N k = 1 N ( D i , 0 | x 0 * , λ * λ ˜ k λ ˜ k T )
If the ensemble size is large enough, the estimation of the tangent linear model is obtained through
M i , 0 | x 0 * , λ * [ 1 N k = 1 N ( x ˜ i k x ˜ 0 k T D i , 0 | x 0 * , λ * λ ˜ k x ˜ 0 k T ) ] ( 1 N k = 1 N x ˜ 0 k x ˜ 0 k T ) 1
D i , 0 | x 0 * , λ * [ 1 N k = 1 N ( x ˜ i k λ ˜ k T M i , 0 | x 0 * , λ * x ˜ 0 k λ ˜ k T ) ] ( 1 N k = 1 N λ ˜ k λ ˜ k T ) 1
where T is the transpose operator. In the above equations, the distributions of perturbations are not determined or fixed yet. It is easy to see that Equations (11) and (12) holds for any small perturbations and these perturbations do not have to be (though they could be) flow-dependent. In other words, the structure of the adjoint model can be determined by a specially designed Monte Carlo method. A probability density that ignores the interactions between the perturbations of state variable initial conditions and that of parameters is acceptable. With the probability density, the estimations can be simplified as
M i , 0 | x 0 * , λ * ( 1 N k = 1 N x ˜ i k x ˜ 0 k T ) ( 1 N k = 1 N x ˜ 0 k x ˜ 0 k T ) 1
D i , 0 | x 0 * , λ * ( 1 N k = 1 N x ˜ i k λ ˜ k T ) ( 1 N k = 1 N λ ˜ 0 k λ ˜ k T ) 1
The tangent linear model here is related through the temporal cross covariances B i 0 = 1 / N k = 1 N x ˜ i k x ˜ 0 k T and P i = 1 / N k = 1 N x ˜ i k λ ˜ k T , the error covariance of ensemble initial condition B 00 = 1 / N k = 1 N x ˜ 0 k x ˜ 0 k T , and the error covariance of the ensemble parameters Λ = 1 / N k = 1 N λ ˜ k λ ˜ kT for any time step t i , i.e., M i , 0 | x 0 * , λ * B i 0 ( B 00 ) 1 and D i , 0 | x 0 * , λ * P i Λ 1 . Especially, P i describes how the initial conditions are related to parameters. Symbols B 0 i and D 0 i are the transposes of B i 0 and D i 0 , that are B 0 i = ( B i 0 ) T and D 0 i = ( D i 0 ) T .
In the derivations above, the tangent linear models and their adjoint models are connected to the second-order statistics of perturbations. Therefore, it is possible to replace adjoint models with these statistics of ensemble members. Besides, the temporal cross covariances are involved in the analytical solution of the cost function in the next subsection to further simplify the calculation. The equations mentioned above hold for any time step t i , but the temporal covariances are only necessary at the time where observations exist (see more details in the next subsection).

2.2. Expressing Analytical Solutions Using Temporal Cross Covariances

In conventional 4DVar schemes, the cost function is defined as
J ( x 0 , λ ) = 1 2 ( x 0 x b ) T B 1 ( x 0 x b ) + 1 2 i [ i ( x i ) y i ] T R i 1 [ i ( x i ) y i ]
in which x b is the background initial condition, λ denotes the parameters in the dynamic model as mentioned in the above subsection, and x i = M i , i 1 [ M 1 , 0 ( x 0 , F 0 ) , F i 1 ] are the state variables at time t i . Here, i , y i and R i are the observation mapping operator, observations, and the error covariance of observations, respectively. The background regularization term for parameters is not included in the cost function above because it is difficult to be well determined in reality. Hence, parameters are estimated only based on the observational information in this paper.
Note that although the observation mapping operator might be nonlinear, it can be linearized as H i by using similar processes discussed in Section 2.1. Considering Equations (1)–(3) and (6), the cost function concerning perturbations is
J ( x ˜ 0 , λ ˜ ) 1 2 ( x 0 * + x ˜ 0 x b ) T B - 1 ( x 0 * + x ˜ 0 x b ) + 1 2 i ( H i M i , 0 x ˜ 0 + H i D i , 0 λ ˜ d i ) T R i 1 ( H i M i , 0 x ˜ 0 + H i D i , 0 λ ˜ d i )
where d i = y i i x i * is the observational innovation at time step t i . In Equation (16) and henceforth, parts of subscripts of adjoint models are omitted to simplify notations. The gradient of the cost function concerning x ˜ 0 and λ ˜ were then calculated using
λ ˜ J = i D i , 0 T H i T R i 1 ( H i M i , 0 x ˜ 0 + H i D i , 0 λ ˜ d i )
x ˜ 0 J = B 1 [ x ˜ 0 ( x b x 0 * ) ] + i M i , 0 T H i T R i 1 ( H i M i , 0 x ˜ 0 + H i D i , 0 λ ˜ d i )
In conventional 4DVar, numerical optimization algorithms such as gradient descend methods are often used. In contrast to the conventional 4DVar, we solve the equations above directly to achieve its analytical solutions. To further simplify the descriptions, we introduce notations below
S ( A , C ) i A i T H i T R i 1 H i C i S ( A , d ) i A i T H i T R i 1 d i
to denote the summation terms. Here, A i and C i represent the tangent linear model M i , 0 and/or D i , 0 , or the temporal cross covariance B i 0 and/or P i . Thus, the minimum solution of the cost function is (See more details in Appendix A for the derivations)
x ˜ 0 = [ B 1 + S ( M , M ) S ( M , D ) S ( D , D ) 1 S ( M , D ) T ] 1 [ B 1 ( x b x 0 * ) + S ( M , d ) S ( M , D ) S ( D , D ) 1 S ( D , d ) ]
λ ˜ = [ S ( D , D ) S ( M , D ) T ( B - 1 + S ( M , M ) ) 1 S ( M , D ) ] 1 [ S ( M , D ) T ( B - 1 + S ( M , M ) ) 1 B 1 ( x b x 0 * ) S ( M , D ) T ( B - 1 + S ( M , M ) ) 1 S ( M , d ) + S ( D , d ) ]
Or eliminating the adjoint models, the equations are equivalent to
x ˜ 0 = B 00 [ B 00 B 1 B 00 + S ( B , B ) S ( B , P ) S ( P , P ) - 1 S ( B , P ) T ] 1 [ B 00 B 1 ( x b x 0 * ) + S ( B , d ) S ( B , P ) S ( P , P ) - 1 S ( P , d ) ]
λ ˜ = Λ [ S ( P , P ) S ( B , P ) T ( B 00 B 1 B 00 + S ( B , B ) ) 1 S ( B , P ) ] 1 [ S ( B , P ) T ( B 00 B 1 B 00 + S ( B , B ) ) 1 B 00 B 1 ( x b x 0 * ) S ( B , P ) T ( B 00 B 1 B 00 + S ( B , B ) ) 1 S ( B , d ) + S ( P , d ) ]
To ensure that the generated perturbations are small enough, we generate ensemble members using B 00 = μ B where μ is a small factor to ensure the high order terms in Taylor expansion could be omitted. In practice, with cycled data assimilation, the perturbations could not always satisfy small perturbations in all windows, especially with nonlinear regimes. Once the perturbations are too large, the ensembles should be regenerated. Theoretically, the value of factor μ is related to the spectral norms of the Hessian matrix of dynamic models and should be carefully determined. However, it is difficult to calculate the spectral norms. An efficient way is set the factor to be as small as possible, and empirically, setting μ to be less than 10−6 is often acceptable. Here, the specially designed construction of B 00 also makes it possible to introduce a flow-dependent covariance estimation method (see next subsections for details). Assuming that the parameters to be estimated are mutually independent, Λ = diag ( σ λ 1 2 , , σ λ m 2 ) is a diagonal matrix with its non-zero elements being the ensemble error variances of the parameters and m is the number of parameters. Then, Equations (22) and (23) leads to
x ˜ 0 = B 00 [ μ B 00 + S ( B , B ) S ( B , P ) S ( P , P ) - 1 S ( B , P ) T ] 1 [ μ ( x b x 0 * ) + S ( B , d ) S ( B , P ) S ( P , P ) - 1 S ( P , d ) ]
λ ˜ = Λ [ S ( P , P ) S ( B , P ) T ( μ B 00 + S ( B , B ) ) 1 S ( B , P ) ] 1 [ μ S ( B , P ) T ( μ B 00 + S ( B , B ) ) 1 ( x b x 0 * ) S ( B , P ) T ( μ B 00 + S ( B , B ) ) 1 S ( B , d ) + S ( P , d ) ]
Equations (24) and (25) shows that the information introduced into the increments are composed of three parts, i.e., the difference between the background initial conditions with the prior approximation ( x b x 0 * ) , observational innovations transmitted through the state variables S ( B , d ) and through parameters S ( P , d ) . The solution of the cost function is analytical for fixed approximations x * and λ * . We call the proposed method analytical 4DEnVar to be consistent with Liang et al. [40]. However, the form of solution is much more complex than what we obtained in the earlier study because the interactions of initial conditions and parameters were included.
It is interesting to note that the increments of state variables and parameters are similar except for the term μ B 00 which is related to the background term in the cost function. In fact, the parameters are regarded as special time-invariant state variables. This treatment process is similar to previous studies of Koyama and Watanabe [51]. In their study, the so-called state augmentation method regards time-varying parameters as state variables. The estimation processes of state variables and parameters are separated from each other. State variables are assimilated in several cycle windows by conventional EnKF firstly, then the parameters are estimated using the imperfect analysis state variables. In contrast to it, the estimation of state variables and parameters proceeded at the same time under the framework of A-4DEnVar in our study.
When considering the single estimation problem for initial conditions or parameters, it is a natural step to assume the parameters (or initial conditions) are accurate. The corresponding perturbation increment reduces to
x ˜ 0 = B 00 [ μ B 00 + S ( B , B ) ] 1 [ μ ( x b x 0 * ) + S ( B , d ) ]
λ ˜ = Λ [ S ( P , P ) ] 1 S ( P , d )
The algorithm proposed is in fact equivalent to the Newton iteration method used in conventional 4DVar [40]. As is well known that the Newton iteration is sensitive to the initial conditions, to make the algorithm stable, a linear search process is necessary. When updating analysis state variables, we add the weighted perturbation ( α 1 x ˜ , α 2 λ ˜ ) to ( x * , λ * ) , and α 1 and α 2 [ 0 , 1 ] are the weights minimizing the cost function under the increment direction ( x ˜ 0 , λ ˜ ) . After the linear search, ( x * , λ * ) is updated and then substituted into the next iteration to regenerate ensemble members and calculate a new increment. The process is finished when the difference of the value of the cost function between two iterations is less than tolerance.
While in real atmospheric and oceanic models, the dimension of state variables is huge, the terms with inverse are difficult to calculate. To overcome the obstruction, A-4DEnVar methods estimate the corresponding pseudo inverses instead of the inverse matrix itself by using a few ensemble members. Since the ensemble size is not very large, if we denote the ensemble perturbations as X ˜ = ( X ˜ 0 1   X ˜ 0 2     X ˜ 0 N ) where N is the ensemble size, its pseudo-inverse is X ˜ 0 + = ( X ˜ 0 T X ˜ 0 ) 1 X ˜ 0 T . Note that X ˜ 0 T X ˜ 0 is an N × N matrix, then X ˜ 0 + could be easily calculated. Besides, any inverse term in Equations (24) and (25) and/or Equations (26) and (27) can be decomposed as ( X 0 + ) T S 1 X 0 + , where S is a symmetric matrix with N rows and N columns. Based on the decomposition, one can easily calculate the value of x ˜ and λ ˜ . We emphasize that a compatible localization approach is also necessary for this situation (see more details in next subsections). However, as a proof of concept, under the simplified models whose degree of freedom is small, like the Lorenz model we used below, the inverses are directly used in this paper.
The flowchart of A-4DEnVar method is shown below (Figure 1).

2.3. Flow-Dependent Covariance and Localization Method

The proposed A-4DEnVar is not a traditional hybrid scheme. The advantage of A-4DEnVar is that the adjoint model is avoided. Although the ensembles in A-4DEnVar are mainly used to estimate the adjoint model, the uncertainty of state variables and parameters are also propagated through these ensembles. It is also possible to propose a flow-dependent scheme in A-4DEnVar.
If an initial condition x 0 * is the center of ensemble initial condition, i.e.,
x 0 n = x 0 * + ε 0 n
where x 0 n denotes the initial condition of ensemble members and ε 0 n is a zero-mean random noise. In 4DEnVar, x 0 * denotes the true initial condition and is unknown. However, in A-4DEnVar, x 0 * denotes the background or the prior initial condition. It is known exactly and is updated in each iteration.
It is easy to calculate E ( x 0 n ) = x 0 * and Var ( x 0 n ) = Var ( ε 0 n ) . Here Ε ( ) and Var ( ) are the expectation and variance operator, respectively. After integration through a dynamic model, say M ( ) , for example, one would obtain the ensemble state variables at time t i ,
x i n = x i * + ε i n = M ( x 0 n ) = M ( x 0 * + ε 0 n )
Hence, the uncertainty of the state variables is
Var ( ε i n ) = Var ( M ( x 0 * + ε 0 n ) M ( x 0 * ) ) Var ( M ( ε 0 n ) )
Here, M is the tangent linear model. However, because the true state is unknown yet, 4DEnVar replaces M ( x 0 * ) with Ε [ M ( x 0 n ) ] , which is only acceptable under a linear or weakly nonlinear situation. In A-4DEnVar, the uncertainty (which is related to the initial state x 0 * ) is calculated directly without any approximation. In addition, a factor is multiplied bythe perturbation in A-4DEnVar, but its influence can be easily removed by dividing the perturbations by the square of the same factor when estimating the variance, i.e.,
Var ( ε i n ) Var ( M ( ε 0 n ) ) 1 μ Var ( M ( x 0 * + μ ε 0 n ) M ( x 0 * ) )
To better consider the flow-dependent error covariance, a combination of traditional 4DEnVar and A-4DEnVar is also useful. However, such issues are beyond this paper. A more detailed discussion combined with localization methods will be addressed in our future work.
It is well known that the dimensions of state variables in real operational system are often huge. The huge dimension usually introduces unexpected problems. One of them is how to localize the background error covariance so that the long-distance pseudo correlations between state variables introduced from a finite of ensemble members are eliminated. Although this paper focuses on the derivation of the algorithm, here we also present a convenient improvement of A-4DEnVar to implement localization. Motivated by Liu et al. (2009), the localization of background error covariance is proposed through extended ensemble members. The process is briefly described below.
In A-4DEnVar, the set of ensemble perturbations is an N -column matrix X ˜ 0 ,
X ˜ 0 = ( x ˜ 0 1 , x ˜ 0 2 , , x ˜ 0 N )
where N is the ensemble size, and the i th ensemble perturbation x ˜ 0 n ( n = 1 , , N ) is a column vector. The background error covariance without localization is
B 1 μ N X ˜ 0 X ˜ 0 T
Generally, to localize B one often modifies it by a correlation function, such as the Schur operator for example. The localized background error covariance is defined as
B L = ρ   ·   B
and ρ denotes the spatial correlation function. Suppose ρ is decomposed as
ρ 1 r ρ ˜ ρ ˜ T
in which ρ ˜ is a matrix composed of r empirical orthogonal functions of ρ . It is demonstrated in Liu et al. [28] that the localized error covariance is
B L 1 μ r N Z ˜ 0 Z ˜ 0 T
where
Z ˜ 0 = ( ρ ˜ X ˜ 0 1 ρ ˜ X ˜ 0 n ρ ˜ X ˜ 0 N )
and X ˜ 0 n ( n = 1 , , N ) is an r -column matrix with its every column x ˜ 0 n .
It is easy to see that the construction of localized background error covariance B L is similar with B itself but with more ensemble members. If the extended ensemble size is also not very large, the solution of A-4DEnVar algorithm (Equations (24) and (25)) is not changed, even with localization in practice.

3. Experiments and Results

3.1. Twin Experiment Setup

Before being performed in operational systems, it is important to study the method in some simple but widely accepted models to test the method. Since this is the first step of our study, a highly nonlinear chaotic ideal model is used here. More complicated models will be studied in the future. In this subsection, we follow the study of Miller et al. [52], Kalnay et al. [22], and Goodliff et al. [34], and use the Lorenz [1] model (Lorenz-63, henceforth) for verification of A-4DEnVar data assimilation. Although the Lorenz-63 model consists of only three variables and three parameters, it is a strong nonlinear model. The six degrees of freedom make it possible to carry out experiments with full rank. The governing equations of the Lorenz-63 model are
d x d t = σ ( y x ) d y d t = r x y x z d z d t = x y b z
where σ is the Prandtl number, b a non-dimensional wave number, and r a normalized Rayleigh number. The standard and default values of parameters are set to be ( σ , b , r ) = ( 10 , 8 / 3 , 28 ) which were widely used in previous studies. The model runs for truth solutions using a fourth-order Runge-Kutta numerical scheme with time step d t = 0.01 and starting from the same initial condition as in Goodliff et al. [34], i.e., (−3.12346395; −3.12529803; 20.69823159), which were obtained after disregarding the first 1000 steps of an integrating process.
For the experiments constructed in this paper, all three model state variables are observed. The simulated observations are sampled from the truth model state variables contaminated by Gaussian noise with its mean zero and variance 1.0 at regular periods. The background initial conditions were obtained by adding Gaussian noise with covariance B to the initial truth. Considering the interactions between variables, the background error covariance is provided by a three-dimensional variational data assimilation (3DVar) process with a 5000 time-step cycle. The cycle includes many end-to-end assimilation windows and only one observation at the end of each window. Starting from an arbitrary symmetric B c and initial conditions, the 3DVar provides an analysis state in a data assimilation window. Then, integrating the analysis state, one would obtain a forecast state x f in the next window. The truth states x t r u t h in the next window can also be calculated easily through integrating the truth initial conditions. Repeat the assimilation and forecast processes until the end of the cycle. A new B c is estimated as
B c = ( x f x t r u t h ) ( x f x t r u t h ) T ¯
from t = 500 to t = 5000 . Again, repeat the estimations. The B c usually converges to B in less than 10 iterations.
The determination of parameters is even more complicated than that of initial conditions. There are two empirical strategies to get the parameters. The first demonstrates that no prior information about the parameters is known so that the variance of three parameters might be the same for convenient. The second multiplies a factor by the value of the parameters as their variances. It is hard to point out which is better. However, the main purpose of this paper is to propose the A-4DEnVar for joint estimation problems and show its consistency with conventional 4DVar. Therefore, the two strategies are both acceptable here. In this paper, the initial parameters are calculated by adding random noise with variance 0.25 to the default parameters. Besides, we also conducted experiments based on the variance considered the difference of the parameters, i.e., the variances are set to be ( 0.1 , 0.027 , 0.28 ) , which is one percent of their values. Obviously, the variances are much less or slightly larger than what is used below. Experiment results also demonstrate that A-4DEnVar performs as well as conventional 4DVar. However, we do not show these results in this paper for the limitation of spaces.
To estimate the adjoint models through temporal cross covariances, we add Gaussian noise with covariance μ B (in all experiments, we set μ = 10 8 ) to the initial conditions and another white noise with variance 10 8 to the background parameters to generate ensemble members. In this paper, these noises are regenerated in every iteration to be consistent with the derivations. To reduce computational cost, it is not always necessary to update them in practice if ensembles are close enough to ( x * , λ * ) . In future works, more efficient generation methods will be discussed, but it is beyond the scope of this paper.
The widely used root-mean-square error ( R M S E ) as a criterion to appraise the performance of A-4DEnVar and conventional 4DVar schemes. Here, the RMSE for the variable x is defined as
R M S E ( x ) = 1 L t = 1 L ( x t a n a l y s i s x t t r u t h ) 2
where L represents the number of time steps contained in all time windows (or cycles) in an experiment, x t a n a l y s i s is the analysis state variable and x t t r u t h is the true state variable at time step t . The R M S E ¯ mentioned below is calculated from all variables,
R M S E ¯ = 1 K k = 1 K R M S E [ x ( k ) ]
where K is the degree of freedom and it is three for both state variables and parameters in the Lorenz-63 model.

3.2. Comparison with Conventional 4DVar

In this subsection, we compare the performance of A-4DEnVar with conventional 4DVar. The experiment settings for window length, observational period, ensemble size, etc., are as follows.
A long data assimilation window may introduce multiple minimums for the cost function. For the single initial condition estimation problem, Kalnay et al. [22] recommended that the optimal window length when observed every 8 time steps is about 32 time steps, and if observed every 25 steps the optimal window length is from 50 to 125 steps. However, as it is known that the simultaneous estimation for initial conditions and parameters is much more complex than the single estimation problems, it is reasonable to use a data assimilation window that is shorter but can also touch the ultimate window length of 4DVar. Thus, we choose 72 steps for a compromise. Besides, we set the observational period to be 12 time steps in the preliminary test. We conducted experiments with an observational period of 24. However, due to the nonlinearity of the chaotic system, it is very hard for both 4DVar and A-4DEnVar to achieve stable performance. For some initial conditions or parameters, algorithms converge to local minimizers that are far away from the truth, which indicates the failure of data assimilation. Therefore, we do not show these results in this paper. The experiment is designed as end-to-end cycles with 14,400 time steps, i.e., 200 windows. One must note that it may be possible to produce special optimal methods to extend the length of assimilation windows and observation periods.
Here, the ensemble size is set to be 50 to validate the performance of A-4DEnVar in full rank. To test the performance of the algorithm, 100 experiments were produced using both A-4DEnVar and conventional 4DVar. In these experiments, the initial conditions for state variables, parameters, and observations are set to be the same for both A-4DEnVar and 4DVar (but they are different values in the 100 experiments). A model control run solution (labeled as CTL) without any data assimilation or correction is also presented.
The results (Figure 2a shows the solutions only in the first 25 windows from one experiment) indicate the effectiveness of A-4DEnVar. Figure 2b shows the mean RMSE of the conventional 4DVar algorithm and A-4DEnVar algorithm in all 100 experiments. Although the RMSE of A-4DEnVar is slightly less than that of 4DVar, both algorithms assimilate the information from observations so that the analysis RMSE for each variable is less than the standard deviations of observations, the unit of which is mentioned above. It is shown that both algorithms perform well under the experiment setting. In Figure 2c, the RMSEs of parameters b and r from A-4DEnVar are less than that from 4DVar. However, 4DVar performs slightly better than A-4DEnVar on the parameter σ .
Due to the complex manifold structure caused by the high nonlinear model, the global minimum of the cost function is difficult to achieve through gradient based optimal methods. The chaos of the Lorenz model makes the optimization even more difficult so that a small perturbation in the gradient would lead to another local minimum of the cost function. Considering that the motivation of A-4DEnVar is replacing the adjoint models with these estimated temporal cross error covariances, these estimations can be close enough to their true values but cannot be the same. The gradient directions calculated from adjoint models in 4DVar are slightly different from what was estimated in A-4DEnVar. Therefore, the RMSE obtained by the two data assimilation schemes might be slightly different from each other. In our conclusions, we do not emphasize whether the performance of 4DVar or that of A-4DEnVar is better, but only state that their performances are all acceptable and are equivalent to each other.
The computational cost also influences the usage of both methods. The computing burden for A-4DEnVar is mainly from integrating ensembles. For conventional 4DVar, the burden is mainly from the calculation of adjoint models. Empirically, running the adjoint model often costs about 6 to 10 times the computational resources of the forward dynamic model. Therefore, if every ensemble member is updated in each iteration, it would be cheaper to implement A-4DEnVar when the ensemble size is less than 10. However, it also should be noticed that regenerating or updating ensembles is only necessary if ensembles are not close to the background. For example, updating the ensembles after several iterations of increments is acceptable when the nonlinearity is limited. Therefore, it is possible for A-4DEnVar to further reduce its computing costs in practice. What also should be emphasized is that, to calculate the gradient, one has to store the value of state variables increments in every time step in 4DVar. In contrast to it, A-4DEnVar only stores the values of state variables where the observations are located. Considering the dimension of the state variables is often much larger than that of observations, A-4DEnVar saves most storage spaces. Besides, the proposed A-4DEnVar is an ensemble-based method, and it can be easily implemented in operational parallel systems to save time.
The usage of factor μ to control the value of perturbations and generate ensemble members is an important difference between A-4DEnVar and 4DEnVar methods. Figure 3 shows the difference between the gradients (in the first iteration step) calculated by A-4DEnVar and the traditional 4DVar method with the change of μ . It is shown that, in Lorenz-63 model, only if the factor is less than 10 6 which is far less than the value of the background error covariance used in other 4DEnVar methods, the gradients (for both variables and parameters) obtained from A-4DEnVar are close enough to that from traditional 4DVar.

3.3. Model Sensitivities with Respect to Observation Period, the Length of Data Assimilation Windows and Ensemble Size

Since the nonlinearity of operational NWP systems increases with the length of the assimilation window, it is of great importance to point out the assimilation capability of A-4DEnVar with different window lengths and observation periods. In this set of experiments, the window of length 24, 48, 72, and 120 steps with observation periods 4, 6, and 12 steps are conducted. Each experiment is carried out 500 times to test the statistical performance and robustness. Again, the initial conditions for state variables, parameters, and observations are set to be the same for A-4DEnVar and 4DVar, but they are different values in these experiments.
The RMSEs of state variables as a function of window lengths and observation periods are presented by box plot in Figure 4. On each box, the central mark indicates the median, and the bottom and top edges of the box indicate the 25th and 75th percentiles, respectively. The whiskers extend to the most extreme data points not considered outliers, and the outliers are plotted individually using the red cross-marker symbol. It can be seen that for experiment settings, most RMSEs of variable states are less than 1.0. The outliers are mainly because of the random choice of backgrounds, and the algorithms find a local minimum. For a fixed window length, the mean of assimilation RMSEs increases with the observational period. The best performance occurs when the window is about 120 time steps. If the window is too short, the lack of observational information also results in large RMSEs. When the assimilation window is 24 steps and the observation period is 24 steps (not shown here), the RMSE is greater than the standard deviation of the observational noise, which indicates the failure of data assimilation.
For parameters in Figure 5, a shorter observational period does not achieve a smaller RMSE. In this case, we find that the minimum point of the cost function is different from the truth because of the interaction between state variables and parameters and the noises contained in observations. Since both A-4DEnVar and 4DVar converge to this point, parameters with high RMSEs are obtained. However, the trajectories are almost the same as the truth. Another suit of experiments (not shown here) which only optimize parameters with perfect initial conditions shows that the interactions can be alleviated but cannot be completely eliminated. Both in Figure 4 and Figure 5, the experiments carried out by 4DVar using the same settings also lead to similar results and demonstrate the equivalence of A-4DEnVar and 4DVar.
In the last set of experiments, we consider the impact of ensemble size on the performance of A-4DEnVar. Generally speaking, the dimensions of variable states in the operational NWP system are huge. It is necessary to evaluate the required ensemble size for data assimilation. However, we only consider a simple model with full rank ensembles here. Because the Lorenz-63 chaos system has six degrees of freedom, to implement data assimilation, the ensemble sizes are chosen as 10, 20, 50, 100, 300, 500, and 1000, while the window length and observational period are set to be 72 and 12 steps as in the preliminary experiment. Each experiment is also repeated 500 times to demonstrate the statistical performance of the A-4DEnVar under the same settings used above. The R M S E ¯ results for variables and parameters are shown in Figure 6.
Although the RMSEs for each variable change a little with different ensemble size, it is clear that all RMSEs are less than observation variance, which shows the effectiveness of the proposed algorithm. The A-4DEnVar is not sensitive to the ensemble sizes for the Lorenz model with full rank ensembles. In fact, the key to A-4DEnVar is estimating the adjoint models by ensemble perturbations. The performance of A-4DEnVar depends on the precision of these estimations. Theoretically, it is the magnitude of ensemble perturbations rather than the ensemble sizes influence the estimation of adjoint models. The experiment results shown in Figure 6 also indicate that RMSE does not improve significantly with the increase of ensemble sizes. While the dimension of the NWP system is large, it is often impossible to generate enough ensemble members. Hence, the localization methods should be introduced to alleviate spurious correlations between state variables. However, we do not focus on the issue in this paper and our related studies are also on the way.

4. Conclusions

The necessity of an adjoint model in conventional 4DVar schemes limits its transplant to other operational NWP systems. To alleviate the limitation, the hybrid data assimilation scheme, namely A-4DEnVar, has been proposed for the initial condition optimal problem in our earlier work. In this paper, the A-4DEnVar is proposed for the combined state and parameter estimation. In A-4DEnVar, the ensemble members are centralized to the state variables integrated with background parameters and from the background initial condition at each iteration step to estimate reasonable temporal cross covariances. The perturbation is very small and model error evolution is described by the temporal cross covariances so that the adjoint model is avoided. The linear research and iteration process are also adopted to ensure convergence.
The A-4DEnVar algorithm is evaluated by the Lorenz-63 model in which all three variables and parameters are estimated simultaneously. Comparisons between A-4DEnVar and adjoint-based conventional 4DVar show that the A-4DEnVar can not only capture the observational information but also obtain a reasonable analysis state in the assimilation window, just as 4DVar does. Sensitive experiments demonstrate the effectiveness of A-4DEnVar in different assimilation window lengths and observational periods. Besides, the proposed algorithm is not sensitive to ensemble size if in full rank.
In this paper, we validate A-4DEnVar in a simplified model. This is the first step of our study, and it will soon be studied in operational systems. Our further work focuses on the localization and the generation of ensemble members to improve its performance in real applications.

Author Contributions

Conceptualization, K.L. and W.L.; methodology, K.L.; validation, G.H.; writing—original draft preparation, K.L.; writing—review and editing, W.L. and G.H.; visualization, S.L. and Y.G. project administration, W.L.; funding acquisition, G.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by grants from the National Key Research and Development Program (2021YFC3101501) and the National Natural Science Foundation (41876014) of China.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Acknowledgments

The authors acknowledge valuable discussions with Hanyu Liu, Lige Cao and Qi Shao from the School of Marine Science and Technology of Tianjin University.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

In this Appendix we present the derivartions of the analytical solutions of A-4DEnVar. Considering the gradient of cost function and set x ˜ 0 J = 0 and λ ˜ J = 0 in Equations (17) and (18), i.e.,
B 1 [ x ˜ 0 ( x b x 0 * ) ] + i M i , 0 T H i T R i 1 ( H i M i , 0 x ˜ 0 + H i D i , 0 λ ˜ d i ) = 0
i D i , 0 T H i T R i 1 ( H i M i , 0 x ˜ 0 + H i D i , 0 λ ˜ d i ) = 0
From Equation (A1), we find that
( B 1 + i M i , 0 T H i T R i 1 H i M i , 0 ) x ˜ 0 = [ B 1 ( x b x 0 * ) + i M i , 0 T H i T R i 1 ( d i H i D i , 0 λ ˜ ) ]
Note that Equation (A2) indicates that
λ ˜ = ( i D i , 0 T H i T R i 1 H i D i , 0 ) 1 i D i , 0 T H i T R i 1 ( H i M i , 0 x ˜ 0 d i )
Substituting Equation (A4) into Equation (A3) yields
[ B 1 + S ( M , M ) S ( M , D ) S ( D , D ) 1 S ( M , D ) T ] x ˜ 0 = [ B 1 ( x b x 0 * ) + S ( M , d ) S ( M , D ) S ( D , D ) 1 S ( D , d ) ]
where
S ( A , C ) i A i T H i T R i 1 H i C i S ( A , d ) i A i T H i T R i 1 d i
It is easy to get from Equation (A5) that
x ˜ 0 = [ B 1 + S ( M , M ) S ( M , D ) S ( D , D ) 1 S ( M , D ) T ] 1 [ B 1 ( x b x 0 * ) + S ( M , d ) S ( M , D ) S ( D , D ) 1 S ( D , d ) ]
Substituting M i , 0 | x 0 * , λ * B i 0 ( B 00 ) 1 and D i , 0 | x 0 * , λ * P i Λ 1 into the above equation, we have
x ˜ 0 = B 00 [ B 00 B 1 B 00 + S ( B , B ) S ( B , P ) S ( P , P ) - 1 S ( B , P ) T ] 1 [ B 00 B 1 ( x b x 0 * ) + S ( B , d ) S ( B , P ) S ( P , P ) - 1 S ( P , d ) ]
Eliminating x ˜ 0 from Equations (A1) and (A2) with similar process, we have
λ ˜ = [ S ( D , D ) S ( M , D ) T ( B - 1 + S ( M , M ) ) 1 S ( M , D ) ] 1 × [ S ( M , D ) T ( B - 1 + S ( M , M ) ) 1 B 1 ( x b x 0 * ) S ( M , D ) T ( B - 1 + S ( M , M ) ) 1 S ( M , d ) + S ( D , d ) ]
and
λ ˜ = Λ [ S ( P , P ) S ( B , P ) T ( B 00 B 1 B 00 + S ( B , B ) ) 1 S ( B , P ) ] 1 × [ S ( B , P ) T ( B 00 B 1 B 00 + S ( B , B ) ) 1 B 00 B 1 ( x b x 0 * ) S ( B , P ) T ( B 00 B 1 B 00 + S ( B , B ) ) 1 S ( B , d ) + S ( P , d ) ]
which are equations in our main sections.

References

  1. Lorenz, E. Deterministic Nonperiodic Flow. J. Atmos. Sci. 1963, 20, 130–141. [Google Scholar] [CrossRef]
  2. Khain, A.P.; Beheng, K.D.; Heymsfield, A.; Korolev, A.; Krichak, S.O.; Levin, Z.; Pinsky, M.; Phillips, V.; Prabhakaran, T.; Teller, A.; et al. Representation of microphysical processes in cloud-resolving models: Spectral (bin) microphysics versus bulk parameterization. Rev. Geophys. 2015, 53, 247–322. [Google Scholar] [CrossRef]
  3. Liang, C.R.; Shang, X.D.; Qi, Y.F.; Chen, G.Y.; Li, Z.J.; Lu, L.H. A Modified Finescale Parameterization for Turbulent Mixing in the Western Equatorial Pacific. J. Phys. Oceanogr. 2021, 51, 1133–1143. [Google Scholar] [CrossRef]
  4. Li, Q.; Zheng, Q.; Zhou, S.-Z. Study on the dependence of the two-dimensional Ikeda model on the parameter. Atmos. Ocean Sci. Lett. 2016, 9, 1128694. [Google Scholar] [CrossRef]
  5. Annan, J.D.; Hargreaves, J.C. Efficient parameter estimation for a highly chaotic system. Tellus Ser. A Dyn. Meteorol. Oceanogr. 2004, 56, 520–526. [Google Scholar] [CrossRef]
  6. Lewis, J.M.; Derber, J.C. The use of adjoint equations to solve a variational adjustment problem with advective constraints. Tellus A 1985, 37A, 309–322. [Google Scholar] [CrossRef]
  7. Courtier, P.; Thépaut, J.-N.; Hollingsworth, A. A strategy for operational implementation of 4D-Var, using an incremental approach. Q. J. R. Meteorol. Soc. 1994, 120, 1367–1387. [Google Scholar] [CrossRef]
  8. Evensen, G. Sequential Data Assimilation with a Nonlinear Quasi-Geostrophic Model Using Monte-Carlo Methods to Forecast Error Statistics. J. Geophys. Res. 1994, 99, 10143–10162. [Google Scholar] [CrossRef]
  9. Li, Z.J.; Navon, I.M. Optimality of variational data assimilation and its relationship with the Kalman filter and smoother. Q. J. R. Meteorol. Soc. 2001, 127, 661–683. [Google Scholar] [CrossRef]
  10. Wu, W.S.; Purser, R.J.; Parrish, D.F. Three-dimensional variational analysis with spatially inhomogeneous covariances. Mon. Weather. Rev. 2002, 130, 2905–2916. [Google Scholar] [CrossRef]
  11. Purser, R.J.; Wu, W.S.; Parrish, D.F.; Roberts, N. Numerical aspects of the application of recursive filters to variational statistical analysis. Part I: Spatially homogeneous and isotropic Gaussian covariances. Mon. Weather. Rev. 2003, 131, 1524–1535. [Google Scholar] [CrossRef]
  12. Bucanek, A.; Brozkova, R. Background error covariances for a BlendVar assimilation system. Tellus Ser. A Dyn. Meteorol. Oceanogr. 2017, 69, 13. [Google Scholar] [CrossRef]
  13. Michel, Y.; Menetrier, B.; Montmerle, T. Objective filtering of the local correlation tensor. Q. J. R. Meteorol. Soc. 2016, 142, 2314–2323. [Google Scholar] [CrossRef]
  14. Weaver, A.T.; Chrust, M.; Menetrier, B.; Piacentini, A. An evaluation of methods for normalizing diffusion-based covariance operators in variational data assimilation. Q. J. R. Meteorol. Soc. 2021, 147, 289–320. [Google Scholar] [CrossRef]
  15. Bannister, R.N. A review of forecast error covariance statistics in atmospheric variational data assimilation. I: Characteristics and measurements of forecast error covariances. Q. J. R. Meteorol. Soc. 2008, 134, 1951–1970. [Google Scholar] [CrossRef]
  16. Stanesic, A.; Horvath, K.; Keresturi, E. Comparison of NMC and Ensemble-Based Climatological Background-Error Covariances in an Operational Limited-Area Data Assimilation System. Atmosphere 2019, 10, 570. [Google Scholar] [CrossRef]
  17. Mahfouf, J.F.; Rabier, F. The ECMWF operational implementation of four-dimensional variational assimilation. II: Experimental results with improved physics. Q. J. R. Meteorol. Soc. 2000, 126, 1171–1190. [Google Scholar] [CrossRef]
  18. Evensen, G. Using the extended Kalman filter with a multilayer quasi-geostrophic ocean model. J. Geophys. Res. 1992, 97, 17905–17924. [Google Scholar] [CrossRef]
  19. Li, H.; Kalnay, E.; Miyoshi, T. Simultaneous estimation of covariance inflation and observation errors within an ensemble Kalman filter. Q. J. R. Meteorol. Soc. 2009, 135, 523–533. [Google Scholar] [CrossRef]
  20. Evensen, G.; van Leeuwen, P.J. An ensemble Kalman smoother for nonlinear dynamics. Mon. Weather. Rev. 2000, 128, 1852–1867. [Google Scholar] [CrossRef]
  21. Bishop, C.H.; Etherton, B.J.; Majumdar, S.J. Adaptive sampling with the ensemble transform Kalman filter. Part I: Theoretical aspects. Mon. Weather. Rev. 2001, 129, 420–436. [Google Scholar] [CrossRef]
  22. Kalnay, E.; Li, H.; Miyoshi, T.; Yang, S.-C.; Ballabrera-Poy, J. 4-D-Var or ensemble Kalman filter? Tellus Ser. A Dyn. Meteorol. Oceanogr. 2007, 59, 758–773. [Google Scholar] [CrossRef]
  23. Gustafsson, N. Discussion on ‘4D-Var or EnKF?’. Tellus Ser. A Dyn. Meteorol. Oceanogr. 2007, 59, 774–777. [Google Scholar] [CrossRef]
  24. Hamill, T.; Snyder, C. A Hybrid Ensemble Kalman Filter / 3D-Variational Analysis Scheme. Mon. Weather. Rev. 2000, 128, 2905. [Google Scholar] [CrossRef]
  25. Buehner, M.; Houtekamer, P.L.; Charette, C.; Mitchell, H.L.; He, B. Intercomparison of Variational Data Assimilation and the Ensemble Kalman Filter for Global Deterministic NWP. Part I: Description and Single-Observation Experiments. Mon. Weather. Rev. 2010, 138, 1550–1566. [Google Scholar] [CrossRef]
  26. Bannister, R.N. A review of operational methods of variational and ensemble-variational data assimilation. Q. J. R. Meteorol. Soc. 2017, 143, 607–633. [Google Scholar] [CrossRef]
  27. Clayton, A.M.; Lorenc, A.C.; Barker, D.M. Operational implementation of a hybrid ensemble/4D-Var global data assimilation system at the Met Office. Q. J. R. Meteorol. Soc. 2013, 139, 1445–1461. [Google Scholar] [CrossRef]
  28. Lorenc, A.C. Recommended nomenclature for EnVar data assimilation methods. Res. Act. Atmos. Ocean. Model. WGNE 2013, 5, 2. [Google Scholar]
  29. Liu, C.; Xiao, Q.; Wang, B. An ensemble-based four-dimensional variational data assimilation scheme. Part I: Technical formulation and preliminary test. Mon. Weather. Rev. 2008, 136, 3363–3373. [Google Scholar] [CrossRef]
  30. Tian, X.; Xie, Z.; Dai, A. An ensemble-based explicit four-dimensional variational assimilation method. J. Geophys. Res. 2008, 113, D21. [Google Scholar] [CrossRef]
  31. Liu, C.; Xiao, Q.; Wang, B. An Ensemble-Based Four-Dimensional Variational Data Assimilation Scheme. Part II: Observing System Simulation Experiments with Advanced Research WRF (ARW). Mon. Weather. Rev. 2009, 137, 1687–1704. [Google Scholar] [CrossRef]
  32. Liu, C.; Xiao, Q. An Ensemble-Based Four-Dimensional Variational Data Assimilation Scheme. Part III: Antarctic Applications with Advanced Research WRF Using Real Data. Mon. Weather. Rev. 2013, 141, 2721–2739. [Google Scholar] [CrossRef]
  33. Lorenc, A.C.; Bowler, N.E.; Clayton, A.M.; Pring, S.R.; Fairbairn, D. Comparison of Hybrid-4DEnVar and Hybrid-4DVar Data Assimilation Methods for Global NWP. Mon. Weather. Rev. 2015, 143, 212–229. [Google Scholar] [CrossRef]
  34. Goodliff, M.; Amezcua, J.; Van Leeuwen, P.J. Comparing hybrid data assimilation methods on the Lorenz 1963 model with increasing non-linearity. Tellus A Dyn. Meteorol. Oceanogr. 2015, 67, 6928. [Google Scholar] [CrossRef]
  35. Tian, X.; Zhang, H.; Feng, X.; Xie, Y. Nonlinear least squares En4DVar to 4DEnVar methods for data assimilation: Formulation, analysis, and preliminary evaluation. Mon. Weather. Rev. 2018, 146, 77–93. [Google Scholar] [CrossRef]
  36. Zhang, H.; Tian, X. A Multigrid Nonlinear least squares four-dimensional variational data assimilation scheme with the advanced research weather research and forecasting model. J. Geophys. Res. Atmos. 2018, 123, 5116–5129. [Google Scholar] [CrossRef]
  37. Tian, X.; Zhang, H. A big data-driven nonlinear least squares four-dimensional variational data assimilation method: Theoretical formulation and conceptual evaluation. Earth Space Sci. 2019, 6, 1771–1774. [Google Scholar] [CrossRef]
  38. Tian, X.; Zhang, H. Implementation of a modified adaptive covariance inflation scheme for the big data-driven NLS-4DVar algorithm. Earth Space Sci. 2019, 6, 2593–2604. [Google Scholar] [CrossRef]
  39. Tian, X.; Han, R.; Zhang, H. An adjoint-free alternating direction method for four-dimensional variational data assimilation with multiple parameter Tikhonov regularization. Earth Space Sci. 2020, 7, e2020EA001307. [Google Scholar] [CrossRef]
  40. Liang, K.; Li, W.; Han, G.; Shao, Q.; Zhang, X.; Zhang, L.; Jia, B.; Bai, L.; Liu, S.; Gong, Y. An analytical four-dimensional ensemble-variational data assimilation scheme. J. Adv. Modeling Earth Syst. 2021, 13, 2818. [Google Scholar] [CrossRef]
  41. Liu, C.S.; Xue, M. Relationships among four-dimensional hybrid ensemble-variational data assimilation algorithms with full and approximate ensemble covariance localization. Mon. Weather. Rev. 2016, 144, 591–606. [Google Scholar] [CrossRef]
  42. Liu, J.; Wang, B.; Xiao, Q. An evaluation study of the DRP-4-DVar approach with the Lorenz-96 model. Tellus Ser. A Dyn. Meteorol. Oceanogr. 2011, 63, 256–262. [Google Scholar] [CrossRef]
  43. Lorenz, E. Predictability: A problem partly solved. In Proceedings of the Seminar on Predictability, Reading, UK, 4–8 September 1995; European Centre for Medium-Range Weather Forecasts (ECMWF): Reading, UK, 1996; Volume 1, pp. 1–19. [Google Scholar]
  44. Ullman, D.S.; Wilson, R.E. Model parameter estimation from data assimilation modeling: Temporal and spatial variability of the bottom drag coefficient. J. Geophys. Res. Ocean 1998, 103, 5531–5549. [Google Scholar] [CrossRef]
  45. Smith, P.J.; Thornhill, G.D.; Dance, S.L.; Lawless, A.S.; Mason, D.C.; Nichols, N.K. Data assimilation for state and parameter estimation: Application to morphodynamic modelling. Q. J. R. Meteorol. Soc. 2013, 139, 314–327. [Google Scholar] [CrossRef]
  46. Aksoy, A.; Zhang, F.; Nielsen-Gammon, J.W. Ensemble-based simultaneous state and parameter estimation in a two-dimensional sea-breeze model. Mon. Weather. Rev. 2006, 134, 2951–2970. [Google Scholar] [CrossRef]
  47. Hu, X.M.; Zhang, F.; Nielsen-Gammon, J.W. Ensemble-based simultaneous state and parameter estimation for treatment of mesoscale model error: A real-data study. Geophys. Res. Lett. 2010, 37, 3017. [Google Scholar] [CrossRef]
  48. Ponsar, S.; Luyten, P.; Ozer, J. Combined model state and parameter estimation with an ensemble Kalman filter in a North Sea station 1-D numerical model. Ocean. Dyn. 2011, 61, 1869–1886. [Google Scholar] [CrossRef]
  49. Ruckstuhl, Y.; Janjić, T. Parameter and state estimation with ensemble Kalman filter based algorithms for convective scale applications. Q. J. R. Meteorol. Soc. 2018, 144, 3257. [Google Scholar] [CrossRef]
  50. Ruckstuhl, Y.; Janjić, T. Combined State-Parameter Estimation with the LETKF for Convective-Scale Weather Forecasting. Mon. Weather. Rev. 2020, 148, 1607–1628. [Google Scholar] [CrossRef]
  51. Koyama, H.; Watanabe, M. Reducing forecast errors due to model imperfections using ensemble kalman filtering. Mon. Weather. Rev. 2010, 138, 3316–3332. [Google Scholar] [CrossRef]
  52. Miller, R.N.; Ghil, M.; Gauthiez, F. Advanced Data Assimilation in Strongly Nonlinear Dynamical Systems. J. Atmos. Sci. 1994, 51, 1037–1056. [Google Scholar] [CrossRef]
Figure 1. The flowchart of proposed A-4DEnVar. The ensemble initial conditions and parameters are generated by adding small perturbations to the background initial conditions and parameters, respectively. After integrating the dynamic model, the ensemble temporal-spatial covariances are calculated to replace the adjoint models in 4DVar. Using the analytical solutions of cost function, state variable and parameter increments are obtained. The linear search processes are conducted to minimize the cost function.
Figure 1. The flowchart of proposed A-4DEnVar. The ensemble initial conditions and parameters are generated by adding small perturbations to the background initial conditions and parameters, respectively. After integrating the dynamic model, the ensemble temporal-spatial covariances are calculated to replace the adjoint models in 4DVar. Using the analytical solutions of cost function, state variable and parameter increments are obtained. The linear search processes are conducted to minimize the cost function.
Atmosphere 13 00993 g001
Figure 2. (a) The solutions of the truth, model control run, data assimilation results from A-4DEnVar, and conventional 4DVar schemes in the first 25 windows. (b) The RMSE of state variables for 200 windows and (c) parameters for A-4DEnVar and 4DVar algorithm for 200 windows, respectively.
Figure 2. (a) The solutions of the truth, model control run, data assimilation results from A-4DEnVar, and conventional 4DVar schemes in the first 25 windows. (b) The RMSE of state variables for 200 windows and (c) parameters for A-4DEnVar and 4DVar algorithm for 200 windows, respectively.
Atmosphere 13 00993 g002
Figure 3. The difference between the gradients obtained from A-4DEnVar and the traditional 4DVar method with the change of factor.
Figure 3. The difference between the gradients obtained from A-4DEnVar and the traditional 4DVar method with the change of factor.
Atmosphere 13 00993 g003
Figure 4. The effect of assimilation window lengths and observation periods on the R M S E ¯ for state variables. (ac) the box plot for A-4DEnVar and (df) for 4DVar. The outliers are marked as red +.
Figure 4. The effect of assimilation window lengths and observation periods on the R M S E ¯ for state variables. (ac) the box plot for A-4DEnVar and (df) for 4DVar. The outliers are marked as red +.
Atmosphere 13 00993 g004
Figure 5. The effect of assimilation window lengths and observation periods on the R M S E ¯ for parameters. (ac) the box plot for A-4DEnVar and (df) for 4DVar. The symbols of box plot are the same as those in Figure 4. The outliers are marked as red +.
Figure 5. The effect of assimilation window lengths and observation periods on the R M S E ¯ for parameters. (ac) the box plot for A-4DEnVar and (df) for 4DVar. The symbols of box plot are the same as those in Figure 4. The outliers are marked as red +.
Atmosphere 13 00993 g005
Figure 6. The effect of assimilation ensemble size on R M S E ¯ for (a) variables and (b) parameters. Note that the window length and observational period are fixed and are 72 and 12 time steps, respectively. The symbols of box plot are the same as those in Figure 4. The outliers are marked as red +.
Figure 6. The effect of assimilation ensemble size on R M S E ¯ for (a) variables and (b) parameters. Note that the window length and observational period are fixed and are 72 and 12 time steps, respectively. The symbols of box plot are the same as those in Figure 4. The outliers are marked as red +.
Atmosphere 13 00993 g006
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Liang, K.; Li, W.; Han, G.; Gong, Y.; Liu, S. Analytical Four-Dimensional Ensemble Variational Data Assimilation for Joint State and Parameter Estimation. Atmosphere 2022, 13, 993. https://doi.org/10.3390/atmos13060993

AMA Style

Liang K, Li W, Han G, Gong Y, Liu S. Analytical Four-Dimensional Ensemble Variational Data Assimilation for Joint State and Parameter Estimation. Atmosphere. 2022; 13(6):993. https://doi.org/10.3390/atmos13060993

Chicago/Turabian Style

Liang, Kangzhuang, Wei Li, Guijun Han, Yantian Gong, and Siyuan Liu. 2022. "Analytical Four-Dimensional Ensemble Variational Data Assimilation for Joint State and Parameter Estimation" Atmosphere 13, no. 6: 993. https://doi.org/10.3390/atmos13060993

APA Style

Liang, K., Li, W., Han, G., Gong, Y., & Liu, S. (2022). Analytical Four-Dimensional Ensemble Variational Data Assimilation for Joint State and Parameter Estimation. Atmosphere, 13(6), 993. https://doi.org/10.3390/atmos13060993

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop