1. Introduction
With the development of the airborne remote sensing system, its higher imaging resolution and three-dimensional imaging is increasingly arousing more interest in new research fields, such as interferometric SAR (InSAR) and array antenna SAR imaging. In both of these new fields, multiantennas are positioned at different locations [
1]. Correspondingly, the motion compensation equipment is developed from a single POS to multiple in different locations with the assistance of ADPOS [
2,
3,
4,
5]. In terms of the ADPOS system architecture, an integrated system with a high-precision inertial measurement unit (IMU) and GNSS [
6] serves as the master POS, which combines strapdown inertial solutions with state correction to fuse the inertial and satellite sensor information using a linear model. Several low-cost slave systems rely on fusing the master system’s information with their own data to improve parameter accuracy using various state estimation algorithms. The complete system provides multi-node, high-frequency, and high-accuracy motion parameters to multi-node loads for higher-resolution three-dimensional imaging [
7].
In order to compensate for motion errors in multinode loads, ADPOS has to address the deterministic errors by calibration and stochastic errors using recursive state estimation. When the inertial sensors for ADPOS are determined, stochastic errors are more important than determined errors in performance improvement. Therefore, filter estimation algorithms play a vital role in the transfer alignment of DPOS, which has a direct impact on the recursive state estimation performance. The transfer model of the ADPOS is inherently nonlinear. However, earlier methods such as the widely-used KF are subject to the divergence that results from approximations during any linearization process [
8,
9]. As a result, the EKF becomes the more widely-applied estimator for the nonlinear model, using a first-order linearization method based on a suboptimal recursive estimation framework. However, EKF has some fatal shortcomings, which dramatically degrade the accuracy and cause divergence of the state estimation when the nonlinear characteristic is very strong [
10].
In order to overcome the shortcomings of the aforementioned traditional state estimation algorithm, a series of nonlinear filtering algorithms have been deduced to further develop nonlinear state estimation in the past years [
9]. In addition to these algorithms, several nonlinear state estimation methods have been generated from deterministic sampling methods or numerical methods. And the linear minimum variance criterion is applied to the filtering update process [
11], as occurs in the widely-used unsecnted Kalman filter [
12] and the central difference Kalman filter [
13]. In light of this, the algorithm flow is obtained using the linear minimum variance criterion approximately. However, this is not the actual case, and therefore more efficient filtering update methods that integrate the Kalman filter framework need to be considered. Nonlinear state estimation algorithms based on deterministic sampling using a new probabilistic framework have been developed, such as the unscented Kalman filter (UKF) and the central difference Kalman filter (CDKF). Both of these methods rely on approximating the probability density distribution of the nonlinear function [
14]. Considering these advantages, such state estimation algorithms avoid the need for Jacobian matrix calculations and truncation errors, making them well-suited for handling nonlinear estimation problems [
1]. Thus, higher estimation accuracy is achieved. Also, compared with the UKF algorithm, the CDKF algorithm requires fewer parameters to generate sigma points for nonlinear estimation while maintaining the same level of accuracy, making it one of the most widely used methods in nonlinear estimation [
15]. However, since the transfer alignment model is high-dimensional, CDKF and modified CDKF may not meet the required estimation accuracy. Addressing how to enhance CDKF performance for high-dimensional models is crucial.
For the filtering update process, one can treat it as an unconstrained optimization problem to be solved. Among these fields, the steepest descent algorithm (gradient descent algorithm), the Newton method and the related Quasi-Newton method, and the conjugate gradient algorithm are widely used [
16,
17]. The conjugate gradient algorithm lies between the steepest descent algorithm and Newton method. It only employs the first order derivative information, overcoming the defect of slow convergence of steepest descent algorithm. In addition, it is immune to the calculation and storage of the Hessian matrix. Because of its simple structure, smaller storage capacity, and high numerical efficiency, the conjugate gradient method has been one of the most efficient methods to solve the nonlinear optimization problem and is used in many fields. Developed from Hestenes and Stiefel in the 1950s, conjugate gradient methods have been studied by many researchers [
18]. There are many derivatives developed from these basic methods, such as the Hestenes–Steifel (HS) method [
19], the Fletcher–Reeves (FR) method [
20], the Polak–Ribiére-Polyak (PRP) method [
21], the conjugate descent (CD) method [
22], the Liu–Storey (LS) method [
23], and the Dai–Yuan (DY) method [
24]. These methods differ from others in their approach to selecting step length factors. Specifically, the FR, CD, and DY methods perform well in convergence performance, but poorly in numerical performance. While the PRP, LS, and HS methods can self-correct when encountering small step sizes continuously, showing better numerical performance, we focus on the PRP method for conducting the conjugate gradient to achieve optimization.
Compared with general filtering estimation methods, the related derivatives using the linear minimum variance criterion assume that the state estimation is a linear regression with the observation [
25]. Variational optimization is an effective method to replace the Kalman linear regression with quadratic terms in the observation to handle nonlinearities [
26,
27]. It combines the conjugate gradient method with variational optimization to estimate state mean, which can achieve higher state estimation accuracy. On the other hand, for error covariance matrix estimation, the original update by general nonlinear filtering is used to estimate the state estimation covariance matrix, which takes place in parallel with the state update. Motivated by the above description, based on the superiority of variational optimization with the conjugate gradient method, CDKF adopts fewer parameters to achieve estimation, and a central difference-based variational filtering algorithm (CDVF) is proposed to contribute to the transform alignment process of the master and slave systems of ADPOS. The main contributions of this paper can be summarized as the following aspects.
(1) For the filtering prediction stage, the CDVF algorithm uses the CDKF interpolation method to substitute derivative computation to propagate the system characteristic. (2) For the filtering update stage, the CDVF algorithm uses variational optimization with conjugate gradient method to conduct the state update instead of the linear minimum variance estimation. Meanwhile, the state estimation covariance matrix update progresses in parallel with the state update by the original CDKF.It can obviously improve the accuracy of the slave system motion parameters. (3) A real flight test is conducted to verify the proposed algorithm. Detailed experiment results are presented and show the superiority of the proposed algorithm.
The outline of the remainder of this work is arranged as follows: In
Section 2, the background theories, including the linear minimum variance criterion, the central difference Kalman filtering algorithm, and the conjugate gradient optimization method, are listed successively. Compared with the linear minimum variance criterion based on CDKF algorithm, combining the central difference interpolation method and the variational optimization method with the conjugate gradient method, the central difference variational filtering algorithm is proposed in
Section 3. The detailed experiment validation process, including experiment design, data processing, and analysis of the results, is located in
Section 4. Finally, the conclusion and future work are displayed in
Section 5.
2. Related Theories
In this section, the related theories, including the linear minimum variance criterion for nonlinear systems and developed nonlinear central difference Kalman filtering and the conjugate gradient method for variational optimization, will be displayed successively to serve for the next section.
The nonlinear system with additive noise can be represented as
where
,
denotes the state vector at discrete-time
k,
denotes the measurement vector at time
k,
is the nonlinear state transition function,
is the process noise coefficient matrix,
is the measurement matrix,
,
are mutually independent noise vectors, and
,
.
2.1. Linear Minimum Variance (LMV) Criterion for Nonlinear Systems [25]
Let
be the variable to be estimated where
is the observation of
. If the following equality holds for
,
then
is the linear minimum variance estimation of
on
. This can also be denoted as
. Also,
and
can be calculated as
Here, the linear minimum variance estimation of
based on
is formulated as
and the estimation error covariance matrix is
For nonlinear systems such as (
1), let
. The linear minimum variance estimation of
based on
can be displayed as
According to the linearity of linear minimum variance estimation, it has
Approximate
as
,
as
, then combine (7). One can check that
and
Considering that
and
is symmetric, then formula (4) is realized as
By the above deduction, the linear minimum variance estimation for nonlinear systems is formulated, but only the approximate results are achieved because of the nonlinearity of the system. Thus, the different nonlinear filtering or smoothing algorithms are devoted to using different numerical calculation methods to derive , and .
2.2. Central Difference Kalman Filtering
For system (
1), under the assumption of Gaussian distribution, applying the linear minimum variance criterion and the Gaussian weighted integrals to the nonlinear filtering equation, the Gaussian filtering is obtained [
11]. Especially, Stirling’s interpolation approximation generates the central difference Kalman filtering and is used to solve the numerical integrals to complete the recursive estimation [
13].
By the above description, the central difference Kalman filtering algorithm in each step for system (
1) is listed as
At time , the Gaussian posterior density function is , where and .
Determine the set of sigma points
according to
where
,
,
,
.
Propagate the set of sigma points in accordance with the system equation as
and the predictive probability density function (pdf)
is approximated by the Gaussian pdf
, where
where
.
Determine the set of predictive sigma points
at time instant
k according to
and
.
Update the state estimation and compute the posterior pdf with the measurement :
Considering that the measurement equation of (
1) is linear, then the covariance matrix update of central difference Kalman filtering can be simplified as
By the above deduction, one can simplify the filtering covariance matrix update. Then it can progress in parallel with the state update using the one-step prediction , measurement matrix , and measurement noise covariance .
2.3. Conjugate Gradient Method for Unconstrained Optimization Problems
For the unconstrained optimization problems, different algorithms are available, such as the steepest descent method, Newton and quasi-Newton methods, conjugate gradient method. Compared with the linear minimum variance criterion, which uses the linear method to complete the filtering update, the conjugate gradient method can be used to improve the filtering update process of the filtering algorithms.
The conjugate gradient method is very efficient for solving large-scale unconstrained optimization problems. A detailed description is presented as follows. For the unconstrained optimization problem (matrix
is positive definite),
then
denotes the gradient of
,
. Here, we will show the algorithm flow of the conjugate gradient method (Algorithm 1). The most important point is to decide the step length factor
, popular algorithms such as FR, PRP, HS, DY, and so on. The difference lies in the selection of
. Here, we list the algorithm flow of the PRP conjugate gradient method as following [
21].
Algorithm 1: Conjugate gradient method (PRP) |
Initialization: , , |
1: |
2: |
3: |
4: |
5: |
3. Central Difference Variational Filtering Algorithm
On the basis of the above subsections, the central difference variational filtering algorithm is developed, where the central difference Kalman filtering algorithm is used to conduct the state prediction, covariance matrix prediction and update. Meanwhile, the conjugate gradient method is used to update the state instead of the LMV criterion. Combining the algorithm flow of the central difference Kalman filtering and the conjugate gradient method, we present the outline of the central difference variational filtering algorithm as follows [
26].
Step 1: Sigma points generation. At time instant
, generate sigma points
as (
14).
Step 2: Prediction. Conduct (
15) and
Step 3: Update.
(1) State update.
Define the cost function
as
The object is to find
and
.
In order to use the conjugate gradient method, the following transformation is required: , where is determined by the one-step prediction covariance matrix by .
By the transformation, the cost function can be revised as
where
. Then it can be easily be seen that
and
For the application of ADPOS, denotes the measurement noise covariance, which is decided by the noise level of the sensors. Then is a diagonal matrix with positive elements, and the second-order derivative is positive definite. In addition, can be obtained from the optimization process of by the conjugate gradient method and .
The specific optimization process is given as follows:
- (1)
Initialize , , = 0, ;
- (2)
Calculate , ;
- (3)
;
- (4)
;
- (5)
;
- (6)
;
Then, by (1)–(6), one can use the conjugate gradient method to obtain the estimation of recursively, and then can be calculated as the state estimation.
(2) Covariance matrix update.
One can use the measurement linearity and Formula (
28) to calculate the error covariance matrix update.
By the above descriptions, the proposed central difference variational filtering algorithm has been presented, the detailed algorithm flowchart is shown in
Figure 1, and the pseudo code is described in Algorithm 2.
Algorithm 2: Central difference variational filtering algorithm |
Initialization: |
Prediction |
For to N do |
1: |
2: , |
, |
3: |
4: |
5: |
|
|
6: |
end |
Update |
Initialization: , , = 0, |
For do |
7. , |
8. |
9. |
11. |
10. |
11. |
12: |
13: |
end |
Return , |