Next Article in Journal
The Dynamical and Kinetic Equations of Four-Five-Six-Wave Resonance for Ocean Surface Gravity Waves in Water with a Finite Depth
Previous Article in Journal
A Transformer and LSTM-Based Approach for Blind Well Lithology Prediction
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

High-Order Extended Kalman Filter for State Estimation of Nonlinear Systems

1
School of Information and Control Engineering, Jilin Institute of Chemical Technology, Jilin 132022, China
2
School of Automation, Guangdong University of Petrochemical Technology, Maoming 525000, China
*
Author to whom correspondence should be addressed.
Symmetry 2024, 16(5), 617; https://doi.org/10.3390/sym16050617
Submission received: 15 April 2024 / Revised: 10 May 2024 / Accepted: 13 May 2024 / Published: 16 May 2024
(This article belongs to the Section Engineering and Materials)

Abstract

:
In general, the extended Kalman filter (EKF) has a wide range of applications, aiming to minimize symmetric loss function (mean square error) and improve the accuracy and efficiency of state estimation. As the nonlinear model complexity increases, rounding errors gradually amplify, leading to performance degradation. After multiple iterations, divergence may occur. The traditional extended Kalman filter cannot accurately estimate the nonlinear model, and these errors still have an impact on the accuracy. To improve the filtering performance of the extended Kalman filter (EKF), this paper proposes a new extended Kalman filter (REKF) method that utilizes the statistical properties of the rounding error to enhance the estimation accuracy. After establishing the state model and measurement model, the residual term is used to replace the higher-order term in the Taylor expansion, and the least squares method is applied to identify the residual term step by step. Then, the iterative process of updating the extended Kalman filter is carried out. Within the Kalman filter framework, a higher-order rounding error-based extended Kalman filter (REKF) is designed for the joint estimation of rounding error and random variables, and the solution method for the rounding error is considered for the multilevel approximation of the original function. Through numerical simulations on a general nonlinear model, the higher-order rounding error-based extended Kalman filter (REKF) achieves better estimation results than the extended Kalman filter (EKF) and improves the filtering accuracy by utilizing the higher-order rounding error information, which also proves the effectiveness of the proposed method.

1. Introduction

The Kalman filter (KF) is a popular state estimation algorithm used in a variety of applications, including localization, navigation, sensor networks, and battery management [1]. The Kalman filter estimates the state for a linear system and obtains an optimal estimate in the presence of stochastic disturbances; therefore, the KF method is also known as the optimal linear estimation method [2]. The KF is based on the minimum mean square error for state estimation, and the state estimation process can be regarded as a filtering process because the state and the observed data are usually affected by system noise [3]. The KF is a recursive estimation algorithm that iteratively obtains updated state values and predicted values. It provides a powerful tool to deal with the state estimation problem of linear systems [4]. However, many engineering systems cannot be described by simple linear systems. For example, the autopilot system of a car and the target tracking system of a drone are complex nonlinear systems. If we only use the Kalman filter (KF) for state estimation, it cannot reach the accuracy we need. When the drone performs tracking tasks or the car turns on the autopilot system, it may cause serious accidents. In 2023, in San Francisco, United States, more than 700 cases of car accidents occurred involving automated vehicles that failed to slow down, resulting in hundreds of deaths. Most of these autopilot incidents were the result of insufficient accuracy. For this reason, improving the accuracy of the algorithm is a very important task. In order to solve the problem of nonlinear systems, they need to be converted into linear systems to be solved. Usually, the extended Kalman filter (EKF) is used to deal with nonlinear systems. The EKF algorithm is an algorithm adapted to a variety of complex systems [5]. Currently, the extended Kalman filter (EKF) has successfully become the most widely used state observer for nonlinear systems [6]. The EKF is used in systems with zero-mean Gaussian process uncertainty and measurement noise [7]. The EKF is an extension of the KF for nonlinear dynamical systems in that it provides a more accurate estimate of the state when the modeling of the dynamical system exactly matches the dynamics it describes. The system is linearized around the predicted state estimates [8]. The state estimation process can also be viewed as a filtering process since the state and the observed data are usually affected by system noise. The traditional extended Kalman filter (EKF) has been successfully applied to many systems. The EKF uses a Taylor expansion of the nonlinear part of the model at the estimates, and linearization is achieved by a first-order approximation, which transforms the nonlinear problem into a linear KF problem [9]. However, when the linearization error is large and the model is uncertain, the performance of the EKF will be greatly reduced, and even divergence phenomena can be observed [10,11]. This paper proposes a method to utilize the rounding error to solve the EKF problem. In this paper, we propose a new extended Kalman filter (EKF) method designed to enhance estimation accuracy by leveraging the statistical properties of the rounding error [12,13]. This approach involves approximating the rounding error using a multilevel approximation of the original function and substituting the rounded higher-order term with the residual term in the traditional extended Kalman filter. This modification aims to address nonlinear equations more accurately and stably, making it particularly suitable for engineering systems [14].
The first section introduces KF and EKF’s importance in applied systems, shows that EKF falls short in certain complex systems’ requirements, and proposes a new REKF method to improve precision. The second part mainly explains the principles and calculations of EKF and REKF, compares their performances, and analyzes their theoretical advantages and disadvantages in accuracy and stability. In the third part, MATLAB simulations are conducted to verify the stability and accuracy of the REKF method. Lastly, the article summarizes REKF’s advantages and looks toward the future.

2. Algorithms of EKF and REKF

2.1. Principle of EKF Algorithm

The mathematical structure of the extended Kalman filter (EKF) is very simple and uses a recursive approach, where the a posteriori estimate obtained from the last computation is used as the a priori estimate for the next computation [15], and each time, the current state estimate is computed recursively from only the measured variables from the previous computation, which is able to give optimal estimates for the system variables, and the second-order and higher terms are omitted from the Taylor expansion to obtain an approximately linear model [16].
A class of systems exists where the equations of state are nonlinear and the observation equations are linear: in this paper, the rounding error rounded off after Taylor expansion is considered, where ξ ( k ) represents the rounding error of the Taylor expansion of a higher order.
Definition 1.
Consider the following time-invariant stochastic differential equation.
x ( k + 1 ) = f ( x ( k ) ) + w ( k )
y ( k + 1 ) = h ( x ( k + 1 ) ) + v ( k + 1 )
where f ( ) and h ( ) represent the state equation and measurement equation, respectively, k T : = { k | k = 1 , 2 , 3 , T 1 } . The time scale is T , x ( k + 1 ) R n is the n dimensional state vector, y ( k + 1 ) R m is the m dimensional measurement vector, and w ( k ) and v ( k + 1 ) are the state modeling error noise vector and measurement error noise vector, respectively. f ( x ( k ) ) is a nonlinear functional model of the state, with x ( k ) , w ( k ) , k as the common actions. For the state, there are x ( k ) with order derivatives continuous up to r , and order derivatives exist at r + 1 . h ( x ( k + 1 ) ) is a nonlinear functional model of the state with x ( k + 1 ) , v ( k + 1 ) , k + 1 . The vectors w ~ N ( 0 , Q ) and v ~ N ( 0 , R ) are independent and obey Gaussian distributions. The Gaussian random variables denote the process noise and the measurement noise, respectively, whose corresponding covariance matrices are Q and R , and they also satisfy the following statistical properties:
{ E { w ( k ) } = 0 , E { w ( k ) w T ( k ) } = Q ( k ) ; E { v ( k + 1 ) } = 0 , E { v ( k + 1 ) v T ( k + 1 ) } = R ( k + 1 ) ; E { w ( k ) v T ( k + 1 ) } = 0 , E { w ( k ) v T ( k + 1 ) } = 0 ;     E { w ( k ) x T ( k ) } = 0 ,   E { v ( k + 1 ) x T ( k ) } = 0 ;
The filtered estimate at time k is regarded as x ^ ( k | k ) . The nonlinear function f ( x ( k ) ) of the state equation in Equation (1) is Taylor-expanded at x ^ ( k | k ) .
f ( x ( k ) ) = f ( x ^ ( k | k ) ) + l = 1 r 1 l ! r f ( x ^ ( k | k ) ) x r ( k ) | x ( k ) = x ^ ( k | k ) [ x ( k ) x ^ ( k | k ) ] r + 1 ( r + 1 ) ! r + 1 f ( x ^ ( k | k ) ) x r + 1 ( k ) | x ( k ) = ξ ( k ) [ x ( k ) x ^ ( k | k ) ] r + 1
By omitting the second-order and its higher-order term, the following is obtained:
f ( x ( k ) ) f ( x ^ ( k | k ) ) + f ( x ^ ( k | k ) ) x ( k ) | x ( k ) = x ^ ( k | k ) [ x ( k ) x ^ ( k | k ) ]
In Equation (5), let:
A ( k + 1 | k ) = f ( x ^ ( k | k ) ) x ( k ) | x ( k ) = x ^ ( k | k )
f ( x ^ ( k | k ) ) = f ( x ^ ( k | k ) ) A ( k + 1 | k ) x ^ ( k | k )
The equation of the state with noise x ( k + 1 ) is as follows:
x ( k + 1 ) A ( k + 1 | k ) x ( k ) + w ( k ) + f ( x ^ ( k | k ) )
The nonlinear part h ( x ( k + 1 ) ) in Formula (2) is expanded by the Taylor formula in the estimate x ^ ( k + 1 | k ) .
h ( x ( k + 1 ) ) = h ( x ^ ( k + 1 | k ) ) + l = 1 r 1 l ! r h ( x ^ ( k + 1 | k ) ) x r ( k + 1 ) | x ( k + 1 ) = x ^ ( k + 1 | k ) [ x ( k + 1 ) x ^ ( k + 1 | k ) ] r + 1 ( r + 1 ) ! r + 1 h ( x ^ ( k + 1 | k ) ) x r + 1 ( k + 1 ) | x ( k + 1 ) = x ^ ( k + 1 | k ) [ x ( k + 1 ) x ^ ( k + 1 | k ) ] r + 1
By omitting the second-order and its higher-order term, the following is obtained:
h ( x ( k + 1 ) ) h ( x ^ ( k + 1 | k ) ) + h ( x ^ ( k + 1 | k ) ) x ( k + 1 ) | x ( k + 1 ) = x ^ ( k + 1 | k ) [ x ( k + 1 ) x ^ ( k + 1 | k ) ]
In Equation (10), let:
H ( k + 1 ) = h ( x ^ ( k + 1 | k ) ) x ( k + 1 ) | x ( k + 1 ) = x ^ ( k + 1 | k )
h ( x ^ ( k + 1 | k ) ) = h ( x ^ ( k + 1 | k ) ) H ( k + 1 ) x ^ ( k + 1 | k )
The measurement equation can be approximately represented as:
y ( k + 1 ) = H ( k + 1 ) x ( k + 1 ) + v ( k + 1 ) + h ( x ^ ( k + 1 | k ) )
where x ^ ( k + 1 | k ) is the state prediction value of the Kalman filter, y ( k + 1 ) is the measurement observation of the Kalman filter, A ( k + 1 | k ) and H ( k + 1 ) are the first-order Jacobi matrices of the nonlinear model of the system where the state equation and the observation equation are derived at x ^ ( k | k ) and x ^ ( k + 1 | k ) , and f ( x ^ ( k | k ) ) and h ( x ^ ( k + 1 | k ) ) are known terms of the Taylor expansion of the equations of state and observation and can be regarded as a constant term.
Extending the prediction and update process of the Kalman filter:
The prediction estimates and prediction error values for the state equations in the model can be obtained according to Equation (8):
Projected estimates:
x ^ ( k + 1 | k ) A ( k + 1 | k ) x ^ ( k | k ) + f ( x ^ ( k | k ) )
Predicted error value:
x ˜ ( k + 1 | k ) A ( k + 1 | k ) x ˜ ( k | k ) + w ( k )
The measurement estimates and measurement estimation errors of the measurement equations in the model can be obtained according to Equation (13):
Measurement estimate:
y ^ ( k + 1 | k ) H ( k + 1 ) x ^ ( k + 1 | k ) + h ( x ^ ( k + 1 | k ) )
Measurement estimation error:
y ˜ ( k + 1 | k ) H ( k + 1 ) x ˜ ( k + 1 | k ) + v ( k + 1 )
Predicted error covariance:
P x x ( k + 1 | k ) A ( k + 1 | k ) P ( k | k ) A ( k + 1 | k ) T + Q ( k )
The filter values for the Kalman observer with time-varying gain are as follows:
x ( k + 1 | k + 1 ) x ^ ( k + 1 | k ) + K ( k + 1 ) ( y ˜ ( k + 1 | k ) )
According to the orthogonality principle E { x ˜ ( k + 1 | k + 1 ) y T ( k + 1 ) } = 0 , the Kalman gain K ( k + 1 ) is solved for:
K ( k + 1 ) P x x ( k + 1 | k ) H ( k + 1 ) T ( H ( k + 1 ) P x x ( k + 1 | k ) H ( k + 1 ) T + R ( k + 1 ) ) 1
The state x ( k + 1 ) estimation error covariance was calculated:
P x x ( k + 1 | k + 1 ) ( I K ( k + 1 ) H ( k + 1 ) ) P x x ( k + 1 | k )
According to the standard extended Kalman filter, the statistical properties of the initial state x ( 0 ) are known: E { x ( 0 ) } = x ^ 0 ,   E { ( x ( 0 ) x ^ 0 ) ( x ( 0 ) x ^ 0 ) T } = P 0 [17], where x ˜ ( k + 1 | k ) is the estimation error, y ^ ( k + 1 | k ) is the prediction observation, y ˜ ( k + 1 | k ) is the error, P x x ( k + 1 | k ) and P x x ( k + 1 | k + 1 ) represent the prediction error covariance and estimation error covariance, K ( k + 1 ) represents the Kalman gain, and x ( k + 1 | k + 1 ) represents the filtered value of the extended Kalman.

2.2. Extended Kalman Filter with Residuals (REKF)

Based on the extended Kalman filter (EKF), the consideration of higher-order terms in the Taylor expansion of the state equation can further improve the model’s accuracy by introducing residuals to replace the higher-order terms. By applying the least squares method step by step to identify the higher-order terms replaced by residuals, the residual extended Kalman filter (REKF) enhances the utilization of the effects of higher-order terms compared to the EKF. This can reflect the real dynamics of the system more accurately so as to improve the filtering accuracy. This method is particularly suitable for systems with strong nonlinearity and can effectively enhance the accuracy of state estimation.

2.2.1. Extended Kalman Filtering Using Residual Terms Instead of Higher-Order Terms

Assumption 1.
In Equation (1), the equation of state  x ( k + 1 )  has continuity up to the order derivative of r, and the order derivative of  r + 1  exists.
The nonlinear function f ( x ( k ) ) in the model when Assumption 1 holds is given by the x ^ ( k | k ) Taylor expansion:
x ( k + 1 ) = f ( x ^ ( k | k ) ) + l = 1 r 1 l ! r f ( x ^ ( k | k ) ) x r ( k ) | x ( k ) = x ^ ( k | k ) [ x ( k ) x ^ ( k | k ) ] r + 1 ( r + 1 ) ! r + 1 f ( x ^ ( k | k ) ) x r + 1 ( k ) | x ( k ) = ξ ( k ) [ x ( k ) x ^ ( k | k ) ] r + 1 + w ( k )
Replace the second-order term and its higher-order term with the residual term of the equation of state:
x ( k + 1 ) = f ( x ( k ) ) + w ( k )   = f ( x ^ ( k | k ) ) + A ( k + 1 | k ) x ( k ) + ξ ( k ) + w ( k )
Among these, ξ ( k ) = ξ 1 ( k ) , ξ ( k ) represents the second order of the Taylor expansion and its higher-order terms.
The residual term in the equation of state is first identified by the least squares method. The identification process is as follows:
Replace the second-order term of the Taylor expansion with a residual term in the equation for which the Taylor expansion is performed at x ^ ( k | k ) :
x ( k + 1 ) = f ( x ( k ) ) + w ( k )   = f ( x ^ ( k | k ) ) + A ( k + 1 | k ) x ( k ) + ξ 1 ( k ) + w ( k )
Bring Equation (24) into the observation equation:
y ( k + 1 ) = H ( k + 1 ) x ( k + 1 ) + v ( k + 1 ) + h ( x ^ ( k + 1 | k )   = H ( k + 1 ) ( f ( x ^ ( k | k ) + A ( k + 1 | k ) x ( k )               + ξ 1 ( k ) + w ( k ) ) + v ( k + 1 ) + h ( x ^ ( k + 1 | k )
Reduce (25) to (26):
y ( 1 ) ( k + 1 ) = H ( k + 1 ) ξ 1 ( k ) + v ( 1 ) ( k + 1 )
Among these, y ( 1 ) ( k + 1 ) and v ( 1 ) ( k + 1 ) . The parameters are as follows:
{ y ( 1 ) ( k + 1 ) = y ( k + 1 ) H ( k + 1 ) f ( x ^ ( k | k ) H ( k + 1 ) A ( k + 1 | k ) x ^ ( k | k ) h ( x ^ ( k + 1 | k ) v ( 1 ) ( k + 1 ) = H ( k + 1 ) w ( k ) + v ( k + 1 ) + H ( k + 1 ) A ( k + 1 | k ) x ˜ ( k | k )
Based on the least squares method, ξ 1 ( k ) , ξ 1 ( k ) in (26) is identified as a Gaussian distribution with ξ ^ 1 ( k | k ) as the variance and P ξ ( 1 ) ( k | k ) as the variance, ξ 1 ( k ) ~ [ ξ ^ 1 ( k | k ) , P ξ ( 1 ) ( k | k ) ] . The identification process is as follows:
{ ξ 1 ( k ) = ξ ^ 1 ( k | k ) + ξ ˜ 1 ( k | k ) ξ ^ 1 ( k | k ) = H T ( k + 1 ) [ H ( k + 1 ) H T ( k + 1 ) ] 1 ( y ( 1 ) ( k + 1 ) )
Judge whether to meet the system accuracy requirements: By setting the threshold λ , determine whether the paradigm ξ ^ 1 ( k | k )   < λ of the estimate of the residual term is valid. If it is valid, then it means that the Taylor expansion only needs to be expanded to the second order to reach the accuracy needed. If it is not valid, then the Taylor expansion needs to continue (the specific parameters are in Appendix A).
It can be recursively proved up to order 2 and r − 1: Appendix B.
Definition 2.
According to the mathematical induction method, it can be deduced to the order of  r + 1 . Assuming that the system accuracy requirement is satisfied when the Taylor expansion is carried out to the order of r + 1 , the residual term ξ r ( k | k ) is used to replace the order term of r + 1 after the Taylor expansion, ξ ^ r ( k | k ) is the valid information extracted from ξ ˜ r 1 ( k | k ) , and ξ ˜ r 1 ( k | k ) is attributed to ξ ˜ r ( k | k ) after the feature extraction. Moreover, ξ ( k ) = ξ 1 ( k ) + + ξ r ( k ) .
x ( k + 1 ) = f ( x ( k ) ) + w ( k )   = f ( x ^ ( k | k ) ) + A ( k + 1 | k ) x ( k ) + ξ 1 ( k ) + + ξ r ( k ) + w ( k )
Bring Equation (29) into the observation equation:
y ( k + 1 ) = H ( k + 1 ) x ( k + 1 ) + v ( k + 1 ) + h ( x ^ ( k + 1 | k )   = H ( k + 1 ) ( f ( x ^ ( k | k ) ) + A ( k + 1 | k ) x ( k )           + ξ 1 ( k ) + + ξ r ( k ) + w ( k ) ) + v ( k + 1 ) + h ( x ^ ( k + 1 | k )
Organize (30) as:
y ( r ) ( k + 1 ) = H ( k + 1 ) ξ r ( k ) + v ( r ) ( k + 1 )
The parameter is y ( r ) ( k + 1 ) . The equation is:
y ( r ) ( k + 1 ) = y ( k + 1 ) H ( k + 1 ) f ( x ^ ( k | k ) H ( k + 1 ) A ( k + 1 | k ) x ^ ( k | k )   H ( k + 1 ) ξ ^ 1 ( k | k ) H ( k + 1 ) ξ ^ r 1 ( k | k ) h ( x ^ ( k + 1 | k )
The parameter is v ( r ) ( k + 1 ) . The equation is:
v ( r ) ( k + 1 ) = H ( k + 1 ) A ( k + 1 | k ) x ˜ ( k | k ) + H ( k + 1 ) w ( k ) + v ( k + 1 )
Using the least squares method, the identification parameters ξ r ( k ) , ξ r ( k ) conform to a Gaussian distribution, with ξ ^ r ( k | k ) as the mean and P ξ ( r ) ( k | k ) as the variance:
ξ ^ r ( k | k ) = H ( k + 1 ) T [ H ( k + 1 ) H ( k + 1 ) T ] 1 ( y ( r ) ( k + 1 ) )
In nonlinear models, when the Taylor expansion is carried out up to the nth order, the approximation is valid, satisfying the required system accuracy, and the parameter identification for all the higher-order terms is complete, with the residual term replacing the higher-order terms (the specific parameters are in Appendix C).
REKF algorithm flow chart (Figure 1):

2.2.2. REKF Implementation Projections and Updates

The nonlinear observation equations and nonlinear state equations have been linearized, allowing the prediction and update steps to be performed in accordance with the Kalman filtering process. This approach streamlines and optimizes the overall filtering procedure. The process is shown in Figure 1, and the steps are as follows:
The state and observation equations for a class of nonlinear system models are:
x ( k + 1 ) = f ( x ( k ) ) + w ( k )   = f ( x ^ ( k | k ) ) + A ( k + 1 | k ) x ( k ) + ξ 1 ( k ) + + ξ r ( k ) + w ( k )
(1)
REKF prediction steps:
The predicted estimate of the state equation:
x ^ ( k + 1 | k ) = f ( x ^ ( k | k ) ) + A ( k + 1 | k ) x ^ ( k | k ) + ξ ^ 1 ( k ) + + ξ ^ r ( k )
The prediction error of the state equation:
x ˜ ( k + 1 | k ) = A ( k + 1 | k ) x ˜ ( k | k ) + ξ ˜ r ( k ) + w ( k )
In this context, ξ ˜ ( k ) represents the prediction error of the residual term in the state equation, while ξ ^ ( k ) denotes the predicted estimate of the residual term in the state equation.
According to the solution formula of Kalman filtering, it can be seen that the estimation error covariance P ( k + 1 | k ) is as follows:
P ( k + 1 | k ) = A ( k + 1 | k ) P ( k | k ) A ( k + 1 | k ) T   + P ξ ( r ) ( k + 1 | k ) + Q ( k )
In this context, P ( k | k ) = { x ˜ ( k | k ) x ˜ ( k | k ) T } represents the estimated error covariance at time k.
The predicted estimate of the measurement equation:
y ^ ( k + 1 | k ) = H ( k + 1 ) x ^ ( k + 1 | k ) + h ( x ^ ( k + 1 | k ) )
Predictive estimation errors in the measurement equations:
y ˜ ( k + 1 | k ) = H ( k + 1 ) x ˜ ( k + 1 | k ) + v ( k + 1 )
(2)
Update steps:
Kalman gain K k ( k + 1 ) :
K k ( k + 1 ) = P ( k + 1 | k ) H ( k + 1 ) T ( H ( k + 1 ) P ( k + 1 | k ) H ( k + 1 ) T + R ( k + 1 ) ) 1
Expanded-Kalman-filtered value of the state equation at k + 1 time:
x ^ ( k + 1 | k + 1 ) = x ^ ( k + 1 | k ) + K k ( k + 1 ) y ˜ ( k + 1 | k )
Find the estimation error covariance at the time of k + 1 :
P ( k + 1 | k + 1 ) = E { x ˜ ( k + 1 | k + 1 ) ( x ˜ ( k + 1 | k + 1 ) ) T }   = ( I K k ( k + 1 ) H ( k + 1 ) ) P ( k + 1 | k )
The extended Kalman prediction and update solution for the remaining substitutions is complete.

2.3. REKF Performance Analysis

(1)
Predictive stage performance analysis
During the prediction stage, the REKF state equation performs better than the EKF in feature extraction. It utilizes a multi-level approximation method for rounding error calculation to improve the estimation accuracy. And the error exists only in the highest-order residual term.
In terms of performance, the smaller the error covariance, the closer the filtered value in the system is to the true value. During the prediction stage, a (r + 1)th order x r + 1 ( k + 1 ) in the Taylor expansion is considered the true value, and the error exists only in the highest-order residual term. Refer to Appendix D for details.
The following content focuses solely on presenting the results.
Taylor expansion to an r + 1 order of error covariance:
p ( r + 1 ) x x ( k + 1 | k ) = A ( k + 1 | k ) p ( k | k ) A ( k + 1 | k ) T + p ( r ) ζ ( k | k ) + Q ( k )
Taylor expansion to an  order of error covariance:
p ( r ) x x ( k + 1 | k ) = p ( r + 1 ) x x ( k + 1 | k ) + ξ ^ ( r ) ( k | k ) ξ ^ ( r ) ( k | k ) T
Estimated error covariance in the Taylor expansion to the first order:
p ( 1 ) x x ( k + 1 | k ) = p ( 2 ) x x ( k + 1 | k ) + ξ ^ ( 1 ) ( k | k ) ξ ^ ( 1 ) ( k | k ) T
ξ ^ ( 1 ) ( k | k ) ξ ^ ( 1 ) ( k | k ) T ξ ^ ( r ) ( k | k ) ξ ^ ( r ) ( k | k ) T are all greater than or equal to 0, and ξ ^ ( 1 ) ( k ) ξ ^ ( 1 ) ( k ) T is equal to 0 only if the matrix is orthogonal to ξ ^ ( 1 ) ( k ) ξ ^ ( 1 ) ( k ) T . Similarly, ξ ^ ( r ) ( k | k ) ξ ^ ( r ) ( k | k ) T is equal to 0 only if the matrix is orthogonal to ξ ^ ( r ) ( k | k ) ξ ^ ( r ) ( k | k ) T .
From this, it follows that:
p ( 1 ) x x ( k + 1 | k ) p ( 2 ) x x ( k + 1 | k ) p ( r + 1 ) x x ( k + 1 | k )
Expand to the r + 1 order error covariance minimum:
p ( r ) x x ( k + 1 | k ) = p ( r + 1 ) x x ( k + 1 | k ) + ξ ^ ( r ) ( k | k ) ξ ^ ( r ) ( k | k ) T p ( r 1 ) x x ( k + 1 | k ) = p ( r ) x x ( k + 1 | k ) + ξ ^ ( r 1 ) ( k | k ) ξ ^ ( r 1 ) ( k | k ) T p ( 2 ) x x ( k + 1 | k ) = p ( 3 ) x x ( k + 1 | k ) + ξ ^ ( 2 ) ( k | k ) ξ ^ ( 2 ) ( k | k ) T p ( 1 ) x x ( k + 1 | k ) = p ( 2 ) x x ( k + 1 | k ) + ξ ^ ( 1 ) ( k | k ) ξ ^ ( 1 ) ( k | k ) T
Upon comparison, it becomes evident that the (r + 1)th order Taylor expansion exhibits the smallest error covariance and the highest accuracy. In the prediction stage, REKF extracts more information, leading to a reduced prediction error covariance matrix. Consequently, the prediction accuracy of REKF is superior to that of the traditional EKF.
(2)
Analysis of performance indicators in the update phase
Our REKF accuracy is greater than EKF in the prediction phase and in the update phase:
p ( 1 ) x x ( k + 1 | k + 1 ) 1 = p ( 1 ) x x ( k + 1 | k ) 1 + ( H T ( k + 1 ) R 1 ( k ) H ( k + 1 ) ) ; p ( 2 ) x x ( k + 1 | k + 1 ) 1 = p ( 2 ) x x ( k + 1 | k ) 1 + ( H T ( k + 1 ) R 1 ( k ) H ( k + 1 ) ) ; p ( r ) x x ( k + 1 | k + 1 ) 1 = p ( r ) x x ( k + 1 | k ) 1 + ( H T ( k + 1 ) R 1 ( k ) H ( k + 1 ) ) ; p ( r + 1 ) x x ( k + 1 | k + 1 ) 1 = p ( r + 1 ) x x ( k + 1 | k ) 1 + ( H T ( k + 1 ) R 1 ( k ) H ( k + 1 ) ) ;
Available by comparison:
p ( 1 ) x x ( k + 1 | k + 1 ) 1 p ( 2 ) x x ( k + 1 | k + 1 ) 1 p ( r + 1 ) x x ( k + 1 | k + 1 ) 1
By analyzing the performance of EKF and REKF, it becomes evident that the new extended Kalman filter leverages the statistical properties of the rounding error to enhance the estimation accuracy. A higher-order extended Kalman filter (REKF) was established for the joint estimation of rounding errors and random variables. The method of solving rounding errors, considering the multilevel approximation of the original function, offers higher accuracy compared to the traditional EKF approach.

3. Simulation Experiments

Simulation experiments can be conducted to verify the enhanced performance of the REKF compared to the EKF. These experiments can be carried out using MATLABR 2021b software.
Simulation experiment I: Consider a system where the measurement equation is linear and the equation of state is nonlinear.
x ( k + 1 ) = [ x 1 ( k + 1 ) x 2 ( k + 1 ) ] = [ 5 sin x 1 ( k + 1 ) + 5 cos x 2 ( k + 1 ) sin x 1 ( k + 1 ) + cos x 2 ( k + 1 ) ] + w k [ y ( k + 1 ) ] = [ 2 x 1 ( k + 1 ) ] + [ 2 x 2 ( k + 1 ) ] + v k + 1
where x ( k + 1 ) = [ x 1 , x 2 ] T , y ( k + 1 ) = [ y 1 , y 2 ] T , w k = [ w 1 , w 2 ] T , v k + 1 = [ v 1 , v 2 ] T w k , and v k + 1 are independent white noises obeying normal distribution in the system, w k ~ N ( 0 , Q ) and v k + 1 ~ N ( 0 , R ) , and satisfy the conditions in Equation (3). w k and v k + 1 are mutually uncorrelated white noise, Q = d i a g { 0.7 , 1.3 } and R = 1 , assuming that the initial true value of the original state model is the initial value of the original model state, x ^ 0 = [ 1 , 1 ] T , and the initial estimation error covariance matrix is P 0 = I R 2 × 2 . The estimation of the target states x 1 and x 2 are compared to the target states using the extended Kalman filter (EKF) with the Kalman filter with high-precision residual term substitution (REKF).
Figure 2 shows that the method proposed in this paper is more accurate.
Table 1 shows that: In X1, the precision of the high-order REKF with residual substitution was improved by about 6.92% compared to the traditional EKF, and the precision of the high-order REKF with residual substitution was improved by about 5.62% compared to the first-order REKF with residual substitution.
In the scenario where the state equation exhibited weak nonlinearity and the observation equation was linear, the REKF demonstrated superior performance compared to the EKF in terms of error reduction and stability improvement.
Table 2 shows that: In X2, the precision of the high-order REKF with residual substitution was improved by about 7.28% compared to the traditional EKF, and the precision of the high-order REKF with residual substitution was improved by about 5.95% compared to the first-order REKF with residual substitution.
A simulation experiment was conducted using MATLAB, with the data in the table representing the root mean square error obtained by averaging the results from 50 iterations of the Kalman filter experiment. In scenarios where the state equation exhibited weak nonlinearity and the measurement was linear, with relatively large random noise in the function model, the figure clearly shows that the REKF outperformed the EKF in terms of accuracy. Furthermore, in the case of an under-measured nonlinear model, the enhancement of the first-order REKF with residual term replacement was smaller, while the enhancement of the higher-order REKF with residual term replacement was larger. This observation is also validated in the table, where it can be seen that the accuracy of the higher-order REKF with residual term replacement was about 3% higher than that of the first order REKF with residual term replacement. Additionally, the accuracy of the higher-order REKF with residual term replacement was about 6% higher than that of the traditional EKF. The REKF with residual term replacement performed better than the traditional EKF method in practical applications.
Simulation Experiment II: Consider systems where both the measurement equation and the equation of state are weakly nonlinear.
[ x 1 ( k + 1 ) x 2 ( k + 1 ) ] = [ x 1 ( k ) x 2 ( k ) 1 6 x 1 3 ( k ) 1 6 x 2 3 ( k ) + 1 120 x 1 5 ( k ) + 1 120 x 2 5 ( k ) 1 1 2 x 1 2 ( k ) 1 2 x 2 2 ( k ) + 1 24 x 1 4 ( k ) + 1 24 x 2 4 ( k ) ] + [ w 1 ( k ) w 2 ( k ) ] [ y ( k + 1 ) ] = [ x 1 ( k + 1 ) + x 2 ( k + 1 ) + 1 2 x 1 2 ( k + 1 ) x 2 ( k + 1 ) + 1 2 x 1 ( k + 1 ) x 2 2 ( k + 1 ) + sin x 1 ( k + 1 ) ] + v ( k + 1 )
where x ( k + 1 ) = [ x 1 , x 2 ] T , y ( k + 1 ) = [ y 1 , y 2 ] T , w k = [ w 1 , w 2 ] T , and v k + 1 = [ v 1 , v 2 ] T . w k and v k + 1 are mutually uncorrelated white noise. Q = d i a g { 0.01 , 0.01 } , R = 0.5 , w k , and v k + 1 are independent white noises in the system obeying normal distribution, w k ~ N ( 0 , Q ) and v k + 1 ~ N ( 0 , R ) , and satisfy the condition in Equation (3). The initial true value of the model is assumed to be the initial value of the original model state, x ^ 0 = [ 0.1 , 0.1 ] T , and the initial estimation error covariance matrix is P 0 = I R 2 × 2 . Figure 3 shows that the method proposed in this paper is more accurate.
Table 3 shows that: In X1, the precision of the high-order REKF with residual substitution was improved by about 5.06% compared to the traditional EKF, and the precision of the high-order REKF with residual substitution was improved by about 3.07% compared to the first-order REKF with residual substitution.
Table 4 shows that: In X2, the precision of the high-order REKF with residual substitution was improved by about 6.71% compared to the traditional EKF, and the precision of the high-order REKF with residual substitution was improved by about 4.88% compared to the first order REKF with residual substitution.
MATLAB was used to carry out the simulation experiment. Since the measurement equation and the state equation are nonlinear functions of the model, a comparison was made between the traditional EKF and the REKF methods mentioned in this paper. The data in the figures are the average values after running the code 50 times. The figure showing the comparison curves of the root mean square errors shows that the residual term increased the accuracy of the traditional EKF method by about 2% compared to the first-order REKF method. Additionally, the higher-order residual term improved the accuracy by about 6% compared to the first-order REKF method and was more accurate than the traditional EKF method. In order to improve the accuracy, the higher-order residual term should be used instead of the extended Kalman filter (REKF). REKF uses Taylor expansion to round off higher-order terms in the model and extract higher-order items of information, which results in better filtering performance compared to the EKF.

4. Conclusions

In this paper, we propose and design an advanced version of the extended Kalman filter (REKF) that specifically accounts for the effects of higher-order terms. Through a series of detailed experiments, we demonstrated that in dealing with nonlinear Gaussian systems, the estimation accuracy of the REKF significantly surpassed that of the traditional extended Kalman filter (EKF). We discovered that the REKF, which includes higher-order terms, not only has greater accuracy but also exhibits higher stability under various conditions. However, it should be noted that our research mainly focuses on scenarios where the state equation is nonlinear. When the measurement equation is also nonlinear, rounding errors are introduced. These rounding errors may have a correlation between the equation of state and the measurement equation. In the error covariance, the product term of the rounding errors of the equation of state and the equation of observation will appear. We have not found an effective solution in our current research. In future work, we plan to delve deeper into and resolve this issue to further optimize and improve the performance of the REKF.

Author Contributions

Conceptualization, L.D. and C.W.; methodology, C.W.; software, L.D.; validation, L.D.; formal analysis, L.D.; investigation, L.D.; resources, L.D.; data curation, L.D.; writing—original draft preparation, L.D.; writing—review and editing, C.W.; visualization, C.W.; supervision, C.W.; project administration, C.W.; funding acquisition, C.W. All authors have read and agreed to the published version of the manuscript.

Funding

(1) National Key R&D Program Intelligent Robot Key Special Project “Robot Joint Drive Control Integrated Chip” 2023YFB4704000; (2) National Natural Science Foundation of China 62125307, U22A2046, 61933013.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

Equation of state with a residual term:
x ( k + 1 ) = f ( x ^ ( k | k ) ) + A ( k + 1 | k ) x ( k ) + ξ 1 ( k ) + w ( k )
Replace the second-order term of the Taylor expansion with a residual term in the equation for which the Taylor expansion is performed at x ^ ( k | k ) :
x ^ ( k + 1 | k ) = f ( x ^ ( k | k ) ) + A ( k + 1 | k ) x ^ ( k | k ) + ξ ^ 1 ( k | k )
x ˜ ( k + 1 | k ) = A ( k + 1 | k ) x ˜ ( k | k ) + ξ ˜ 1 ( k | k ) + w ( k )
Bring Equation (A1) into the observation equation:
y ( k + 1 ) = H ( k + 1 ) x ( k + 1 ) + v ( k + 1 ) + h ( x ^ ( k + 1 | k )   = H ( k + 1 ) ( f ( x ^ ( k | k ) + A ( k + 1 | k ) x ( k )               + ξ 1 ( k ) + w ( k ) ) + v ( k + 1 ) + h ( x ^ ( k + 1 | k )
Reduce (A4) to (A5):
y ( 1 ) ( k + 1 ) = H ( k + 1 ) ξ 1 ( k ) + v ( 1 ) ( k + 1 )
where y ( 1 ) ( k + 1 ) and v ( 1 ) ( k + 1 ) . The parameters are as follows:
{ y ( 1 ) ( k + 1 ) = y ( k + 1 ) H ( k + 1 ) f ( x ^ ( k | k ) H ( k + 1 ) A ( k + 1 | k ) x ^ ( k | k ) h ( x ^ ( k + 1 | k ) v ( 1 ) ( k + 1 ) = H ( k + 1 ) w ( k ) + v ( k + 1 ) + H ( k + 1 ) A ( k + 1 | k ) x ˜ ( k | k )
v ( 1 ) ( k + 1 ) obeys a normal distribution with a mean of 0 and a variance of R ( 1 ) ( k + 1 ) :
R ( 1 ) ( k + 1 ) = E { v ( 1 ) ( k + 1 ) v ( 1 ) ( k + 1 ) T } = H ( k + 1 ) A ( k + 1 | k ) p x x ( 1 ) ( k | k ) A ( k + 1 | k ) T H ( k + 1 ) T         + H ( k + 1 ) Q ( k ) H ( k + 1 ) T + R ( k + 1 )
where p x x ( 1 ) ( k | k ) = E { x ˜ ( k | k ) x ˜ ( k | k ) T } .
Based on the least squares method, ξ 1 ( k ) is identified as a Gaussian distribution in (A5), with ξ ^ 1 ( k | k ) as the variance and P ξ ( 1 ) ( k | k ) as the variance, ξ 1 ( k ) ~ [ ξ ^ 1 ( k | k ) , P ξ ( 1 ) ( k | k ) ] . The identification process is as follows:
{ ξ 1 ( k ) = ξ ^ 1 ( k | k ) + ξ ˜ 1 ( k | k ) ξ ^ 1 ( k | k ) = H T ( k + 1 ) [ H ( k + 1 ) H T ( k + 1 ) ] 1 ( y ( 1 ) ( k + 1 ) )
When m > n: p ξ ( 1 ) ( k + 1 | k ) = [ H T ( k + 1 ) R ( 1 ) ( k + 1 ) 1 H ( k + 1 ) ] 1 .
When m < n:
p ξ ( 1 ) ( k + 1 | k ) = H T ( k + 1 ) [ H ( k + 1 ) H T ( k + 1 ) ] 1 R ( 1 ) ( k + 1 )   [ H ( k + 1 ) H T ( k + 1 ) ] 1 H T ( k + 1 ) .
When m = n: p ξ ( 1 ) ( k + 1 | k ) . Pick one of these to apply.

Appendix B

Assuming that the Taylor expansion to the second order does not meet the requirements, perform Taylor expansion to the third order, and replace the third-order term and its higher-order terms with the residual term ξ 2 ( k ) . ξ ^ 2 ( k | k ) is the effective information extracted from ξ ˜ 1 ( k | k ) , and after feature extraction, ξ ˜ 1 ( k | k ) is attributed to ξ ˜ 2 ( k | k ) . Moreover, ξ ( k ) = ξ 1 ( k ) + ξ 2 ( k ) .
x ( k + 1 ) = f ( x ( k ) ) + w ( k )   = f ( x ^ ( k | k ) ) + A ( k + 1 | k ) x ( k ) + ξ 1 ( k ) + ξ 2 ( k ) + w ( k )
The prediction estimates and prediction errors for the Taylor expansion of the nonlinear function f ( x ( k ) ) at x ^ ( k | k ) are as follows:
x ^ ( k + 1 | k ) = f ( x ^ ( k | k ) ) + A ( k + 1 | k ) x ^ ( k | k ) + ξ ^ 1 ( k | k ) + ξ ^ 2 ( k | k )
x ˜ ( k + 1 | k ) = A ( k + 1 | k ) x ˜ ( k | k ) + ξ ˜ 2 ( k | k ) + w ( k )
y ( k + 1 ) = H ( k + 1 ) ( f ( x ^ ( k | k ) ) + A ( k + 1 | k ) x ( k )   + ξ 1 ( k ) + ξ 2 ( k ) + w ( k ) ) + v ( k + 1 ) + h ( x ^ ( k + 1 | k ) )
Simplify and organize (A12) as:
y ( 2 ) ( k + 1 ) = H ( k + 1 ) ξ 2 ( k ) + v ( 2 ) ( k + 1 )
The parameter is y ( 2 ) ( k + 1 ) . The equation is as follows:
y ( 2 ) ( k + 1 ) = y ( k + 1 ) H ( k + 1 ) f ( x ^ ( k | k ) H ( k + 1 ) A ( k + 1 | k ) x ^ ( k | k ) H ( k + 1 ) ξ ^ 1 ( k | k ) h ( x ^ ( k + 1 | k )
The parameter is v ( 2 ) ( k + 1 ) . The equation is as follows:
v ( 2 ) ( k + 1 ) = H ( k + 1 ) A ( k + 1 | k ) x ˜ ( k | k )   + H ( k + 1 ) w ( k ) + v ( k + 1 )
v ( 2 ) ( k + 1 ) obeys a normal distribution with a mean of 0 and a variance of R ( 2 ) ( k + 1 ) :
R ( 2 ) ( k + 1 ) = E { v ( 2 ) ( k + 1 ) v ( 2 ) ( k + 1 ) T } = H ( k + 1 ) A ( k + 1 | k ) p x x ( 1 ) ( k | k ) A ( k + 1 | k ) T H ( k + 1 ) T         + H ( k + 1 ) Q ( k ) H ( k + 1 ) T + R ( k + 1 )
Using the least squares method, identify the parameters ξ 2 ( k ) and ξ 2 ( k ) in (A13) as a normal distribution, with ξ ^ 2 ( k | k ) as the mean and P ξ ( 2 ) ( k | k ) as the variance:
ξ ^ 2 ( k | k ) = H ( k + 1 ) T [ H ( k + 1 ) H ( k + 1 ) T ] 1 ( y ( 2 ) ( k + 1 ) )
When m > n: p ξ ( 2 ) ( k + 1 | k ) = [ H T ( k + 1 ) R ( 2 ) ( k + 1 ) 1 H ( k + 1 ) ] 1 .
When m < n: p ξ ( 2 ) ( k + 1 | k ) = H T ( k + 1 ) [ H ( k + 1 ) H T ( k + 1 ) ] 1 R ( 2 ) ( k + 1 ) [ H ( k + 1 ) H T ( k + 1 ) ] 1 H T ( k + 1 )
When m = n: p ξ ( 2 ) ( k + 1 | k ) . Pick one of these to apply.
Judge whether it meets the system accuracy requirement: By setting a threshold λ , determine whether the paradigm ξ ^ 2 ( k | k )   < λ of the estimate of the residual term is valid. If it is valid, then it means there is no need to expand to a higher order. If it is not valid, then Taylor’s expansion needs to be continued.
Assuming that the Taylor expansion to the order of r 1 is still not enough to meet the requirements, expand the Taylor expansion to the order of r , and use the residual term ξ r 1 ( k | k ) to replace the order term of r and its higher-order terms after the Taylor expansion. ξ ^ r 1 ( k | k ) is the valid information extracted from ξ ˜ r 2 ( k | k ) , and after the extraction, ξ ˜ r 2 ( k | k ) is attributed to ξ ˜ r 1 ( k | k ) . Moreover, ξ ( k ) = ξ 1 ( k ) + + ξ r 1 ( k ) .
x ( k + 1 ) = f ( x ( k ) ) + w ( k )   = f ( x ^ ( k | k ) ) + A ( k + 1 | k ) x ( k ) + ξ 1 ( k ) + + ξ r 1 ( k ) + w ( k )
The estimated value x ^ ( k | k ) and the error value x ˜ ( k | k ) of the Taylor expansion of the nonlinear function f ( x ( k ) ) at x ^ ( k | k ) are as follows:
x ^ ( k + 1 | k ) = f ( x ^ ( k | k ) ) + A ( k + 1 | k ) x ^ ( k | k ) + ξ ^ 1 ( k | k ) + + ξ ^ r 1 ( k | k )
x ˜ ( k + 1 | k ) = A ( k + 1 | k ) x ˜ ( k | k ) + ξ ˜ r 1 ( k | k ) + w ( k )
Bring Equation (A18) into the observation equation:
y ( k + 1 ) = H ( k + 1 ) ( f ( x ^ ( k | k ) ) + A ( k + 1 | k ) x ( k )   + ξ 1 ( k ) + + ξ r 1 ( k ) + w ( k ) ) + v ( k + 1 ) + h ( x ^ ( k + 1 | k ) )
Simplify and organize (A21) as:
y ( r 1 ) ( k + 1 ) = H ( k + 1 ) ξ r 1 ( k ) + v ( r 1 ) ( k + 1 )
The parameter is y ( r 1 ) ( k + 1 ) . The equation is as follows:
y ( r 1 ) ( k + 1 ) = y ( k + 1 ) H ( k + 1 ) f ( x ^ ( k | k ) H ( k + 1 ) A ( k + 1 | k ) x ^ ( k | k )   H ( k + 1 ) ξ ^ 1 ( k | k ) H ( k + 1 ) ξ ^ 2 ( k | k ) H ( k + 1 ) ξ ^ r 2 ( k | k ) h ( x ^ ( k + 1 | k ) )
The parameter is v ( r 1 ) ( k + 1 ) . The equation is as follows:
v ( r 1 ) ( k + 1 ) = H ( k + 1 | k ) A ( k + 1 ) x ˜ ( k | k ) + H ( k + 1 ) w ( k ) + v ( k + 1 )
R ( r 1 ) ( k + 1 ) = E { v ( r 1 ) ( k + 1 ) v ( r 1 ) ( k + 1 ) T } = H ( k + 1 ) A ( k + 1 | k ) p x x ( 1 ) ( k | k ) A ( k + 1 | k ) T H ( k + 1 ) T         + H ( k + 1 ) Q ( k ) H ( k + 1 ) T + R ( k + 1 )
Using the least squares method, the parameters ξ r 1 ( k ) and ξ r 1 ( k ) are identified to be normally distributed, with ξ ^ r 1 ( k | k ) as the mean and P ξ ( r 1 ) ( k | k ) as the variance:
ξ ^ r 1 ( k | k ) = H ( k + 1 ) T [ H ( k + 1 ) H ( k + 1 ) T ] 1 ( y ( r 1 ) ( k + 1 ) )
p ξ ( r 1 ) ( k + 1 | k ) = { [ H T ( k + 1 ) R ( r 1 ) ( k + 1 ) 1 H ( k + 1 ) ] 1 m > n H T ( k + 1 ) [ H ( k + 1 ) H T ( k + 1 ) ] 1 R ( r 1 ) ( k + 1 ) [ H ( k + 1 ) H T ( k + 1 ) ] 1 H T ( k + 1 ) m < n
When m = n, p ξ ( r 1 ) ( k + 1 | k ) ; either one of these applies.
Determine whether the system accuracy requirement is met: After the Taylor expansion has reached the order of r , continue to determine whether ξ ^ r 1 ( k | k )   < λ is valid. If it is valid, then there is no need to expand to higher orders; if not, then the Taylor expansion needs to be continued.

Appendix C

Equation of state with a residual term:
x ( k + 1 ) = f ( x ( k ) ) + w ( k )   = f ( x ^ ( k | k ) ) + A ( k + 1 | k ) x ( k ) + ξ 1 ( k ) + + ξ r ( k ) + w ( k )
The prediction estimates x ^ ( k + 1 | k ) and prediction errors x ˜ ( k + 1 | k ) for the Taylor expansion of the nonlinear function f ( x ( k ) ) at x ^ ( k | k ) are given below:
x ^ ( k + 1 | k ) = f ( x ^ ( k | k ) ) + A ( k + 1 | k ) x ^ ( k | k ) + ξ ^ 1 ( k | k ) + + ξ ^ r ( k | k )
x ˜ ( k + 1 | k ) = A ( k + 1 | k ) x ˜ ( k | k ) + ξ ˜ r ( k | k ) + w ( k )
Bring Equation (A27) into the observation equation:
y ( k + 1 ) = H ( k + 1 ) x ( k + 1 ) + v ( k + 1 ) + h ( x ^ ( k + 1 | k )   = H ( k + 1 ) ( f ( x ^ ( k | k ) ) + A ( k + 1 | k ) x ( k )           + ξ 1 ( k ) + + ξ r ( k ) + w ( k ) ) + v ( k + 1 ) + h ( x ^ ( k + 1 | k )
Organize (A30) as:
y ( r ) ( k + 1 ) = H ( k + 1 ) ξ r ( k ) + v ( r ) ( k + 1 )
The parameter is y ( r ) ( k + 1 ) . The equation is:
y ( r ) ( k + 1 ) = y ( k + 1 ) H ( k + 1 ) f ( x ^ ( k | k ) H ( k + 1 ) A ( k + 1 | k ) x ^ ( k | k )   H ( k + 1 ) ξ ^ 1 ( k | k ) H ( k + 1 ) ξ ^ r 1 ( k | k ) h ( x ^ ( k + 1 | k )
The parameter is v ( r ) ( k + 1 ) . The equation is:
v ( r ) ( k + 1 ) = H ( k + 1 ) A ( k + 1 | k ) x ˜ ( k | k ) + H ( k + 1 ) w ( k ) + v ( k + 1 )
R ( r ) ( k + 1 ) = E { v ( r ) ( k + 1 ) v ( r ) ( k + 1 ) T } = H ( k + 1 ) A ( k + 1 | k ) p x x ( 1 ) ( k | k ) A ( k + 1 | k ) T H ( k + 1 ) T         + H ( k + 1 ) Q ( k ) H ( k + 1 ) T + R ( k + 1 )
Using the least squares method, the identification parameters ξ r ( k ) and ξ r ( k ) conform to a Gaussian distribution, with ξ ^ r ( k | k ) as the mean and P ξ ( r ) ( k | k ) as the variance:
ξ ^ r ( k | k ) = H ( k + 1 ) T [ H ( k + 1 ) H ( k + 1 ) T ] 1 ( y ( r ) ( k + 1 ) )
When m > n: p ξ ( r ) ( k + 1 | k ) = [ H T ( k + 1 ) R ( r ) ( k + 1 ) 1 H ( k + 1 ) ] 1
When m < n: p ξ ( r ) ( k + 1 | k ) = H T ( k + 1 ) [ H ( k + 1 ) H T ( k + 1 ) ] 1 R ( r ) ( k + 1 ) [ H ( k + 1 ) H T ( k + 1 ) ] 1 H T ( k + 1 )
When m = n: p ξ ( r ) ( k + 1 | k ) . Pick one of these to apply.

Appendix D

From the error covariance analysis, in the prediction stage, take the remainder term instead of the r + 1 order of the Taylor expansion as the closest to the true value. The predicted and error values are as follows, where x r + 1 ( k + 1 ) represents the true value, x ^ r + 1 ( k + 1 | k ) represents the forecast estimate, x ˜ r + 1 ( k + 1 | k ) represents the forecast error, and the error exists only in the highest-order residual term:
x r + 1 ( k + 1 ) = Δ f ( x ^ ( k | k ) ) + A ( k + 1 | k ) x ( k ) + ξ ( 1 ) ( k ) + + ξ ( r ) ( k ) + w ( k ) x ^ r + 1 ( k + 1 | k ) = Δ f ( x ^ ( k | k ) ) + A ( k + 1 | k ) x ^ ( k | k ) + ξ ^ ( 1 ) ( k | k ) + + ξ ^ ( r ) ( k | k ) x ˜ r + 1 ( k + 1 | k ) = A ( k + 1 | k ) x ˜ ( k | k ) + ξ ˜ ( r ) ( k | k ) + w ( k )
Taylor expansion to r + 1 order of error covariance:
p ( r + 1 ) x x ( k + 1 | k ) = E { x ˜ r + 1 ( k + 1 | k ) x ˜ r + 1 T ( k + 1 | k ) }   = A ( k + 1 | k ) p ( k | k ) A ( k + 1 | k ) T + p ( r ) ζ ( k | k ) + Q ( k )
Predicted and error values for the Taylor expansion to the r order: replace the remainder term with the Taylor expansion to the r + 1 order as the closest true value, and calculate the error for the last remainder term only.
x r ( k + 1 ) = Δ f ( x ^ ( k | k ) ) + A ( k + 1 | k ) x ( k ) + ξ ( 1 ) ( k ) + + ξ ( r 1 ) ( k ) + w ( k ) x ^ r ( k + 1 | k ) = Δ f ( x ^ ( k | k ) ) + A ( k + 1 | k ) x ^ ( k | k ) + ξ ^ ( 1 ) ( k | k ) + + ξ ^ ( r 1 ) ( k | k ) x ˜ r ( k + 1 | k ) = x r ( k + 1 ) x ^ r ( k + 1 | k )   x r + 1 ( k + 1 ) x ^ r ( k + 1 | k )   = A ( k + 1 | k ) x ˜ ( k | k ) + ξ ˜ ( r ) ( k | k ) + ξ ^ ( r ) ( k | k ) + w ( k )
Taylor expansion to  order of error covariance:
p ( r ) x x ( k + 1 | k ) = E { x ˜ r ( k + 1 | k ) x ˜ r T ( k + 1 | k ) }   = A ( k + 1 | k ) p ( k | k ) A ( k + 1 | k ) T + p ( r ) ζ ( k | k )           + ξ ^ ( r ) ( k | k ) ξ ^ ( r ) ( k | k ) T + Q ( k )   = p ( r + 1 ) x x ( k + 1 | k ) + ξ ^ ( r ) ( k | k ) ξ ^ ( r ) ( k | k ) T ;
The predicted and error values for the Taylor expansion to the first order can be obtained by recursion:
x 1 ( k + 1 ) = Δ f ( x ^ ( k | k ) ) + A ( k + 1 | k ) x ( k ) + w ( k ) x ^ 1 ( k + 1 | k ) = Δ f ( x ^ ( k | k ) ) + A ( k + 1 | k ) x ^ ( k | k )
Estimated error covariance in the Taylor expansion to the first order:
p ( 1 ) x x ( k + 1 | k ) = E { x ˜ 1 ( k + 1 | k ) x ˜ 1 ( k + 1 | k ) T }   = A ( k + 1 | k ) p ( k | k ) A ( k + 1 | k ) T + p ( r ) ζ ( k | k )           + ξ ^ ( 1 ) ( k | k ) ξ ^ ( 1 ) ( k | k ) T + + ξ ^ ( r ) ( k | k ) ξ ^ ( r ) ( k | k ) T + Q ( k )   = p ( 2 ) x x ( k + 1 | k ) + ξ ^ ( 1 ) ( k | k ) ξ ^ ( 1 ) ( k | k ) T

References

  1. Bai, Y.; Yan, B.; Zhou, C.; Su, T.; Jin, X. State of art on state estimation: Kalman filter driven by machine learning. Annu. Rev. Control. 2023, 56, 100909. [Google Scholar] [CrossRef]
  2. Muhammad, W.; Ahsan, A. Airship aerodynamic model estimation using unscented Kalman filter. J. Syst. Eng. Electron. 2020, 31, 1318–1329. [Google Scholar] [CrossRef]
  3. Huang, C.; Qiao, F.; Ma, H. Noise reduction of acoustic Doppler velocimeter data based on Kalman filtering and autoregressive moving average models. Acta Oceanol. Sin. 2020, 39, 106–113. [Google Scholar] [CrossRef]
  4. Xiaohui, S.; Chenglin, W.; Tao, W. A Novel Step-by-Step High-Order Extended Kalman Filter Design for a Class of Complex Systems with Multiple Basic Multipliers. Chin. J. Electron. 2021, 30, 313–321. [Google Scholar] [CrossRef]
  5. Majeed, R. Sensorless Field Oriented Control of Induction Motor Drive using Extended Kalman Filter (EKF). In Proceedings of the 2023 International Conference on Circuit Power and Computing Technologies (ICCPCT), Nagercoil, India, 18–19 March 2023; pp. 1044–1048. [Google Scholar]
  6. Rotondo, D. The Weighted Kalman Filter. IFAC-PapersOnLine 2023, 56, 8869–8874. [Google Scholar] [CrossRef]
  7. Valipour, M.; Ricardez-Sandoval, L.A. An Extended Moving Horizon Estimation embedded with an Abridged Gaussian Sum Extended Kalman Filer to handle non-Gaussian noises. IFAC-PapersOnLine 2022, 55, 25–30. [Google Scholar] [CrossRef]
  8. Lee, A.S.; Hilal, W.; Gadsden, S.A.; Al-Shabi, M. Combined Kalman and Sliding Innovation Filtering: An Adaptive Estimation Strategy. Measurement 2023, 218, 113228. [Google Scholar] [CrossRef]
  9. Xiaohui, S.; Chenglin, W.; Tao, W. High-order extended Kalman filter design for a class of complex dynamic systems with polynomial nonlinearities. Chin. J. Electron. 2021, 30, 508–515. [Google Scholar] [CrossRef]
  10. He, L.; Hu, M.; Wei, Y.; Liu, B.; Shi, Q. State of charge estimation by finite difference extended Kalman filter with HPPC parameters identification. Sci. China Technol. Sci. 2020, 63, 410–421. [Google Scholar] [CrossRef]
  11. Jiang, Y.; Huang, Y.; Xue, W.; Fang, H. On designing consistent extended Kalman filter. J. Syst. Sci. Complex. 2017, 30, 751–764. [Google Scholar] [CrossRef]
  12. Yuqi, Y.; Di, Z.; Junlong, L.; Chaofei, L. Time-varying parameters estimation with adaptive neural network EKF for missile-dual control system. J. Syst. Eng. Electron. 2024, 1–12. [Google Scholar]
  13. Al-Masri, W.M.; Abdel-Hafez, M.F.; El-Hag, A.H. Toward high-accuracy estimation of partial discharge location. IEEE Trans. Instrum. Meas. 2016, 65, 2145–2153. [Google Scholar] [CrossRef]
  14. Saleh, S.; El-Wakeel, A.S.; Noureldin, A. 5G-Enabled Vehicle Positioning Using EKF With Dynamic Covariance Matrix Tuning. IEEE Can. J. Electr. Comput. Eng. 2022, 45, 192–198. [Google Scholar] [CrossRef]
  15. Mallick, M.; Tian, X.; Liu, J. Evaluation of measurement converted KF, EKF, UKF, CKF, and PF in GMTI filtering. In Proceedings of the 2021 International Conference on Control, Automation and Information Sciences (ICCAIS), Hanoi, Vietnam, 27–29 November 2021; pp. 21–27. [Google Scholar]
  16. Kumar, K.; Bhaumik, S.; Date, P. Extended Kalman Filter Using Orthogonal Polynomials. IEEE Access 2021, 9, 59675–59691. [Google Scholar] [CrossRef]
  17. Sun, X.; Wen, C.; Wen, T. Maximum correntropy high-order extended Kalman filter. Chin. J. Electron. 2022, 31, 190–198. [Google Scholar] [CrossRef]
Figure 1. The flow chart represents the relationship between the nonlinear system and the Kalman filter, and the updating process of the REKF.
Figure 1. The flow chart represents the relationship between the nonlinear system and the Kalman filter, and the updating process of the REKF.
Symmetry 16 00617 g001
Figure 2. Comparison of EKF with residual term and conventional EKF. (a) The X axis represents the number of tests, and the Y axis represents the state estimate for X1. (b) The X axis represents the number of tests, and the Y axis represents the state estimate for X2. (c) The X axis represents the number of tests, and the Y axis represents the root mean square error of X1. (d) The X axis represents the number of tests, and the Y axis represents the root mean square error of X2.
Figure 2. Comparison of EKF with residual term and conventional EKF. (a) The X axis represents the number of tests, and the Y axis represents the state estimate for X1. (b) The X axis represents the number of tests, and the Y axis represents the state estimate for X2. (c) The X axis represents the number of tests, and the Y axis represents the root mean square error of X1. (d) The X axis represents the number of tests, and the Y axis represents the root mean square error of X2.
Symmetry 16 00617 g002
Figure 3. Comparison of EKF with residual term and conventional EKF. (a) The X axis represents the number of tests, and the Y axis represents the state estimate for X1. (b) The X axis represents the number of tests, and the Y axis represents the state estimate for X2. (c) The X axis represents the number of tests, and the Y axis represents the root mean square error of X1. (d) The X axis represents the number of tests, and the Y axis represents the root mean square error of X2.
Figure 3. Comparison of EKF with residual term and conventional EKF. (a) The X axis represents the number of tests, and the Y axis represents the state estimate for X1. (b) The X axis represents the number of tests, and the Y axis represents the state estimate for X2. (c) The X axis represents the number of tests, and the Y axis represents the root mean square error of X1. (d) The X axis represents the number of tests, and the Y axis represents the root mean square error of X2.
Symmetry 16 00617 g003aSymmetry 16 00617 g003b
Table 1. X1 performance comparison.
Table 1. X1 performance comparison.
X1 Performance EnhancementValueEKFFirst-Order EKF
EKF root mean square error1.2922//
EKF first-order root mean square error1.27571.28%/
EKF higher-order root mean square error1.20286.92%5.62%
Table 2. X2 performance comparison.
Table 2. X2 performance comparison.
X2 Performance EnhancementValueEKFFirst-Order EKF
EKF root mean square error1.2723//
EKF first-order root mean square error1.25431.41%/
EKF higher-order root mean square error1.17977.28%5.95%
Table 3. Comparison of the performance of the EKF with the first-order EKF remainder substitution and the higher-order EKF remainder substitution for X1.
Table 3. Comparison of the performance of the EKF with the first-order EKF remainder substitution and the higher-order EKF remainder substitution for X1.
X1 PerformanceValueEKFFirst-Order EKF
EKF root mean square error0.2528//
EKF first-order root mean square error0.24762.06%/
EKF higher-order root mean square error0.24005.06%3.07%
Table 4. Comparison of the root mean square error of the first-order EKF residual replacement and the higher-order EKF residual replacement for X2.
Table 4. Comparison of the root mean square error of the first-order EKF residual replacement and the higher-order EKF residual replacement for X2.
X2 PerformanceValueEKFFirst-Order EKF
EKF root mean square error0.2384//
EKF first-order root mean square error0.23381.93%/
EKF higher-order root mean square error0.22246.71%4.88%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ding, L.; Wen, C. High-Order Extended Kalman Filter for State Estimation of Nonlinear Systems. Symmetry 2024, 16, 617. https://doi.org/10.3390/sym16050617

AMA Style

Ding L, Wen C. High-Order Extended Kalman Filter for State Estimation of Nonlinear Systems. Symmetry. 2024; 16(5):617. https://doi.org/10.3390/sym16050617

Chicago/Turabian Style

Ding, Linwang, and Chenglin Wen. 2024. "High-Order Extended Kalman Filter for State Estimation of Nonlinear Systems" Symmetry 16, no. 5: 617. https://doi.org/10.3390/sym16050617

APA Style

Ding, L., & Wen, C. (2024). High-Order Extended Kalman Filter for State Estimation of Nonlinear Systems. Symmetry, 16(5), 617. https://doi.org/10.3390/sym16050617

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop