Next Article in Journal
Transformative Technologies in Digital Agriculture: Leveraging Internet of Things, Remote Sensing, and Artificial Intelligence for Smart Crop Management
Previous Article in Journal
Beta Maximum Power Extraction Operation-Based Model Predictive Current Control for Linear Induction Motors
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Distributed Consensus Multi-Distribution Filter for Heavy-Tailed Noise

by
Guan-Nan Chang
1,2,
Wen-Xing Fu
1,
Tao Cui
3,
Ling-Yun Song
3 and
Peng Dong
3,*
1
Unmanned System Research Institute, Northwestern Polytechnical University, Xi’an 710072, China
2
Xi’an Modern Control Technology Research Institute, Xi’an 710065, China
3
School of Aeronautics and Astronautics, Shanghai Jiao Tong University, Shanghai 200240, China
*
Author to whom correspondence should be addressed.
J. Sens. Actuator Netw. 2024, 13(4), 38; https://doi.org/10.3390/jsan13040038
Submission received: 8 April 2024 / Revised: 26 June 2024 / Accepted: 26 June 2024 / Published: 28 June 2024

Abstract

:
Distributed state estimation is one of the critical technologies in the field of target tracking, where the process noise and measurement noise may have a heavy-tailed distribution. Traditionally, heavy-tailed distributions like the student-t distribution are employed, but our observation reveals that Gaussian noise predominates in many instances, with occasional outliers. This sporadic reliance on heavy-tailed distributions can degrade performances or necessitate frequent parameter adjustments. To overcome this, we introduce a novel distributed consensus multi-distribution state estimation method that combines Gaussian and student-t filters. Our approach establishes a system model using both Gaussian and student-t distributions. We derive a multi-distribution filter for a single sensor, assigning probabilities to Gaussian and student-t noise models. Parallel estimation under both distributions, utilizing Gaussian and student-t filters, allows us to calculate the likelihood of each distribution. The fusion of these results yields a mixed-state estimation and corresponding error matrix. Recognizing the increasing degrees of freedom in the student-t distribution over time, we provide an effective approximation. An information consensus strategy for multi-distribution filters is introduced, achieving global estimation through consensus on fused local filter results via interaction with neighboring nodes. This methodology is extended to the distributed case, and the recursive process of the distributed multi-distribution consensus state estimation method is presented. Simulation results demonstrate that the estimation accuracy of the proposed algorithm improved by at least 20% compared to that of the traditional algorithm in scenarios involving both Gaussian and heavy-tailed distributions.

1. Introduction

1.1. Background

Sensor networks have garnered significant attention in recent years in distributed applications, such as environmental monitoring, power grid systems and traffic management [1,2,3]. Distributed target tracking or state estimation is one of the classical applications, and the core of it is the use of local communication and measurement through information exchange to obtain the state estimation of the entire system. However, some interference may lead to heavy-tailed measurement noise, as shown in Figure 1, which could degrade the estimation performance. Thus, how to improve the estimation performance in distributed scenarios with heavy-tailed noise is a troublesome problem.

1.2. Related Work

There are three main types of distributed state estimation methods, including consensus methods, gossip methods and diffusion methods. For gossip-based method [4,5], each sensor selects one or several connected neighbors randomly to send information to. Although they have a low communication cost, gossip-based methods show slow convergence. The diffusion methods use a one-step convex combination of the received data [6,7], so they cannot converge to a consensus result. In contrast, the consensus method can ensure global convergence in many cases [8,9]; thus, it is the most widely used. Combining this method with Kalman filtering is the most direct method, and there are four main types: the consensus on estimation (CE) method, consensus on measurement (CM) method, consensus on information (CI) method and hybrid CM-CI (HCMCI) method [10,11]. Since the CE method does not consider the error covariance matrix, the estimation results are conservative [12]. In order to solve this problem, the CM method is proposed. However, the bounded estimation error of the CM method can only be achieved when consensus steps are large enough [8]. The CI method conforms the information vector and the information matrix, which can ensure the bounding of the estimation error of any number of consensus steps, while the new information is inevitably under-weighted [13,14,15]. The HCMCI method can handle both a priori and likelihood consensus, combining the advantages of the CM and CI methods. In addition, there are consensus algorithms [16] and consensus algorithms based on the Luenberger observer [17].
However, the above systems assume that the system noise obeys a Gaussian distribution, which is easy to be violated in practice. Due to unmodeled anomalies, sudden disturbances from the environment, unreliable sensors, target maneuvers, system failures or attacks on the system, the system will suffer from outliers [18]. In such scenarios, relying solely on the Gaussian noise assumption for state estimation leads to severe performance degradation and potential system divergence [19]. Outliers in distributions typically exhibit heavy-tailed characteristics, prompting the utilization of heavy-tailed distributions to accurately characterize system states and measurements [20]. The student-t distribution serves as an effective model for simulating heavy-tailed distributions, especially in handling non-Gaussian outliers, rendering it a widely adopted solution in this context [21]. In [22], the sensor measurement noise was modeled as a multivariable student-t distribution using the CI consensus strategy, where the state and noise parameters are estimated at the same time. Combined with the variational Bayesian (VB) method, the joint posterior distribution is processed.
However, the above method can only deal with the situation where the measurement noise follows a heavy-tailed distribution. In order to deal with the situation that both process noise and measurement noise follow a heavy-tailed distribution, ref. [23] established both process and measurement noises as student-t distributions and proposed a distributed state estimation method based on a student-t filter and CI consensus strategy. In practical real-world settings, noise predominantly adheres to a Gaussian distribution, with outliers occurring sporadically. Given this scenario, an intuitive approach involves modeling the system where a mixture of Gaussian and student-t distributions encapsulates the dynamic nature of the noise. This method capitalizes on the strengths inherent in both Gaussian and student-t distributions, aiming to attain precise and consistent state estimations. Grounded in this conceptual framework, a resilient filtering algorithm, utilizing a blend of Gaussian and student-t distributions, is introduced for single-sensor state estimation [24]. Nevertheless, the iterative calculations of the parameters pose a challenge, particularly in the context of distributed systems characterized by lightweight architectures and limited communication bandwidth.
We can find that many studies have focused on the robustness. However, in practical scenarios, heavy-tailed noise is a low-probability event, and the model noise can still present as a Gaussian distribution in most cases. Although the mentioned robust filter can provide a robust performance in the presence of outliers, it will lose estimation accuracy when the noise is normal. Therefore, how to balance the robustness and estimation accuracy as well as obtain consensus results are significant problems.
Motivated by these, this paper proposes a distributed consensus state estimation method based on the Gaussian distribution and the student-t distribution. A comparison of related studies is shown in Table 1.
The main contribution of this paper can be highlighted as follows:
(1)
The derivation of a method for local sensors that models process and measurement noises in both Gaussian and student-t distributions. Each distribution is assigned a probability, enabling parallel state estimation using Gaussian and student-t filters. The fusion of the results from these distributions provides a mixed-state estimation and corresponding error matrix.
(2)
Addressing the challenge of the increasing degrees of freedom in the student-t distribution over time by providing an effective approximate solution, ensuring stability and accuracy in the estimation process.
(3)
The introduction of an information consensus strategy for multi-distribution filters, enabling global estimation by achieving consensus on fused local filter results through interaction with neighboring nodes. The results are extended to the distributed case, and the recursive process presented further validates its efficacy, as supported by simulation results showcasing its performance in scenarios involving both Gaussian and heavy-tailed distributions.
The organization of this paper is shown in Figure 2. Section 2 presents the problem, Section 3 is the proposed distributed multiple-consensus-state estimation algorithm, the simulation results are presented in Section 4, and the conclusions are presented in Section 5.

2. Preliminaries and Problem Formulation

We consider a sensor network with node N and edge A . Then, the network can be represented as a two-tuple ( N , A ) . The set of neighbor nodes (including node i itself) of node i N is N i . The process equation of the system is assumed to be discrete-time linear:
x k = F k 1 x k 1 + w k 1
where x k means the n x -dimension state vector at time k, F k means the state transition matrix, and w k means the process noise at time k.
The measurement equation of each node i N is also linear:
z k i = H k i x k + v k i , i N
where z k i means the n z i -dimension measurement vector of node i N at time k, H k i means the measurement matrix of node i N , and v k i means the measurement noise of node i N at time k.
In general, the noises can be supposed to follow a Gaussian distribution. However, some unknown perturbations may lead to outliers characterized by heavy-tailed noise. In order to deal with the problem, many studies have used distributions that are insensitive to outliers to model the noise.
The student-t distribution is one of them. A random variable x obeys the student-t distribution if its probability density satisfies the following [20]:
St ( x ; m , P , η ) = M ( 1 + 1 η ( x m ) T P 1 ( x m ) ) η + n x 2
where
M = Γ ( η + n x 2 ) Γ ( η 2 ) 1 ( η π ) n x 2 1 det ( P )
m is the mean, η is the degrees of freedom (dof), P is a positive definite symmetric matrix, Γ denotes the Gamma function, and n x is the dimension of x. The variables of the system model are summarized in Table 2.
It should be noted that the variance in the student-t distribution is η η 2 P for dof η greater than 2; otherwise, the variance is undefined. When η = 1 , it converges to the Cauchy distribution, and while η tends to infinity, it becomes the Gaussian distribution. Compared with the Gaussian distribution, the student-t distribution shows heavy-tailed characteristics: when moving away from the mean, the density does not decrease as quickly as with the Gaussian distribution. Figure 3 presents logarithmic presentations of the Gaussian and student-t distribution. It can be seen from the figure that the value of the Gaussian distribution quickly drops below −16, while the value of the student-t distribution does not drop so fast.
Due to the long-tailed property, the student-t distribution is effective in dealing with heavy-tailed noise. However, it may not describe normal noise well. In practical scenarios, the Gaussian noise can adapt to most situations, while the heavy-tailed noise is a low-probability event. Thus, only student-t distribution molding can improve the robustness of the algorithm, but it will loss estimation accuracy.
The problem is how to balance and obtain the robustness and estimation accuracy of the consensus estimation results for sensor networks in the presence of possible heavy-tailed noise.

3. Main Results

The proposed multi-distribution filter, based on Gaussian and student-t distributions for a single local sensor operating without information interaction, is introduced, addressing inherent algorithm limitations. An approximate method is provided to mitigate these shortcomings. Subsequently, a CI-based consensus strategy is proposed for scenarios involving mixed Gaussian and student-t distributions. Finally, building upon the aforementioned consensus strategy and the single-sensor multi-distribution filter, an algorithm for distributed multi-distribution filtering is proposed.

3.1. The Multi-Distribution Filter Based on Gaussian Distribution and Student-t Distribution

In this subsection, we present a multi-distribution filter for a single sensor exposed to both a heavy-tailed process and measurement noises. The exclusive reliance on a student-t distribution filter often results in prolonged performance degradation or necessitates frequent parameter readjustments during normal system operations. Conversely, relying solely on Gaussian filters tends to cause divergence in scenarios where system outliers manifest. To harness the strengths of both filters, we present the following two hypotheses for a single sensor node (omitting the superscript representing the sensor node in this subsection):
H 0 : Suppose that the process and measurement noises obey the Gaussian distribution as follows:
p ( w k ) = N ( w k ; 0 , Q k 0 )
p ( v k ) = N ( v k ; 0 , R k 0 )
where N x ; m , P denotes that x obeys the Gaussian distribution with mean m and covariance P.
H 1 : Suppose that the process and measurement noises obey the student-t distribution as follows:
p ( w k ) = St ( w k ; 0 , Q k 1 , η k )
p ( v k ) = St ( v k ; 0 , R k 1 , η k )
where the probability of H 0 is μ k 0 , the probability of H 1 is μ k 1 , r H μ k r = 1 , and H = { 0 , 1 } .
Now, we need to assign a filter to each distribution. For the hypothesis of the Gaussian distribution, since the system model is linear, the standard Kalman filter can be used. The steps are as follows: given the initial values x ^ 0 0 and P ^ 0 0 , when time step k > 1 , the following recursive process is performed:
(1) Time update
x ^ k | k 1 0 = F k x ^ k 1 0
P k | k 1 0 = F k P k 1 0 F k T + Q k 1 0
(2) Measurement update
z ˜ k 0 = z k 0 H k x ^ k | k 1 0
S k 0 = H k P k | k 1 0 ( H k ) T + R k 0
K k 0 = P k | k 1 0 ( H k ) T / S k 0
x ^ k 0 = x ^ k | k 1 0 + K k 0 z ˜ k 0
P k 0 = P k | k 1 0 K k 0 S k 0 ( K k 0 ) T
For the hypothesis of the student-t distribution, we use the student-t filter described as follows: given the initial values x ^ 0 1 , P 0 1 and η 0 , when time step k > 1 , the following recursive process is performed:
(1) Time update
x ^ k | k 1 1 = F k x ^ k 1 1
P k | k 1 1 = F k P k 1 1 F k T + Q k 1 1
(2) Measurement update
z ˜ k 1 = z k 1 H k x ^ k | k 1 1
S k 1 = H k P k | k 1 1 ( H k ) T + R k 1
K k 1 = P k | k 1 1 ( H k ) T / S k 1
x ^ k 1 = x ^ k | k 1 1 + K k 1 z ˜ k 1
P k 1 = η k 1 + Δ z , k 1 η ˜ k 1 + d ( P k | k 1 1 K k 1 S k 1 ( K k 1 ) T )
Δ z , k 1 = ( z ˜ k 1 ) T ( S k 1 ) 1 z ˜ k i , 1
η k = η k 1 + n z
The detailed derivation of the student-t filter can be seen in [25].
In assuming that the state posterior obeys the mixed distribution of Gaussian and student-t distributions, their probabilities are μ k 0 and μ k 1 , respectively. According to the full probability theorem,
p ( x k | Z k ) = r H μ k r p ( x k | Z k , H r ) = μ k 0 N ( x k ; x ^ k 0 , P k 0 ) + μ k 1 St ( x k ; x ^ k 1 , P k 1 , η k )
where Z k = { z 1 , z 2 , , z k } represents the measurement set up to time k. In order to obtain the posterior probability density function (PDF), the required parameters are the probability μ k r corresponding to the two distributions, the state estimation x ^ k r and the matrix P k r . For the student-t distribution, the dof η k is also required. In addition to the distribution probability μ k r , the others are obtained by the two parallel filters. Given the distribution probability μ k 1 i of the previous time, the probability of assuming that H r is correct can be given by
μ k r = Δ p ( H r | Z k ) = p ( H r | z k , Z k 1 ) = p ( z k | Z k 1 , H r ) p ( H r | Z k 1 ) p ( z k | Z k 1 ) = p ( z k | Z k 1 , H r ) p ( H r | Z k 1 ) j H p ( z k | Z k 1 , H j ) p ( H j | Z k 1 ) = p ( z k | Z k 1 , H r ) μ k 1 r j H p ( z k | Z k 1 , H j ) μ k 1 j
where p ( z k | Z k 1 , H r ) is the measurement likelihood of H r at time k, that is, Λ k r = Δ p ( z k | Z k 1 , H r ) = p ( z ˜ k r ) . For the Gaussian filter,
Λ k 0 = 1 det ( 2 π S k 0 ) e 1 2 ( z ˜ k 0 ) T ( S k 0 ) 1 z ˜ k 0
For the student-t distribution filter,
Λ k 1 = Γ ( η k 1 + n z 2 ) Γ ( η k 1 2 ) 1 ( η k 1 π ) n z 2 1 det ( S k 1 ) ( 1 + 1 η k 1 ( z ˜ k 1 ) T ( S k 1 ) 1 z ˜ k 1 ) η k 1 + n z 2
The mean of the mixed posterior distribution is
x ^ k = E [ x k | Z k ] = x k [ μ k 0 N ( x k ; x ^ k 0 , P k 0 ) + μ k 1 St ( x k ; x ^ k 1 , P k 1 , η k ) ] d x k = μ k 0 x k N ( x k ; x ^ k 0 , P k 0 ) d x k + μ k 1 x k St ( x k ; x ^ k 1 , P k 1 , η k ) d x k = μ k 0 E [ x k | Z k , H 0 ] + μ k 1 E [ x k | Z k , H 1 ] = r H μ k r x ^ k r .
The covariance corresponding to the mixed posterior distribution is
P k = E [ ( x k x ^ k ) ( x k x ^ k ) T | Z k ] = ( x k x ^ k ) ( x k x ^ k ) T [ μ k 0 N ( x k ; x ^ k 0 , P k 0 ) + μ k 1 St ( x k ; x ^ k 1 , P k 1 , η k ) ] d x k = μ k 0 ( x k x ^ k ) ( x k x ^ k ) T N ( x k ; x ^ k 0 , P k 0 ) d x k + μ k 1 ( x k x ^ k ) ( x k x ^ k ) T St ( x k ; x ^ k 1 , P k 1 , η k ) d x k = r H μ k r E [ ( x k x ^ k ) ( x k x ^ k ) T | Z k , H r ] = r H μ k r E [ ( x k x ^ k r + x ^ k r x ^ k ) ( x k x ^ k r + x ^ k r x ^ k ) T | Z k , H r ] = r H μ k r E [ ( x k x ^ k r ) ( x k x ^ k r ) T | Z k , H r ] + r H μ k r ( x ^ k r x ^ k ) ( x ^ k r x ^ k ) T .
For the Gaussian distribution, the variance is
P k 0 = E [ ( x k x ^ k ) ( x k x ^ k ) T | Z k , H 0 ]
and for the student-t distribution, the variance is
P ¯ k 1 = E [ ( x k x ^ k ) ( x k x ^ k ) T | Z k , H 1 ] = η k η k 2 P k 1 .
Therefore, the expression of the covariance of the mixed posterior distribution can be obtained by
P k = μ k 0 P k 0 + μ k 1 P ¯ k 1 + r H μ k r ( x ^ k r x ^ k ) ( x ^ k r x ^ k ) T
Thus, we obtain a complete recursive step for the multi-distribution filter.
With each measurement update, the dof increase according to Equation (24). In turn, this also requires an increase in the dof of noise, making the problem more and more Gaussian. In fact, the algorithm will converge to the Kalman filter after several time steps. Therefore, it is necessary to find some approximate methods.
One of the simplest ways is to enforce η k = η in Equation (24), where η is a constant, so that the dof will not increase all the time. However, the actual posterior density is p ( x k ) = St ( x k ; x ^ k 1 , P k 1 , η + n z ) instead of St ( x k ; x ^ k 1 , P k 1 , η ) . Notice that we have omitted the condition here. At this point, we need to find a posterior density q ( x k ) = St ( x k ; x ^ k 1 , P ˜ k 1 , η ) to approximate p ( x k ) . The qualitative characteristics should be retained. Therefore, the adjusted matrix parameters should be the scaled version of the original matrix. The general expression of this problem is how to find the probability density q ( x ) = St ( x ; x ^ , c P , η ˜ ) to approximate the probability density p ( x ) = St ( x ; x ^ , c P , η ) so that the two are as close as possible. This density is controlled by the scaling factor c>0. Therefore, our problem is how to find c so that the two probability densities are close in a certain sense. A suggested method is to use the moment matching method, that is, to make the variances of the p ( x ) and q ( x ) equal. The advantages of this method are that it is simple to apply and does not need any parameters to be adjusted. We can obtain the following conditions:
η η 2 P = η ˜ η ˜ 2 c P
where η > 2 and η ˜ > 2 . Then, the scale factor c can be obtained:
c = η ( η ˜ 2 ) ( η 2 ) η ˜
The PDF of process and measurement noises can be approximated in the same way.

3.2. Consensus on Mixed Density

The basic idea of consensus is to calculate the aggregation of the whole network by iteratively computing the same type of region on each node in the network that only contains the subset of adjacent nodes. Consensus is used to average the PDF of states in each node and the PDF of states received from neighbors. Given PDFs p ( · ) and q ( · ) and weight π , define the following information fusion and weighting operations:
p ( x ) q ( x ) = p ( x ) q ( x ) p ( x ) q ( x ) d x
π p ( x ) = p ( x ) π p ( x ) π d x
The Kullback–Leibler average (KLA), which depends on relative entropy, is an average information definition of a PDF. The weighted KLA in PDF { p i ( · ) } is defined as
p ¯ = arg inf p ( · ) i N π i K L ( p | | p i )
where π i > 0 , and i N π i = 1 represents the relative weight. K L ( p | | p i ) is the Kullback–Leibler divergence (KLD) between p ( · ) and p i ( · ) . The problem of the consensus algorithm can be described as finding a way to make
lim l + p l i = p ¯ ( x ) , i N
where asymptotic PDF p ¯ ( · ) represents the KLA with the same weight. The solution to this problem is the collective KLA: the weighted KLA is given by the normalized weighted geometric mean of PDF
p ¯ ( x ) = i N p i ( x ) π i i N p i ( x ) π i d x = i N ( π i p i ( x ) )
It can be calculated by updating local data in a distributed manner using a convex combination with data from neighbors
p i ( x ) = j N i ( π i , j p 1 j ( x ) ) , i N
where is the number of consensus iteration steps, π i , j is the consensus weight satisfying π i , j 0 , and j N i π i , j = 1 , and it is also the ( i , j ) component of the consensus matrix Π (if j N i , π i , j = 0 ). In addition, the initial value of iteration is p 0 i ( · ) = p i ( · ) . Let π l i , j be the ( i , j ) component of matrix Π . Then,
p i ( x ) = j N ( π i , j p j ( x ) ) , i N .
When π i , j is chosen so that matrix Π is primitive and doubly stochastic,
lim + π i , j = 1 | N | , i N
Therefore, as the number of consensus steps increases, each local PDF tends to focus on the unweighted KLA.
For PDF p ( x i ) = N ( x ; m i , P i ) with the Gaussian distribution, it can be proved that the probability density consensus algorithm can be simplified to an algebraic expression involving only its information vector q i = ( P i ) 1 m i and information matrix Ω i = ( P i ) 1 :
Ω i = j N i π i , j Ω 1 j , i N , = 0 , 1 ,
q i = j N i π i , j q 1 j , i N , = 0 , 1 ,
This is the so-called CI consensus method.
In the above section, the multiple-distribution filter for a single sensor is given. To extend it to the distributed case, two important problems need to be solved: (1) for the mixed posterior distribution, which information should be transferred by the adjacent nodes and what strategy should be adopted to realize the collective KLA; and (2) since the mixed distribution is non-Gaussian, whether the CI consensus strategy for Gaussian distribution can be directly applied to the mixed distribution of Gaussian and student-t distributions.
For the first problem, the following consensus strategy is given: the distribution probability μ k r can be the first to obtain the consensus distribution probability, and then the fused PDF p i ( x ) = r = 0 1 μ k r p i , r ( x ) is given as the initial value, and then the consensus is run on the fused PDF. This strategy only needs to transfer the distribution probability and the fused PDF.
It should be noted that the above consensus method is for a continuous PDF. When the distribution probability is discrete, its distribution is a probability mass function (PMF). Given a PMF distribution μ i ( i N ), the weighted KLA is defined as
μ ¯ = arg inf μ ( · ) i N π i K L ( μ | | μ i )
For PMF μ , ν and weight π > 0 , the following information fusion and weighting operations are defined:
μ ν = η = c o l ( η j ) j M w i t h η j = μ j ν j h M μ h ν h = Δ μ j ν j , j M ,
π μ = η = c o l ( η j ) j M w i t h η j = ( μ j ) π h M ( μ h ) π = Δ π μ j , j M
Then, the KLA probability mass function can be expressed as
μ ¯ = i N ( π i μ i )
The collective fusion of the PMF can be obtained in a distributed manner:
μ i = j N i ( π i , j μ 1 j )
where π i , j > 0 is the consensus weight, and j N i π i , j = 1 .
Before answering the second question, let us first look at how to use a Gaussian distribution to approximate a student-t distribution. That is, how to find a scalar c to minimize the difference between q ( x ) = N ( x ; x ^ , c P ) and p ( x ) = St ( x ; x ^ , P , η ) under certain criteria. We know that when the dof of the student-t distribution tends to infinity, its distribution tends to the Gaussian distribution. Thus, we can obtain q ( x ) = N ( x ; x ^ , c P ) St ( x ; x ^ , c P , η ˜ ) . In this way, the problem becomes an approximation between two student-t distributions, so we can use the moment matching method to obtain the value of c, that is
c = lim η ˜ η ( η ˜ 2 ) ( η 2 ) η ˜ = η η 2
Therefore, q ( x ) = N ( x ; x ^ , η η 2 P ) can be used to approximate p ( x ) = St ( x ; x ^ , P , η ) . We can see that the fusion process can be seen as the fusion of two Gaussian distributions. Thus, we can directly use the CI consensus method based on the Gaussian distribution.

3.3. The Distributed Multi-Distribution Filter

We have previously obtained the algorithm for a single sensor and the consensus strategy for multi-sensor mixed density. Now, we need to extend the results to the distributed case.
For local node i, the initial values x ^ 0 i , r , P 0 i , r , and μ 0 r are given for r H . For the student-t filter, the common dof η is also given. When k > 0 , start the following recursive process.
Step 1 
Parallel filtering
For the Gaussian filter, we use Equations (9)–(15). For the student-t distribution filter, we make an approximation so that
η ˜ k 1 i = η , c k i = ( η + n z ) ( η 2 ) ( η + n z 2 ) η
P ˜ k 1 1 = c k i P k 1 1 , Q ˜ k 1 1 = c k i Q k 1 1 , R ˜ k 1 1 = c k i R k 1 1
Then, Equations (17), (19) and (24) become
P k | k 1 i , 1 = F k P ˜ k 1 i , 1 F k T + Q ˜ k 1 i , 1
S k i , 1 = H k P k | k 1 i , 1 ( H k ) T + R ˜ k i , 1
η k = η ˜ k 1 + n z
Therefore, we can use Equations (16), (54), (18), (55), (20)–(23) and (56) for the student-t distribution filter. After that, we can obtain x ^ k i , r , P k i , r , z ˜ k i , r , and S k i , r for r H and η k i .
Step 2 
Calculate distribution probability
Update probability
μ k , 0 i , r = Λ k i , r μ k 1 i , r j H Λ k i , j μ k 1 i , j
where the likelihoods of the Gaussian and student-t filter Λ k i , 0 and Λ k i , 1 can be obtained by Equations (27) and (28).
Step 3 
Consensus on distribution probability
For L-step consensus,
μ k , i , r = j N i ( π i , j μ k , 1 j , r ) , = 1 , , L r H
μ k i , r = μ k , L i , r
where = 1 , 2 , , L is the number of consensus steps, π i , j is the consensus weight, π i , j 0 , and j N i π i , j = 1 .
Step 4 
Fuse the mixed PDF
x ^ k , 0 i = r H μ k i , r x ^ k i , r
P ¯ k i , 1 = η k η k 2 P k i , 1
P k , 0 i = μ k i , 0 P k i , 0 + μ k i , 1 P ¯ k i , 1 + r H μ k i , r ( x ^ k i , r x ^ k , 0 i ) ( x ^ k i , r x ^ k , 0 i ) T
Step 5 
Consensus on fused PDF
Ω k , 0 i = ( P k , 0 i ) 1 , q k , 0 i = Ω k , 0 i x k , 0 i
For L-step consensus,
Ω k , i = j N i π i , j Ω k , 1 j , = 1 , , L
q k , i = j N i π i , j q k , 1 j , = 1 , , L
Step 6 
Reinitialization
x ^ k i , 0 = ( Ω k , L i ) 1 q k , L i , P k i , 0 = ( Ω k , L i ) 1
x ^ k i , 1 = ( Ω k , L i ) 1 q k , L i , P k i , 1 = η k 2 η k ( Ω k , L i ) 1
The workflow of the proposed algorithm is shown in Figure 4, and the pseudocode is summarized in Algorithm 1.
Algorithm 1: Distributed consensus multi-distribution filter (DCMDF)
Given the initial values x ^ 0 i , r , P 0 i , r and μ 0 r for r H and the common dof η , for each time step k at every node i, start the following recursive process:
Step1 Parallel filtering:
    Gaussian filtering: Calculate x ^ k 0 , i , P k 0 , i , S k 0 , i by Equations (9)–(15)
    Student-t filtering: Calculate x ^ k 1 , i , P k 1 , i , S k 1 , i by Equations (16)–(22) and (54), (55)
     η k = η ˜ k 1 + n z
Step2 Calculate distribution probability:
    Calculate μ k , 0 i , r according to (57)
Step3 Consensus on distribution probability:
    for = 1 to L
        Calculate μ k , i , r according to (58)
    end for
     μ k i , r = μ k , L i , r
Step4 Fuse the mixed PDF:
     Calculate x ^ k , 0 i and P k , 0 i according to (60)–(62)
Step5 Consensus on fused PDF:
     Ω k , 0 i = ( P k , 0 i ) 1 , q k , 0 i = Ω k , 0 i x k , 0 i
     for = 1 to L
        Calculate Ω k , i and q k , i according to (63) and (64)
     end for
Step6 Reinitialization:
     x ^ k i , 0 = ( Ω k , L i ) 1 q k , L i , P k i , 0 = ( Ω k , L i ) 1
     x ^ k i , 1 = ( Ω k , L i ) 1 q k , L i , P k i , 1 = η k 2 η k ( Ω k , L i ) 1
Return: x ^ k i , 0 , P k i , 0 , x ^ k i , 1 and P k i , 1

4. Numerical Simulation

Consider a linear dynamic system in which the state contains x = [ p x , p ˙ x , p y , p ˙ y ] T :
F k = 1 T 0 0 0 1 0 0 0 0 1 T 0 0 0 1 , Q k = G k Δ G k T
G k = T 2 / 2 T 0 0 0 0 T 2 / 2 T T
where the sampling time is T = 1s, Δ = diag ( [ w x 2 , w y 2 ] ) , and w x 2 = w y 2 = 0.1 .
There are 20 sensor nodes in the sensor network, and their graphical topology is shown in Figure 5. The measurement model is
H k i = 1 0 0 0 0 0 1 0
The standard measurement noise variance is R = diag ( [ ( 15 m ) 2 , ( 15 m ) 2 ] ) .
Assume that the initial true state is
x 0 = [ 2600 m , 20 m / s , 3800 m , 10 m / s ] T
In the simulation, the initial state of the filter is randomly selected from N ( x 0 , P 0 ) , where
P 0 = diag ( [ 50 2 m 2 , 5 2 m 2 / s 2 , 5 0 2 m 2 , 5 2 m 2 / s 2 ] )
Process noise and measurement noise with outliers are generated by the following model [20,23]:
w k N ( 0 , Q ) with probability 1 p o N ( 0 , 100 Q ) with probability p o
v k i N ( 0 , R ) with probability 1 p o N ( 0 , 100 R ) with probability p o
where p o is the probability of measuring outliers. It should be noted that this model for heavy-tailed noise is widely used.
In this section, the following three methods are compared: (1) distributed consensus Kalman Filter [14], referred to as DCKF; (2) distributed consensus student-t filter, proposed in [23], referred to as DCSTF; (3) the proposed algorithm, referred to as DCMDF; and (4) the multiple-model method with two Gaussian distributions, referred to as DCKFIMM.
The number of consensus steps was set to L = 3 , and the consensus weight of the sensor node was set to π i , j = 1 / | N i | if j N i and π i , j = 0 if j N i . The dof of the student-t distribution in DCSTF and DCMDF was η = 10 . One model of DCKFIMM is the same as the DCKF, and anther model has the following parameters: the measurement noise variance is 100R, and the process noise variance is 100Q. The simulation results were obtained through 100 Monte Carlo simulations and evaluated using the root mean square errors (RMSEs) of the position and velocity.
The occurrence probability of outliers was set to p o = 0 , which means that the noises are Gaussian. The RMSEs of the position and velocity of the four algorithms are shown in Figure 6 and Figure 7. It can be found that under the condition of a Gaussian distribution, the performance of the DCMDF algorithm is almost the same as that of DCKF, while the error of DCSTF is relatively large. The performance of the DCKFIMM algorithm is better than the DCSTF and little worse than the DCMDF. This result shows that Gaussian models are better than the single student-t model in dealing with estimation problems in normal noise environments. At the same time, it is worth noting that DCMDF adopts a mixed model of Gaussian distribution and student-t distribution; therefore, it can obtain a higher estimation accuracy through model adaptation.
The occurrence probability of outliers was set to p o = 0.3 . Figure 8 presents the true and estimated trajectories for one simulation. Figure 9 and Figure 10 show the RMSEs of the position and velocity of the three algorithms. It can be found that DCKF has the worst performance, which is caused by the natural defects of the Gaussian distribution. Although it can achieve good performance in a normal noise environment, it struggles to dealing with the estimation in the presence of outliers. The DCKFIMM has a better performance than DCKF; however, the performance of the DCKDIMM is worse than those of DCSTF and DCMDF because DCSTF and DCMDF adopt the student-t distribution, which is more in line with the distribution of noise while outliers occur. It is worth noting that DCMDF has the best performance because the algorithm uses a hybrid distribution model, which can be adaptive for the model in the estimation process, making the algorithm more robust and accurate.
Table 3 and Table 4 show the simulation results of the three algorithms under different outlier probabilities. It can be found that the RMSEs of the four algorithms will increase with the increase in outlier probabilities. However, the estimation performance of DCMDF always shows a relatively better estimation performance, and the estimation accuracy of the proposed algorithm improved by at least 20% compared to the other algorithms when p 0 0 , which further proves that DCMDF has a good performance in robustness and estimation accuracy.
To further test the performance of the algorithm, we reduced the number of nodes and density of network. The topology of a sensor network with 15 nodes is shown in Figure 11. Since the nodes and density of network were reduced, we increased the number of consensus steps to 6, the RMSEs of the position and velocity are shown in Table 5 and Table 6. It can be seen that the proposed algorithm still has better performances in estimation accuracy.
In order to handle the estimation problem with heavy-tailed noise, in this study, we made two hypotheses for the model noise, which include the Gaussian distribution and student-t distribution. The reason why the student-t distribution was chosen to represent the heavy-tailed noise distribution is discussed in detail in [23], which is briefly reviewed here. By adjusting the dof, the student-t distribution can tend toward a Gaussian distribution and Cauchy distribution, which have good flexibility. Similar to the traditional multiple-model algorithm, we calculate the state estimation results based on two different hypothetical noises. Then, the two results are fused based on different weights. Therefore, the proposed algorithm can balance the robustness and estimation accuracy.
The computation and communication burden are important issues in distributed-state estimation. Compared with the traditional CI algorithm, the computation of the proposed algorithm is mainly in the step of local estimation. The amount of computation has doubled, but the strategy of parallel computing can be adopted. The other step is to calculate the distribution probability of consensus, which has a small increase in calculation. The increase in communication also comes from this step. It should be noted that the quantities that we are concerned with are the state and its error matrix. This step can also be omitted in the application. Then, the increases in computation and communication from the distribution probability calculation and consensus will be eliminated, which can further improve the computational efficiency and reduce the communication burden.

5. Conclusions and Outlook

For the problem of distributed state estimation in which both process noise and measurement noise are heavy-tailed, a distributed consistent multi-distribution state estimation algorithm based on the Gaussian distribution and student-t distribution parallel filter is proposed in this paper, with the following steps:
(1)
The system model is established as two models based on the Gaussian distribution and student-t distribution, and each model is assigned a filter based on the distribution. An algorithm based on single-sensor observation is derived, and a combined posterior distribution based on a mixed posterior distribution is presented. To solve the problem of the filter converging to the Gaussian distribution quickly, the moment matching method is used to approximate the filter to keep the heavy-tailed characteristics.
(3)
Aiming at the consensus problem of mixed probability density, a mixed strategy based on CI consensus is proposed: firstly, the consensus on the discrete distributed probability PMF is carried out to obtain the KLA PMF; then, the posterior of the multi-distributions is combined using the probability of the distribution to obtain the posterior of the combination; finally, the CI method is used to obtain the consensus of the posterior of the combination. Based on the above hybrid probability consensus strategy and single-sensor algorithm, the recursive steps of distributed consensus multi-distribution filtering are presented.
(3)
The simulation results show that the proposed algorithm can achieve good results in both Gaussian noise and heavy-tailed noise scenarios.
Although the method proposed this paper deals with state estimation with heavy-tailed noise, it is still a type of rough processing method. The algorithm can ensure a low-bound estimation accuracy, but there is still potential for improvement.
For example, in target tracking, the process transfer model is time-variant. The two types of noise hypotheses in this paper can model the uncertainty of a process transfer, but this modeling method is still not precise enough. In a broad sense, the algorithm proposed in this paper can be regarded as a special case of the multi-model method. Therefore, the concepts of probability transfer and input interaction of the multi-model method can also be applied to the algorithm proposed in this paper. How to integrate these technologies is one of the problems to be solved in the future.

Author Contributions

Conceptualization, G.-N.C. and W.-X.F.; methodology, G.-N.C. and P.D.; software, Guan-nan Chan, T.C. and L.-Y.S.; validation, G.-N.C., W.-X.F. and T.C.; formal analysis, G.-N.C. and W.-X.F.; investigation, W.-X.F.; resources, P.D.; data curation, W.-X.F.; writing—original draft preparation, G.-N.C. and L.-Y.S.; writing—review and editing, G.-N.C. and T.C.; visualization, G.-N.C. and L.-Y.S.; supervision, P.D.; project administration, W.-X.F.; funding acquisition, P.D. All authors have read and agreed to the published version of the manuscript.

Funding

This work was jointly supported by the National Natural Science Foundation of China (No. 61803260).

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors on request.

Conflicts of Interest

The authors declare no conflicts of interest.

Acronyms and Nomenclatures

NotationsDefinitionsNotationsDefinitions
CEconsensus on estimationCMconsensus on measurement
CIconsensus on informationHCMCIhybrid CM-CI
dofdegrees of freedomPDFprobability density function
KLAKullback–Leibler averageKLDKullback–Leibler divergence
DCKFdistributed consensus Kalman FilterDCSTFdistributed consensus student-t filter
DCMDFdistributed consensus multiple-distribution filterDCKFIMMdistributed multiple-model method with two Gaussian distributions
RMSEroot mean square errorPMFprobability mass function

References

  1. Ding, D.; Han, Q.L.; Wang, Z.; Ge, X. A survey on model-based distributed control and filtering for industrial cyber-physical systems. IEEE Trans. Ind. Inform. 2019, 15, 2483–2499. [Google Scholar] [CrossRef]
  2. He, S.; Shin, H.; Xu, S.; Tsourdos, A. Distributed estimation over a low-cost sensor network: A review of state-of-the-art. Inform. Fusion 2020, 54, 21–43. [Google Scholar] [CrossRef]
  3. Huang, S.; Li, Y.; Wu, J. Distributed state estimation for linear time-invariant dynamical systems: A review of theories and algorithms. Chin. J. Aeronaut. 2022, 35, 1–17. [Google Scholar] [CrossRef]
  4. Sirocchi, C.; Bogliolo, A. Community-Based Gossip Algorithm for Distributed Averaging. In Proceedings of the IFIP International Conference on Distributed Applications and Interoperable Systems, Lisbon, Portugal, 19–23 June 2023. [Google Scholar]
  5. Loizou, N.; Richtarik, P. Revisiting randomized gossip algorithms: General framework, convergence rates and novel block and accelerated protocols. IEEE Trans. Ind. Inform. 2019, 15, 2483–2499. [Google Scholar] [CrossRef]
  6. Cai, P.; Wang, S.; Chen, Y. Diffusion Mixture Minimum Total Error Entropy Adaptive Filtering Algorithm and Its Performance Analysis. IEEE Trans. Signal Inf. Process. Over Netw. 2023, 9, 397–411. [Google Scholar] [CrossRef]
  7. Liang, Y.; Li, Y.; Chen, Y.; Sheng, A. Event-triggered diffusion nonlinear estimation for sensor networks with unknown cross-correlations. Syst. Control Lett. 2023, 175, 105506. [Google Scholar] [CrossRef]
  8. Olfati-Saber, R. Distributed Kalman filter with embedded consensus filters. In Proceedings of the 44th IEEE Conference on Decision and Control, Seville, Spain, 15 December 2005. [Google Scholar]
  9. Olfati-Saber, R.; Murray, R.M. Consensus problems in networks of agents with switching topology and time-delays. IEEE Trans. Autom. Control 2004, 49, 1520–1533. [Google Scholar] [CrossRef]
  10. Battistelli, G.; Chisci, L.; Fantacci, C. Parallel consensus on likelihoods and priors for networked nonlinear filtering. IEEE Signal Process. Lett. 2014, 21, 787–791. [Google Scholar] [CrossRef]
  11. Battistelli, G.; Chisci, L.; Mugnai, G.; Farina, A.; Graziano, A. Consensus-based linear and nonlinear filtering. IEEE Trans. Autom. Control 2015, 60, 1410–1415. [Google Scholar] [CrossRef]
  12. Olfati-Saber, R. Distributed Kalman filtering for sensor networks. In Proceedings of the 2007 46th IEEE Conference on Decision and Control, New Orleans, LA, USA, 12–14 December 2007. [Google Scholar]
  13. Battistelli, G.; Chisci, L.; Morrocchi, S.; Papi, F. An information-theoretic approach to distributed state estimation. IFAC Proc. Vol. 2011, 44, 12477–12482. [Google Scholar] [CrossRef]
  14. Battistelli, G.; Chisci, L. Kullback–Leibler average, consensus on probability densities, and distributed state estimation with guaranteed stability. Automatica 2014, 50, 707–718. [Google Scholar] [CrossRef]
  15. He, X.; Xue, W.; Fang, H. Consistent distributed state estimation with global observability over sensor network. Automatica 2018, 92, 162–172. [Google Scholar] [CrossRef]
  16. Qu, H.; Yang, F.; Han, Q.; Zhang, Y. Distributed H∞ consensus filtering for attitude tracking using ground-based radars. IEEE Trans. Cybern. 2021, 51, 3767–3778. [Google Scholar] [CrossRef] [PubMed]
  17. Rego, F.C.; Pu, Y.; Alessandretti, A.; Aguiar, A.P.; Pascoal, A.M.; Jones, C.N. A distributed Luenberger observer for linear state feedback systems with quantized and rate-limited communications. IEEE Trans. Autom. Control 2021, 66, 3922–3937. [Google Scholar] [CrossRef]
  18. Ting, J.A.; Theodorou, E.; Schaal, S. A Kalman filter for robust outlier detection. In Proceedings of the 2007 IEEE/RSJ International Conference on Intelligent Robots and Systems, San Diego, CA, USA, 29 October–2 November 2007. [Google Scholar]
  19. Fu, H.; Cheng, Y. A Novel Robust Kalman Filter Based on Switching Gaussian-Heavy-Tailed Distribution. IEEE Trans. Circuits Syst. II Express Briefs 2022, 69, 3012–3016. [Google Scholar] [CrossRef]
  20. Huang, Y.; Zhang, Y.; Li, N.; Wu, Z.; Chambers, J.A. A novel robust student’s t based Kalman filter. IEEE Trans. Aerosp. Electron. Syst. 2019, 55, 1545–1554. [Google Scholar]
  21. Huang, Y.; Zhang, Y.; Xu, B.; Wu, Z.; Chambers, J. A New Outlier-Robust Student’s t Based Gaussian Approximate Filter for Cooperative Localization. IEEE-ASME Trans. Mechatron. 2017, 22, 2380–2386. [Google Scholar] [CrossRef]
  22. Dong, P.; Jing, Z.; Shen, K.; Li, M. A distributed consensus filter for sensor networks with heavy-tailed measurement noise. Sci. China Inform. Sci. 2018, 61, 119201. [Google Scholar] [CrossRef]
  23. Wang, J.; Dong, P.; Shen, K.; Song, X.; Wang, X. Distributed consensus student-t filter for sensor networks with heavy-tailed process and measurement noises. IEEE Access 2020, 8, 167865–167874. [Google Scholar] [CrossRef]
  24. Huang, Y.; Zhang, Y.; Zhao, Y.; Chambers, J.A. A novel robust Gaussian–student’s t mixture distribution based Kalman filter. IEEE Trans. Signal Process. 2019, 67, 3606–3620. [Google Scholar] [CrossRef]
  25. Roth, M.; Ozkan, E.; Gustafsson, F. A student’s t filter for heavy tailed process and measurement noise. In Proceedings of the 2013 IEEE International Conference on Acoustics, Speech and Signal Processing, Vancouver, BC, Canada, 26–31 May 2013. [Google Scholar]
Figure 1. Research background and problem.
Figure 1. Research background and problem.
Jsan 13 00038 g001
Figure 2. The organization of this paper.
Figure 2. The organization of this paper.
Jsan 13 00038 g002
Figure 3. Logarithmic heat maps of Gaussian distribution (left) and student-t distribution (right).
Figure 3. Logarithmic heat maps of Gaussian distribution (left) and student-t distribution (right).
Jsan 13 00038 g003
Figure 4. The workflow of the proposed algorithm.
Figure 4. The workflow of the proposed algorithm.
Jsan 13 00038 g004
Figure 5. Topology of sensor network.
Figure 5. Topology of sensor network.
Jsan 13 00038 g005
Figure 6. Position root mean square errors of different algorithms.
Figure 6. Position root mean square errors of different algorithms.
Jsan 13 00038 g006
Figure 7. Velocity root mean square errors of different algorithms.
Figure 7. Velocity root mean square errors of different algorithms.
Jsan 13 00038 g007
Figure 8. The true and estimated trajectories for one simulation.
Figure 8. The true and estimated trajectories for one simulation.
Jsan 13 00038 g008
Figure 9. Position root mean square errors of different algorithms.
Figure 9. Position root mean square errors of different algorithms.
Jsan 13 00038 g009
Figure 10. Velocity root mean square errors of different algorithms.
Figure 10. Velocity root mean square errors of different algorithms.
Jsan 13 00038 g010
Figure 11. Topology of sensor network with 15 nodes.
Figure 11. Topology of sensor network with 15 nodes.
Jsan 13 00038 g011
Table 1. Comparison with related studies.
Table 1. Comparison with related studies.
ReferenceConsensueGaussian DistributionHeavy-Tailed Distribution
[10,11,12,16,17]
[19,20,21]
[22,23]
[24]
the proposed algorithm
Table 2. Variables of system model.
Table 2. Variables of system model.
VariablesMeaningVariablesMeaning
xStatePCovariance
zMeasurement η dof
mMean value Γ Gamma function
Table 3. Root mean square errors of position under different outlier probabilities.
Table 3. Root mean square errors of position under different outlier probabilities.
p 0 DCKFDCSTFDCMDFDCKFIMM
03.83395.51943.78994.4447
0.111.49918.95497.04059.0710
0.215.39509.43738.168511.5212
0.318.246611.48509.681814.0321
0.420.929514.826212.202018.3049
Table 4. Root mean square errors of velocity under different outlier probabilities.
Table 4. Root mean square errors of velocity under different outlier probabilities.
p 0 DCKFDCSTFDCMDFDCKFIMM
00.94051.06020.93761.0908
0.12.43782.28422.16372.2934
0.23.09312.61952.55652.3951
0.33.50502.91432.77513.0377
0.43.91303.31943.039843.4926
Table 5. Root mean square errors of position under different outlier probabilities.
Table 5. Root mean square errors of position under different outlier probabilities.
p 0 DCKFDCSTFDCMDFDCKFIMM
03.73624.73153.70214.1563
0.111.54208.49576.99219.8612
0.215.26419.59248.081311.6715
0.318.853611.40429.762415.6121
0.419.791213.812611.462017.9736
Table 6. Root mean square errors of velocity under different outlier probabilities.
Table 6. Root mean square errors of velocity under different outlier probabilities.
p 0 DCKFDCSTFDCMDFDCKFIMM
00.76230.83250.75460.8216
0.11.98201.77412.08311.8742
0.23.03622.61352.52722.6718
0.33.49312.86352.74523.1803
0.43.88423.31463.06533.6943
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chang, G.-N.; Fu, W.-X.; Cui, T.; Song, L.-Y.; Dong, P. Distributed Consensus Multi-Distribution Filter for Heavy-Tailed Noise. J. Sens. Actuator Netw. 2024, 13, 38. https://doi.org/10.3390/jsan13040038

AMA Style

Chang G-N, Fu W-X, Cui T, Song L-Y, Dong P. Distributed Consensus Multi-Distribution Filter for Heavy-Tailed Noise. Journal of Sensor and Actuator Networks. 2024; 13(4):38. https://doi.org/10.3390/jsan13040038

Chicago/Turabian Style

Chang, Guan-Nan, Wen-Xing Fu, Tao Cui, Ling-Yun Song, and Peng Dong. 2024. "Distributed Consensus Multi-Distribution Filter for Heavy-Tailed Noise" Journal of Sensor and Actuator Networks 13, no. 4: 38. https://doi.org/10.3390/jsan13040038

APA Style

Chang, G. -N., Fu, W. -X., Cui, T., Song, L. -Y., & Dong, P. (2024). Distributed Consensus Multi-Distribution Filter for Heavy-Tailed Noise. Journal of Sensor and Actuator Networks, 13(4), 38. https://doi.org/10.3390/jsan13040038

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop