Next Article in Journal
Mobile Contactless Fingerprint Recognition: Implementation, Performance and Usability Aspects
Previous Article in Journal
Impact Damage Detection Using Chirp Ultrasonic Guided Waves for Development of Health Monitoring System for CFRP Mobility Structures
Previous Article in Special Issue
A Novel Clutter Suppression Method Based on Sparse Bayesian Learning for Airborne Passive Bistatic Radar with Contaminated Reference Signal
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Performance Evaluation Metrics and Approaches for Target Tracking: A Survey

1
Key Laboratory of Information Fusion Technology, Northwestern Polytechnical University, Xi’an 710072, China
2
Key Laboratory of Science and Technology on ATR, National University of Defense Technology, Changsha 410073, China
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(3), 793; https://doi.org/10.3390/s22030793
Submission received: 5 December 2021 / Revised: 15 January 2022 / Accepted: 18 January 2022 / Published: 20 January 2022
(This article belongs to the Special Issue Advanced Sensing Technologies in Automation and Computer Sciences)

Abstract

:
Performance evaluation (PE) plays a key role in the design and validation of any target-tracking algorithms. In fact, it is often closely related to the definition and derivation of the optimality/suboptimality of an algorithm such as that all minimum mean-squared error estimators are based on the minimization of the mean-squared error of the estimation. In this paper, we review both classic and emerging novel PE metrics and approaches in the context of estimation and target tracking. First, we briefly review the evaluation metrics commonly used for target tracking, which are classified into three groups corresponding to the most important three factors of the tracking algorithm, namely correctness, timeliness, and accuracy. Then, comprehensive evaluation (CE) approaches such as cloud barycenter evaluation, fuzzy CE, and grey clustering are reviewed. Finally, we demonstrate the use of these PE metrics and CE approaches in representative target tracking scenarios.

1. Introduction

Target tracking is widely involved in many problems of significance, such as military defense, automation/driverless transportation, intelligent robots, and so on. There are many outstanding works to provide guidance on the implementation of target-tracking algorithms. For the issue of multitarget tracking, the International Society of Information Fusion (ISIF) took track estimation, data association, and performance evaluation into account many years ago. The textbook [1], which is a tutorial of the known target-tracking algorithms and the class material [2], can assist researchers in re-implementing these methods and developing advanced methods. Some common toolboxes include the Recursive Bayesian Estimation Library (ReBEL) [3,4,5], the nonlinear estimation framework [6,7,8], and the Tracker Component Library [9]. The Open Source Tracking and Estimation Working Group (OSTEWG), which is the working group of the ISIF, revisited the current widely used methods to make up an open-source framework in [10,11], which is named Stone Soup and is available from the website https://github.com/dstl/Stone-Soup/ (accessed on 10 November 2021). Another ISIF working group, the Evaluation of Techniques of Uncertainty Reasoning Working Group (ETURWG), develops the uncertainty description in the domain of target tracking and introduced the Uncertainty Representation and Reasoning Evaluation Framework (URREF) [12] to Stone Soup.
This paper focuses on the performance evaluation (PE) of target-tracking algorithms, which plays a key role in the comparison of existing algorithms and in putting forward a new algorithm. PE is referred to the assessment and the evaluation of various performance metrics of a system [13,14], whose significance lies in providing evaluation results of the system performance [15], as well as the reference basis for the optimization of the system performance.
The basic process required for tracking evaluation is that when both truth targets and tracks are available, the first step is to find an association between the true target and a track so that performance measures can be computed. However, it was assumed that there is a unique association in this paper. The association algorithms can be found in [16,17,18,19]. After the tracks were assigned to targets, we computed the various performance measures to analyze the target-tracking algorithms and optimize these algorithms.
For the practitioner, the PE problems can be divided into two stages. The first is to choose the relevant effective metrics, and the second is to evaluate a single score through these metrics [20]. In this paper, we reviewed the measures to evaluate the performance of the target tracking system. These metrics were grouped into the categories of the correctness, timeliness, and accuracy. The assessed results of each metric were weighted and combined to give an overall performance measure. Further, we designed simulations that employed several PE approaches based on these metrics to illustrate their use in the target-tracking problem.
The rest of this paper is organized as follows. Section 2 introduces general evaluation metrics and sorts categories by the characteristics of each metric. Section 3 reviews classic PE approaches. The simulation results of our approach in the context of target tracking are given in Section 4. Section 5 concludes our work and faces remaining challenges.

2. A Classification of the Comprehensive Evaluation Metrics

In the context of target tracking, a variety of evaluation metrics with physical significance have been proposed, which can evaluate the practicability of the tracking algorithm and the consistency of the expected and assessed results. These metrics can be divided into three categories: effectiveness, timeliness, and accuracy, which can be seen in [21,22,23,24]. This paper also followed this division criterion for convenience. What the correctness [18,25,26,27] usually represents is the number of missed/false targets, etc. The timeliness assesses the time performance of the estimated track [23,28], which is a crucial measure for online target tracking. The accuracy metrics can be defined in different ways according to the different scenario requirements, in which the (root-) mean-squared error ((R)MSE) is commonly used in trajectory error (TE), tracking position error (TPE), and tracking velocity error (TVE) [29], and the other accuracy metrics refer to [28]. References [30,31] combined the multiple-object-tracking precision (MOPT) and the multiple-object-tracking accuracy (MOTA) to describe the effectiveness and the timeliness of the multiple-object-tracking systems, but they disregarded the effect of the error. In addition, the measures, such as the cross-platform commonality, the track purity, the processor loading, etc., were considered in [7,26,32,33,34,35,36]. The comprehensive evaluation (CE) metric system is shown in Figure 1.

2.1. Correctness Measures

The correctness measures the numerical characteristics of the acquired data and calculates how many mistakes the tracker made in terms of misses, false tracks, and so forth, which can be briefly described by Figure 2, where the small and large circles represent the truth target and the estimated track, respectively. The solid and dashed lines denote the trajectory of the target and the curve of the track, respectively. Given a time interval t t 1 , t 2 , the correspondence between the target and the track is established in Figure 2. These measures are explained in detail below.
  • Number of valid tracks (NVT):
    If a track is assigned to a target and the target has only one track, then the track is validated. N v a l i d ( t ) denotes the NVT at time t;
  • Number of missed targets (NMT):
    A target is missed if it is not associated with any track. N m i s s e d ( t ) is the NMT at time t [26];
  • Number of false tracks (NFT):
    A track is false if it is not assigned to any target. N f a l s e ( t ) denotes the NFT at time t;
  • Number of spurious tracks (NST):
    A track is defined as spurious if it is assigned to more than one target, and the NST is denoted as N s p u r ( t ) ;
  • Average number of swaps in tracks (ANST):
    Different confirmed tracks may be assigned to a particular truth. This can happen when crossing targets and targets come close to each other. Assume N s w a p ( t ) as the ANST at time t [26];
  • Average number of broken tracks (ANBT):
    There is also probable that no track is assigned to the truth for several time steps. If there is no assigned track to the truth, the number of broken tracks is counted at each time step. Reference [26] employed the ANBT to check the track segment associated with the truth;
  • Tracks redundancy (TR):
    TR is represented as the ratio of validated tracks and total assigned tracks:
    T R ( t ) = N v a l i d ( t ) N v a l i d ( t ) + N s p u r ( t ) .

2.2. Timeliness Measures

The performance measure provides more information about the track persistence, which is also an indispensable part of the evaluation metrics [37]. Some timeliness metrics for the PE are given as follows:
  • Rate of false alarms (RFA):
    The RFA [38] is defined as the NFT per time step, which can be denoted as follows:
    F L R ( t ) = N f a l s e ( t ) t .
  • Track probability of detection (TPD):
    In the time interval [ t 1 , t 2 ] , assume t f i r s t i and t l a s t i as the first and last time that the ith target is present, respectively. According to [39], the TPD of each target is represented as:
    P d i = t t l a s t i t f i r s t i ,
    where t denotes the persistent duration where the ith target is assigned to a valid track;
  • Rate of track fragmentation (RTF):
    It is likely that the track obtained through some tracking algorithms may not be continuous sometimes. The track segment is assigned to the ith truth, the number of changes that the continuous track becomes fragmental is defined as T F R i when the track segment is assigned to the ith truth. The smaller the RTF is, the more persistent the tracking estimated by the algorithm is [39];
  • Track latency (TL):
    The TL, the delay from the moment that the target arises in the view of the sensor to the moment that target is detected by the tracker in the running period, is a measure of the track timeliness;
  • Total execution time (TET):
    The computational cost is another important factor to be considered in the PE of target tracking. Therefore, the total time that is taken to run the tracker is expressed as the TET for each tracking algorithm.

2.3. Accuracy Measures

The measure, favored by the majority of researchers, is a primary choice in evaluating the target tracking, in which several measures can be defined as based on the type of distance between the set of truths and tracks:
  • RMSE:
    The RMSE is defined in terms of the estimation error e k , which is the average difference between the estimated state X ˜ k and the truth state X k , as:
    R l ( k ) = 1 n k = 1 n e k 2 ,
    where n denotes the number of the lth targets detected at the tth time step, e k 2 = e k T e k . Th MSE/RMSE has long been the dominant quantitative performance metric in the field of signal processing. For the traditional target-tracking algorithms, the aim is to minimize them between the target truth and the estimated track [40], which is not suitable for the track assignments that do not have a one-to-one correspondence. At present, the CE metrics have been widely used in the Hausdorff distance [41], the Wasserstein distance [42,43], and the optimal subpattern assignment (OSPA) distance [20,44,45,46];
  • Hausdorff distance:
    The Hausdorff distance is a common method of measuring the distance between two sets of objects, which can be used to measure the similarity between tracks and is given by:
    d H ( X , X ˜ ) = Δ m a x { max x X min x ˜ X ˜ d ( x , x ˜ ) , max x ˜ X ˜ min x X d ( x , x ˜ ) } ,
    where X = { x 1 , x 2 , , x k } , X ˜ = { x ˜ 1 , x ˜ 2 , , x ˜ k } . x i and x ˜ i are the ith target and the ith track, respectively. d ( x , x ˜ ) is the Euclidean distance between x and x ˜ . It has been proven that the Hausdorff distance is very useful in assessing the multitarget data fusion algorithms [47,48,49]. Meanwhile, the distance is relatively insensitive to differences in the numbers of objects;
  • Wasserstein distance:
    The Wasserstein metric was initially used to measure the similarity of probability distributions [50] and was proposed for the sets of targets in [51]. The Wasserstein distance between X and X ˜ is:
    d W ( X , X ˜ ) = min π Π n 1 n i = 1 n x i x ^ π ( i ) 2 2 ,
    where X = { x 1 , x 2 , , x m } , X ˜ = { x ˜ 1 , x ˜ 2 , , x ˜ n } . Π n is the set of all permutations of { 1 , 2 , , n } . π Π n can be written as π = ( π ( 1 ) , π ( 2 ) , , π ( n ) ) . x i x ^ π ( i ) 2 2 = ( x i x ^ π ( i ) ) T ( x i x ^ π ( i ) ) . The Wasserstein distance extends and provides a rigorous theoretical basis for a natural multitarget miss distance. However, it lacks a physically consistent interpretation when the sets have different cardinalities [52];
  • OSPA distance:
    The OSPA was proposed to overcome the insensitive shortcoming, whose parameters can deal with the problem that the numbers of elements in the two sets do not match [53]. The OSPA metric between X and X ˜ is:
    d O S P A ( X , X ˜ ) = Δ 1 n min π Π n i = 1 m d ( c ) ( x i , x ˜ π ( i ) ) p + c p ( n m ) 1 p ,
    where d c ( x , x ˜ ) is the cut-off distance between x and x ˜ and d c ( x , x ˜ ) = min d ( x , x ˜ ) , c , c denotes the truncation parameter, and p is the OSPA metric order parameter. The choice of parameters was given in [53]. The OSPA distance has been used widely in the literature [44,54,55,56] and has better properties for the multitarget error evaluation than the Hausdorff metric.
    In addition, there are various improved methods based on the OSPA [57], which are be enumerated as follows:
  • Generalized OSPA (GOSPA):
    In the GOSPA metric, we look for an optimal assignment between the truth targets and the estimated tracks, leaving missed and false targets unsigned [58]. The GOSPA metric penalizes localization errors for properly detected targets, the NMT, and the NFT [59]. The GOSPA can be represented as an optimization over assignment sets:
    d p c , 2 X , X ˜ = min γ Γ i , j γ d p x i , x ˜ j + c p α n + m 2 γ 1 p ,
    where γ 1 , 2 , , n × 1 , 2 , , m , ( i , j ) , ( i , j ) γ j = j , and ( i , j ) , ( i , j ) γ i = i . Γ denotes the set of all possible assignment sets γ . α is the additional parameter to control the cardinality mismatch penalty; in general, α =2. The terms c p α ( m γ ) and c p α ( n γ ) represent the costs (to the pth power) for the NMT and the NFT, respectively;
  • OSPA-on-OSPA (OSPA) 2 metric:
    The OSPA 2 metric [60,61] is the distance between two sets of tracks, which establishes an assignment between the real and the estimated trajectories that is not allowed to change with time and enables capturing the tracking errors of fragmentation and track switching. It is also simple to compute and flexible enough to capture many important aspects of the tracking performance. The OSPA 2 distance is defined as follows:
    d p , q ( c , 2 ) ( X , X ˜ ; ω ) = Δ 1 n min π Π n i = 1 m d q ( c ) ( x i , x ˜ π ( i ) ; ω ) p + c p ( n m ) 1 p ,
    where q is the order of the base distance and w is a collection of convex weights.
Finally, we introduce the single integrated air picture (SIAP) metric. Despite the terminology, it is applicable to tracking in general and not just in relation to an air picture. It is made up of multiple individual metrics [62,63]. The SIAP metric requires an association between tracks and targets. We used a unique association in this paper. A description of the key metrics of the SIAP is given in Table 1.

3. CE Approaches

The above metrics provide the criteria for the PE of the tracking algorithm, which are combined to give the overall performance by the CE model. In this section, we review several CE approaches to analyze and judge the performance of target tracking.

3.1. The Weight of Each Evaluation Metric Set

The analytic hierarchy process (AHP) is given to ascertain the proportion of each metric, which was used to model and analyze the evaluation metric of the PE according to layers [64,65,66]. The specific steps are as follows.
Step 1 
The assessment metric system:
According to Section 2, all levels of evaluation metrics are established. U i ( i [ 1 , m ] ) in U = U 1 , U 2 , , U m is the ith metric of the primary metric in the system; U i j ( j [ 1 , l ] ) in U i = U i 1 , U i 2 , , U i m is the jth metric of U i ; U i j = U i j 1 , U i j 2 , , U i j m , U i j k is the kth metric of U i j . The rest can be established in the same manner;
Step 2 
The comparison matrix A i j :
Citing the numbers 1-9 as a scale, each influencing metric U i in the above metric set U is determined according to the importance of the element to its corresponding quantized value A i j . A = ( A i j ) n × n , A j i = 1 A i j , A i i = 1 ;
Step 3 
The maximum eigenvalue λ m a x of A and the corresponding normalized eigenvector W:
W is denoted as W = ( W 1 , W 2 , W n ) , i = 1 n W i = 1 , where W i denotes the weight of the ith evaluation metric.

3.2. Cloud Barycenter Evaluation

Based on the traditional fuzzy set and probability theory, the cloud theory provides a powerful method by combining the qualitative information with the quantitative data [67]. As a kind of mathematical model, cloud theory describes the mapping relationship between quality and quantity through a fuzzy and stochastic relation completely [68]. The cloud barycenter evaluation developed from cloud theory is a CE method that has been extensively used in numerous complex systems, especially in the military field [67]. The cloud barycenter evaluation method is a qualitative and quantitative method to achieve the transformation between the conception and the data.
The cloud is represented by three digital characteristics, including the expected value E x , the entropy E n , and the hyper entropy H e [67,69], where E x is the central value of the fuzzy concept in the defined domain, E n represents the degree of the fuzziness of the qualitative concept, and H e is the fuzzy measurement and the entropy of E n , which is a mapping of the uncertainty of the qualitative concept.
The cloud barycenter evaluation is realized by establishing the cloud model of each metric. The specific evaluation processes are as follows:
Step 1 
The cloud model of the comment set:
The comment set of metrics was ascertained by experts. For example, we set S= excellent, good, fair, worse, poor to denote the comment set of target tracking, which is shown in Table 2. We set the comment as the corresponding continuous number field interval [0,1]. The formula of the cloud model is represented as:
E x i 0 = c min + c max 2 , E n i 0 = c max c min 6 ,
where E x i 0 , E n i 0 are the expected values and entropy of some qualitative comments, respectively.
Step 2 
The quantitative and the qualitative variables for the given metric set;
(a)
The cloud model of quantitative metrics:
The corresponding quantitative metrics values were established by n experts as E x 11 , E x 21 , , E x n 1 , which can be denoted by the cloud model:
E x 1 = E x 11 + E x 21 + + E x n 1 n ,
E n 1 = max ( E x 11 , E x 21 , , E x n 1 ) min ( E x 11 , E x 21 , , E x n 1 ) 6 .
(b)
The cloud model of qualitative metrics:
In the same way, every qualitative metric, which are represented by the linguistic value, can also be described by the cloud model:
E x 2 = E x 12 E n 12 + E x 22 E n 22 + + E x n 2 E n n 2 E n 12 + E n 22 + + E n n 2 ,
E n 2 = E n 12 + E n 22 + + E n n 2 ,
where E x 12 , E x 22 , , E x n 2 and E n 12 , E n 22 , , E n n 2 denote the expected values and entropy of the cloud model, respectively;
Step 3 
The weighted departure degree:
S = ( S 1 , S 2 , , S n ) is the n-dimensional integrated barycenter vector, each dimension value of which is calculated by S i = g i × h i ( i = 1 , 2 , , n ) , where g i = ( E x 1 , E x 2 , , E x n ) is the cloud barycenter position and h i = ( W 1 , W 2 , , W n ) is the cloud barycenter height calculated by the AHP. S 0 = ( S 1 0 , S 2 0 , , S n 0 ) denotes the ideal cloud vector. The synthesized vector is normalized as follows:
S i T = S i 0 S i S i 0 , S i < S i 0 . S i S i 0 S i 0 , S i S i 0 .
Finally, the weighted departure degree is given by:
θ = j = 1 n ( S j T × W j ) ( 0 θ 1 ) ;
Step 4 
Result analysis:
The comment set is put in a consecutive interval. Meanwhile, each comment value is realized by a cloud model. The cloud-generator model can be established as Figure 3 shows. The comment set can be divided into five categories: excellent, good, fair, worse, and poor. For a specific case, assessment results can be output by inputting 1 + θ into the cloud-generator model.

3.3. Fuzzy CE Method

The fuzzy CE is based on fuzzy mathematics, which quantitatively expresses the objective attributes of some uncertain things [70,71,72]. The specific process is as follows:
Step 1 
The metric set:
We analyzed the result of target tracking and establish the evaluation metric set U as follows:
U = U 1 , U 2 , , U n ,
where U i is the ith evaluation metric;
Step 2 
The evaluation level set:
The evaluation level set is given by V = { V 1 , V 2 , , V m } , where V i is the ith grey category. V is the remark collection, which is made up of remarks of the research object;
Step 3 
The evaluation matrix:
Starting from a single factor for the evaluation, we determine the degree of membership about evaluation objects to the evaluation level set and make the fuzzy evaluation. Then, combining the single-factor set, a multi-factor evaluation set is given by:
R = R 1 R 2 R m = r 11 r 12 r 1 m r 21 r 22 r 2 m r n 1 r n 2 r n m ,
where r i j denotes the membership degree of U i corresponding to V j ;
Step 4 
The fuzzy CE value:
C = B R = b 1 b 2 b n r 11 r 12 r 1 m r 21 r 22 r 2 m r n 1 r n 2 r n m = c 1 c 2 c j ,
where C is the fuzzy CE set and B is the weight of the metric.
According to the principle of the maximum membership degree, the comprehensive value of the PE is obtained; thereby, the corresponding performance levels [70] are calculated.

3.4. Grey Clustering

Grey theory is a useful methodology for incomplete information systems. Grey relational analysis can be used to analyze relationships between the uncertainty and the gray category [73,74]. The main steps of the method are as follows:
Step 1 
Triangular whitenization weight functions are established and obtained as follows:
f 1 j k ( d i j ) = f 1 j k ( c 1 j k , ) = d i j c 1 j k , d i j [ 0 , c 1 j k ] 1 , d i j ( c 1 j k , ) 0 , d i j [ 0 , ) ,
f 2 j k ( d i j ) = f 2 j k ( , c 2 j k , + ) = d i j c 2 j k , d i j [ 0 , c 2 j k ] d i j 2 c 2 j k c 2 j k , d i j ( c 2 j k , 2 c 2 j k ] 0 , d i j [ 0 , 2 c 2 j k ] ,
f 3 j k ( d i j ) = f 3 j k ( 0 , c 3 j k ) = 1 , d i j [ 0 , c 3 j k ] d i j 2 c 3 j k c 3 j k , d i j ( c 3 j k , 2 c 3 j k ] 0 , d i j [ 0 , 2 c 3 j k ] ,
where d i j ( i = 1 , 2 , , n ; j = 1 , 2 , , m ) is the sample of the ith algorithm about jth evaluation metric and c j k is the midpoints of the jth clustering metric belonging to the kth grey category. The type of measure determines the choice of three functions: If the measure is extremely large data (preferably larger), we select f 1 j k ( d i j ) . If it is the moderate measure, f 2 j k ( d i j ) is selected. If it is extremely small, then f 3 j k ( d i j ) is the first choice;
Step 2 
The clustering coefficient:
σ i k = j = 1 m f j k ( d i j ) w j ,
δ i k = σ i k k = 1 s σ i k ,
where w j is the weight of the jth clustering and is determined by the AHP. σ i k is the weight cluster coefficient of the ith algorithm about the kth grey category, and δ i k is the normalized weight cluster coefficient, respectively [75];
Step 3 
The integrated clustering coefficient η i for each monitoring point with respect to the grey classes k can be calculated with the following equation [76]:
η i = k = 1 s k · δ i k , ( i = 1 , 2 , , n ) ;
Step 4 
According to the integrated clustering coefficient, the evaluation result is determined. The value range of the integrated clustering coefficient is divided into s intervals of the same length, which are: [ 1 , 1 + s 1 / s ] , [ 1 + s 1 / s , 1 + 2 ( s 1 ) / s ] , ⋯, [ s s 1 / s , s ] . The track algorithm is judged as the kth grey category, when η i belongs to 1 + ( k 1 ) ( s 1 ) / s , 1 + k ( s 1 ) / s .

4. Rating and Overall Performance

Most simulations are run in the Monte Carlo scenario to describe the characteristics of the performance metrics. In [21], the analysis and assessment of the tracking algorithm were performed both with simulated and real data, where the real data were measured with the Multi-Static Primary Surveillance Radar (MSPSR) L-Band demonstrator, and the metrics were calculated for the performance evaluation such as the mean variance statistic of the NMT, RTF, RMSE, etc. Refs. [25,77] calculated the GOSPA varying values of c and γ using a multiple target tracking example in the MATLAB code. The COCO 2017 validation set and the MOTChallenge (MOT17) dataset were used in terms of the Hausdorff, Wasserstein, and OSPA metrics [78]. In the paper, the metrics mentioned in Section 2 were combined to give a score or a membership value by the aforementioned CE approaches. Three measures were taken together to judge the efficiency of target tracking in the cloud barycenter evaluation (the synthetic measures do not involve simulations). At the same time, the realization of the fuzzy theory and the grey clustering is in the absence of the correctness and the timeliness scenario. In order to demonstrate it more vividly, our simulation was performed by using a graphical user interface (GUI).

4.1. Application of Cloud Theory for Target Tracking

In this section, we discuss the application of cloud theory for target tracking. In this scenario, three categories of the performance metrics are involved, and the last is the accuracy, which is divided into the TE, TPE, and TVE. The judgment matrix was given by experts and displayed in Figure 4, and then, the metric weights were ascertained. W 1 = (0.5436, 0.1634, 0.2970) is the weight of three primary metrics. Then, W 2 = (0.3746, 0.1835, 0.0943, 0.1184, 0.0983, 0.0503, 0.0806) is the weight of the correctness measure; W 3 = (0.0854, 0.5424, 0.2133, 0.1588) is the weight of the timeliness measure; W 4 = (0.1692, 0.4434, 0.3874) is the weight of the accuracy measure. We can draw some conclusions in which the tracking accuracy and correctness are more important than the timeliness in judging the efficiency of target tracking, where the sequential ratio of those is 0.2970, 0.5346, and 0.1634.
According to the evaluation of experts, S = excellent , good , fair , worse , poor was placed in [0,1]. Therefore, Table 3 represents the results of the cloud model of comments.
S = 0.9 0.7 0.9 0.7 0.5 0.5 0.3 0.3 0.5 0.3 0.1 0.7 0.7 0.7 0.9 0.9 0.5 0.7 0.5 0.5 0.5 0.3 0.5 0.1 0.3 0.5 0.5 0.5 0.9 0.5 0.7 0.5 0.5 0.7 0.3 0.5 0.9 0.5 0.5 0.7 0.7 0.7 0.7 0.7 0.7 0.5 0.5 0.5 0.7 0.7 0.7 0.7 0.5 0.5 0.9 0.7 0.9 0.9 0.5 0.7 0.7 0.5 0.5 0.3 0.5 0.5 0.7 0.9 0.9 0.9 0.7 0.5 0.9 0.7 0.5 0.1 0.9 0.1 0.9 0.9 0.3 0.5 0.9 0.7 0.9 0.7 0.5 0.7 0.3 0.7 0.5 0.7 0.7 0.7 0.5 0.9 0.7 0.9 0.9 0.7 0.5 0.5 0.3 0.5 0.7 0.7 0.3 0.3 0.5 0.7 0.7 0.7 0.9 0.5 0.7 0.7 0.3 0.3 0.5 0.5 0.3 0.3 0.9 0.9 0.9 0.9 0.9 0.5 0.5 0.7 0.9 0.5 0.7 0.5 0.5 0.5 0.7 0.7 0.7 0.9
S is the decision matrix, which denotes the integrated cloud gravity center. Combining Equations (13) and (14) and Table 3, the cloud model of the parameter status is calculated and given by Table 4.
Then, the conclusion can be obtained in the part of “performance evaluation”. W = [0.2022, 0.0990, 0.0509, 0.0639, 0.0530, 0.0271, 0.0435, 0.0140, 0.0886, 0.0349, 0.0260, 0.0277, 0.0725, 0.0633]. In the ideal state, E = [0.9, 0.9, 0.9, 0.9, 0.9, 0.9, 0.9, 0.9, 0.9, 0.9, 0.9, 0.9, 0.9, 0.9]; the integrated evaluation vector of the cloud barycenter is S 0 = W × E = (0.1819, 0.0891, 0.0458, 0.0575, 0.0477, 0.0244, 0.0391, 0.0126, 0.0798, 0.0314, 0.0234, 0.0249, 0.0652, 0.0570). The actual integrated vector is S = ( W 1 , W 2 , , W 14 ) × ( E x 1 , E x 2 , , E x 14 ) = (0.1739, 0.0653, 0.0326, 0.0409, 0.0265, 0.0130, 0.0244, 0.0064, 0.0603, 0.0167, 0.0130, 0.0194, 0.0551, 0.0481). S is normalized as S T = ( S 1 T , S 2 T , , S 14 T ) = (−0.0444, −0.2667, −0.2889, −0.2889, −0.4444, −0.4667, −0.3778, −0.4889, −0.2444, −0.4667, −0.4444, −0.2222, −0.1556, −0.1556). Finally, the weighted departure degree can be acquired as θ = −0.20477.
In the ideal state, θ for PE is zero; however, the actual θ is 0.20477 . When 1 + θ = 0.79523 is input into the cloud generator model, the cloud drops close to the “good” cloud object. Then, the PE result of target tracking is deemed to be good. We came up with a conjecture that the measure of the accuracy and the correctness are more effective than the timeliness in target tracking. Then, we verified it in the following simulation.

4.2. Application of Fuzzy CE for Target Tracking

Fuzzy CE uses the fuzzy mathematics tool to depict the influence of various metrics, which was applied in the situation without the timeliness measure. Here, U= NMT U 1 , NVT U 2 , NST U 3 , NFT U 4 , TSE U 5 , TBE U 6 , TE U 7 . B = (0.1621, 0.1284, 0.0773, 0.1120, 0.0836, 0.2000, 0.2365) is the weight of the seven metrics. V = V 1 , V 2 , V 3 , V 4 denotes the comment set, corresponding four levels: “excellent”, “good”, “medium”, and “poor”. Here, the subordinate degree was determined by the expert assessment. We obtain R , C as follows.
R = R 1 R 2 R 3 R 4 R 5 R 6 R 7 = 0.2 0.4 0.4 0 0.3 0.4 0.3 0 0.1 0.4 0.5 0 0.2 0.4 0.4 0 0.1 0.5 0.4 0 0.2 0.3 0.5 0 0 0.5 0.5 0
C = B R = 0.1621 0.1284 0.0773 0.1120 0.2000 0.1922 0.2365 T 0.2 0.4 0.4 0 0.3 0.4 0.3 0 0.1 0.4 0.5 0 0.2 0.4 0.4 0 0.1 0.5 0.4 0 0.2 0.3 0.5 0 0 0.5 0.5 0 = 0.1494 0.4120 0.4386 0
The simulation result is given in Figure 5. According to the metric information, we could obtain the fuzzy CE of all the evaluation information. The fuzzy CE set was calculated and performed the “medium” in terms of the maximum membership degree on the basis of the principle. To combine the analysis above, the total score of the PE for target tracking can be given by: D = 0.1494 × 90 + 0.4120 × 80 + 0.4386 × 70 = 77.108, and the result showed that the performance for target tracking performed medium.

4.3. PE in Target Tracking Using Grey Clustering

In this section, representative metrics of the timeliness and the accuracy measures were chosen, which included the TPE, TL, TVE, TPD, and RFA. w = 0.29783, 0.088788, 0.15777, 0.29783, 0.15777 is the weight of these. Four grey categories were established, where K = 1 , 2 , 3 , 4 denotes “excellent”, “good”, “medium”, and “poor”. Based on the scenario [79], the following tracking methods were used for the performance evaluation where results were available for 200 Monte Carlo runs:
(1)
Pseudo-observation (PS) [80];
(2)
Projection (PRO) [81,82];
(3)
Karush–Kuhn–Tucker (KKT) [83];
(4)
Karush–Kuhn–Tucker–Kalman filter (KKT_KF) [79];
(5)
Unconstrained Kalman filter (UKF) [84];
(6)
Trajectory function of time (T-FoT) [85,86].
Performance metrics were now found based on the given tracking results, which are given in Table 5. According to Table 5, we can determine the whitenization weight functions in Table 6. Taking the first metric (TPE) as an example, the four whitenization weight functions are:
f 1 1 6 , = d i j 6 , d i j 0 , 6 1 , d i j 6 , 0 , d i j [ 0 , )
f 1 2 , 4.5 , + = d i j c 2 j k , d i j 0 , 4.5 d i j 9 4.5 , d i j ( 4.5 , 9 ] 0 , d i j 0 , 9
f 1 3 , 3 , + = d i j 3 , d i j 0 , 3 d i j 6 3 , d i j ( 3 , 6 ] 0 , d i j 0 , 6
f 1 4 0 , 1 = 1 , d i j 0 , 1 d i j 2 1 , d i j ( 1 , 2 ] 0 , d i j 0 , 2
f 1 j k ( c 1 j k , ) was applied in the excellent grey category, which is an upper measure. The good grey category and medium grey category were selected by f 2 j k ( , c 2 j k , + ) . f 3 j k ( 0 , c 3 j k ) is a whitenization weight function of lower measure, which was used in the poor grey category next. From the weight of the evaluation metric and the grey clustering, the weighted clustering coefficient matrix σ can be determined.
σ = ( σ i k ) = 0.7331 0.8083 0.4698 0 0.8656 0.8006 0.3427 0 0.9255 0.7552 0.2425 0 0.6936 0.8847 0.5296 0 0.8024 0.8481 0.2258 0 0.8966 0.6882 0.1425 0
Furthermore, the normalized clustering coefficient matrix can be obtained:
δ i k = 0.3645 0.4019 0.2336 0 0.4309 0.3985 0.1706 0 0.4813 0.3927 0.1261 0 0.3290 0.4197 0.2513 0 0.4277 0.4520 0.1203 0 0.5191 0.3984 0.0825 0
The simulation result is shown in Figure 6. The integrated clustering coefficients of the six algorithms were calculated. For example, the integrated clustering coefficient of the PS in the first algorithm was η 1 = k = 1 4 k · δ i k = 1×0.3645 + 2×0.4019 + 3×0.2336 + 4×0 = 1.8691, and the other coefficients were η 2 = 1.7397 (PRO), η 3 = 1.6448 (KKT), η 4 = 1.9222 (KKT_KF), η 5 = 1.6927 (UKF), η 6 = 1.5634 (T-FoT), respectively. The value range of the integrated clustering coefficient was divided into four intervals of the same length. According to Step 4 in Section 3.4, η 2 , η 3 , η 5 , η 6 1 , 1 + 3 / 4 , η 1 , η 4 1 + 3 / 4 , 1 + 6 / 4 , and the evaluated result was that the methods of the PRO, KKT, UKF, and T-FoT had “excellent” performance and the methods of the PS and KKT_KF “good” performance. It is not difficult to find that the CE method for target tracking achieved a satisfactory result by comparing the above algorithms.
The simulation showed that the three measures contained sufficient information for evaluation by the comparative analysis. However, there were slight differences in the scenarios in which the one lacks correctness and the other one lacks timeliness. The results of grey clustering were consistent with cloud theory, and the fuzzy CE showed poor results. In other words, for this case, it can be concluded that the timeliness measure was less informative than the correctness and accuracy measures.
So far, on the one hand, PE has a tendency to focus on improving OSPA. On the other hand, PE metrics have been redefined using different association algorithms. A more effective PE method needs to be explored to enhance the algorithm efficiency.

5. Conclusions and Remaining Challenges

This paper reviewed PE metrics for target tracking and some CE approaches. The measures were divided into three categories and described for each category. Finally, the simulation result showed that a combination of metrics from different classes can provide a criterion for PE in target tracking.
Instead of estimating the discrete-time state of the target, it is actually of greater interest to estimate the continuous-time trajectory of the target, which contains more information than the discrete time series of point estimates. The T-FoT framework [85,87,88] is promising and powerful as it completely describes the movement pattern/dynamic behavior of the targets over time and enables the use of many curve-fitting tools such as the Gauss process and neural networks, in addition to various flexible parametric regression analysis methods. However, based on this framework, the output of the estimator or tracker is a spatio-temporal trajectory function, or functions in the case of multiple targets. How to evaluate the quality of the trajectory functions/curves in the presence of missing detection and false alarms remains an open problem.

Author Contributions

This paper is a collaborative work by all the authors. Y.S.: methodology, investigation, writing—original draft, formal analysis, and software. Z.H.: resources, formal analysis, and validation. T.L.: conceptualization, methodology, and writing—review and editing. H.F.: writing—review and editing. All authors have read and agreed to the published version of the manuscript.

Funding

This work was partially supported by the National Natural Science Foundation of China (62071389), by the JWKJW Foundation (2021-JCJQ-JJ-0897, 2020-JCJQ-ZD-150-12), and by the Key Laboratory Foundation of National Defence Technology (JKWATR-210504).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data will be made available upon request.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
CEComprehensive evaluation
PEPerformance evaluation
TPETrack position error
TETrajectory error
TVETrack velocity error
NVTNumber of valid tracks
NMTNumber of missed targets
NFTNumber of false tracks
NSTNumber of spurious tracks
ANSTAverage number of swaps in tracks
ANBTAverage number of broken tracks
TRTracks redundancy
RFARate of false alarms
TPDTrack probability of detection
RTFRate of track fragmentation
TLTrack latency
TETTotal execution time
OSPAOptimal subpattern assignment
AHPAnalytic hierarchy process
SIAPSingle integrated air picture
MOPTMultiple-object-tracking precision
MOTAMultiple-object-tracking accuracy
GUIGraphical user interface
PSPseudo-observation
PROProjection
KKTKarush–Kuhn–Tucker
KKT_KFKarush–Kuhn–Tucker–Kalman filter
UKFUnconstrained Kalman filter
T-FoTTrajectory function of time

References

  1. Chong, C.Y. Tracking and data fusion: A handbook of algorithms (bar-shalom, y. et al; 2011)[bookshelf]. IEEE Control. Syst. Mag. 2012, 32, 114–116. [Google Scholar]
  2. Blasch, E. Target Tracking Toolbox lecture notes and software to support EE716. Ph.D. Thesis, Wright State University, Dayton, OH, USA, 2001. [Google Scholar]
  3. Available online: https://www.pdx.edu/biomedical-signal-processing-lab/signal-point-kalman-filters-and-the-rebel-toolkit (accessed on 10 September 2021).
  4. Wan, E.A.; Van Der Merwe, R.; Haykin, S. The unscented Kalman filter. Kalman Filter. Neural Net. 2001, 5, 221–280. [Google Scholar]
  5. Paul, A.S. Sigma-point Kalman Smoothing: Algorithms and Analysis with Applications to Indoor Tracking. Ph.D. Thesis, Oregon Health & Science University, Portland, OR, USA, 2010. [Google Scholar]
  6. Straka, O.; Flídr, M.; Duník, J.; Ŝimandl, M. A software framework and tool for nonlinear state estimation. IFAC Proc. Vol. 2009, 42, 510–515. [Google Scholar] [CrossRef]
  7. Straka, O.; Flídr, M.; Duník, J.; Simandl, M.; Blasch, E. Nonlinear estimation framework in target tracking. In Proceedings of the 2010 13th International Conference on Information Fusion, Edinburgh, UK, 26–29 July 2010. [Google Scholar]
  8. Blasch, E.P.; Straka, O.; Duník, J.; Šimandl, M. Multitarget tracking performance analysis using the non-credibility index in the nonlinear estimation framework (NEF) toolbox. In Proceedings of the IEEE 2010 National Aerospace & Electronics Conference, Dayton, OH, USA, 14–16 July 2010; pp. 107–115. [Google Scholar]
  9. Crouse, D.F. The tracker component library: Free routines for rapid prototyping. IEEE Aerosp. Electron. Syst. Mag. 2017, 32, 18–27. [Google Scholar] [CrossRef]
  10. Thomas, P.A.; Barr, J.; Balaji, B.; White, K. An open source framework for tracking and state estimation (’Stone Soup’). In Signal Processing, Sensor/Information Fusion, and Target Recognition XXVI; International Society for Optics and Photonics: Bellingham, WA, USA, 2017; Volume 10200, p. 1020008. [Google Scholar]
  11. Last, D.; Thomas, P.; Hiscocks, S.; Barr, J.; Kirkland, D.; Rashid, M.; Li, S.B.; Vladimirov, L. Stone Soup: Announcement of beta release of an open-source framework for tracking and state estimation. In Signal Processing, Sensor/Information Fusion, and Target Recognition XXVIII; International Society for Optics and Photonics: Bellingham, WA, USA, 2019; Volume 11018, p. 1101807. [Google Scholar]
  12. Costa, P.C.; Laskey, K.B.; Blasch, E.; Jousselme, A.L. Towards unbiased evaluation of uncertainty reasoning: The URREF ontology. In Proceedings of the 2012 15th International Conference on Information Fusion, Singapore, 9–12 July 2012; pp. 2301–2308. [Google Scholar]
  13. Xu, Z. Performance evaluation of business administration training room in application-oriented universities. In Proceedings of the 2020 2nd International Conference on Computer Science Communication and Network Security (CSCNS2020), Sanya, China, 22–23 December 2021; Volume 336, p. 09015. [Google Scholar]
  14. Zhang, G.; Hui, G.; Zhang, G.; Hu, Y.; Zhao, Z. A Novel Comprehensive Evaluation Method of Forest State Based on Unit Circle. Forests 2019, 10, 5. [Google Scholar] [CrossRef] [Green Version]
  15. Li, X. Application Research on the Model of the Performance Evaluation of Enterprise Informatization. J. Inf. 2008, 12, 15–17. [Google Scholar]
  16. Bar-Shalom, Y.; Li, X.R.; Kirubarajan, T. Estimation with Applications to Tracking and Navigation: Theory, Algorithms and Software; John Wiley & Sons: Hoboken, NJ, USA, 2004; p. 584. [Google Scholar]
  17. Popp, R.L.; Kirubarajan, T.; Pattipati, K.R. Survey of assignment techniques for multitarget tracking. Multitarg.-Multisens. Tracking Appl. Adv. 2000, 3, 77–159. [Google Scholar]
  18. Colegrove, S.B.; Cheung, B.; Davey, S.J. Tracking system performance assessment. In Proceedings of the Sixth International Conference of Information Fusion, Cairns, Australia, 8–10 July 2003; Volume 2, pp. 926–933. [Google Scholar]
  19. Sheng, X.; Chen, Y.; Guo, L.; Yin, J.; Han, X. Multitarget Tracking Algorithm Using Multiple GMPHD Filter Data Fusion for Sonar Networks. Sensors 2018, 18, 3193. [Google Scholar] [CrossRef] [Green Version]
  20. Ristic, B.; Vo, B.N.; Clark, D.; Vo, B.T. A Metric for Performance Evaluation of Multi-Target Tracking Algorithms. IEEE Trans. Signal Process. 2011, 59, 3452–3457. [Google Scholar] [CrossRef]
  21. Kulmon, P.; Stukovska, P. Assessing Multiple-Target Tracking Performance Of GNN Association Algorithm. In Proceedings of the 2018 19th International Radar Symposium (IRS), Bonn, Germany, 20–22 June 2018. [Google Scholar]
  22. Evirgen, E.A. Multi sensor track fusion performance metrics. In Proceedings of the 2016 24th Signal Processing and Communication Application Conference (SIU), Zonguldak, Turkey, 16–19 May 2016; pp. 97–100. [Google Scholar]
  23. Gorji, A.A.; Tharmarasa, R.; Kirubarajan, T. Performance measures for multiple target tracking problems. In Proceedings of the 14th International Conference on Information Fusion, Chicago, IL, USA, 5–8 July 2011. [Google Scholar]
  24. de Villiers, J.P.; Focke, R.W.; Pavlin, G.; Jousselme, A.L.; Dragos, V.; Laskey, K.B.; Costa, P.; Blasch, E. Evaluation metrics for the practical application of URREF ontology: An illustration on data criteria. In Proceedings of the 2017 20th International Conference on Information Fusion (Fusion), Xi’an, China, 10–13 July 2017. [Google Scholar]
  25. García-Fernández, Á.F.; Rahmathullah, A.S.; Svensson, L. A time-weighted metric for sets of trajectories to assess multi-object tracking algorithms. arXiv 2021, arXiv:2110.13444. [Google Scholar]
  26. Rothrock, R.L.; Drummond, O.E. Performance metrics for multiple-sensor multiple-target tracking. In Proceedings of the SPIE Conference on Signal and Data Processing of Small Targets, San Diego, CA, USA, 30 July–2 August 2001; Volume 4048, pp. 521–531. [Google Scholar]
  27. Drummond, O.; Fridling, B. Ambiguities in evaluating performance of multiple target tracking algorithms. Proc. Spie Int. Soc. Opt. Eng. 1992, 1698, 326–337. [Google Scholar]
  28. Li, X.; Zhao, Z. Evaluation of estimation algorithms part I: Incomprehensive measures of performance. Aerosp. Electron. Syst. IEEE Trans. 2006, 42, 1340–1358. [Google Scholar] [CrossRef]
  29. Blackman, S.; Popoli, R. Design and Analysis of Modern Tracking Systems; Artech House Publishers: London, UK, 1999; p. 1015. [Google Scholar]
  30. Bernardin, K.; Stiefelhagen, R. Evaluating multiple object tracking performance: The clear mot metrics. Eurasip J. Image Video Process. 2008, 2008, 246309. [Google Scholar] [CrossRef] [Green Version]
  31. Milan, A.; Schindler, K.; Roth, S. Challenges of ground truth evaluation of multitarget tracking. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Portland, OR, USA, 23–28 June 2013; pp. 735–742. [Google Scholar]
  32. Pelletier, M.; Sivagnanam, S.; Blasch, E.P. A track scoring MOP for perimeter surveillance radar evaluation. In Proceedings of the 2012 15th International Conference on Information Fusion, Singapore, 9–12 July 2012; pp. 2028–2034. [Google Scholar]
  33. Blasch, E.P.; Straka, O.; Yang, C.; Qiu, D.; Šimandl, M.; Ajgl, J. Distributed tracking fidelity-metric performance analysis using confusion matrices. In Proceedings of the 2012 15th International Conference on Information Fusion, Singapore, 9–12 July 2012. [Google Scholar]
  34. Evers, C.; Löllmann, H.W.; Mellmann, H.; Schmidt, A.; Barfuss, H.; Naylor, P.A.; Kellermann, W. LOCATA challenge-overview of evaluation measures. Trans. Signal Process. 2008, 56, 3447–3457. [Google Scholar]
  35. Mori, S.; Chang, K.C.; Chong, C.Y.; Dunn, K.P. Tracking performance evaluation-prediction of track purity. In Signal and Data Processing of Small Targets 1989; International Society for Optics and Photonics: Bellingham, WA, USA, 1989; Volume 1096, pp. 215–223. [Google Scholar]
  36. Blasch, E.P.; Valin, P. Track purity and current assignment ratio for target tracking and identification evaluation. In Proceedings of the 14th International Conference on Information Fusion, Chicago, IL, USA, 5–8 July 2011. [Google Scholar]
  37. Blasch, E. Fusion Evaluation Tutorial. In Proceedings of the International Conference on Information Fusion, Stockholm, Sweden, 28 June–1 July 2004. [Google Scholar]
  38. Coraluppi, S.; Grimmett, D.; de Theije, P. Benchmark Evaluation of Multistatic Trackers. In Proceedings of the 2006 9th International Conference on Information Fusion, Florence, Italy, 10–13 July 2006. [Google Scholar]
  39. Grimmett, D.; Coraluppi, S.; Cour, B.; Hempel, C.G.; Lang, T.; Theije, P.; Willett, P. MSTWG multistatic tracker evaluation using simulated scenario data sets. In Proceedings of the 2008 11th International Conference on Information Fusion, Cologne, Germany, 30 June–3 July 2008. [Google Scholar]
  40. Guerriero, M.; Svensson, L.; Svensson, D.; Willett, P. Shooting two birds with two bullets: How to find Minimum Mean OSPA estimates. In Proceedings of the 2010 13th International Conference on Information Fusion, Edinburgh, UK, 26–29 July 2010. [Google Scholar]
  41. Chang, F.; Chen, Z.; Wang, W.; Wang, L. The Hausdorff distance template matching algorithm based on Kalman filter for target tracking. In Proceedings of the 2009 IEEE International Conference on Automation and Logistics, Shenyang, China, 5–7 August 2009; pp. 836–840. [Google Scholar]
  42. Da, K.; Li, T.; Zhu, Y.; Fu, Q. A Computationally Efficient Approach for Distributed Sensor Localization and Multitarget Tracking. IEEE Commun. Lett. 2020, 24, 335–338. [Google Scholar] [CrossRef]
  43. Hoffman, J.R.; Mahler, R. Multitarget miss distance and its applications. In Proceedings of the Fifth International Conference on Information Fusion. FUSION 2002, Annapolis, MD, USA, 8–11 July 2002; Volume 1, pp. 149–155. [Google Scholar]
  44. Schuhmacher, D.; Vo, B.T.; Vo, B.N. A Consistent Metric for Performance Evaluation of Multi-Object Filters. IEEE Trans. Signal Process. 2008, 56, 3447–3457. [Google Scholar] [CrossRef] [Green Version]
  45. Ristic, B.; Vo, B.N.; Clark, D. Performance evaluation of multitarget tracking using the OSPA metric. In Proceedings of the 2010 13th International Conference on Information Fusion, Edinburgh, UK, 26–29 July 2010. [Google Scholar]
  46. Nagappa, S.; Clark, D.E.; Mahler, R. Incorporating track uncertainty into the OSPA metric. In Proceedings of the 14th International Conference on Information Fusion, Chicago, IL, USA, 5–8 July 2011. [Google Scholar]
  47. El-Fallah, A.I.; Ravichandran, R.B.; Mehra, R.K.; Hoffman, J.R.; Alford, M.G. Scientific performance evaluation for distributed sensor management and adaptive data fusion. In Signal Processing, Sensor Fusion, and Target Recognition X; International Society for Optics and Photonics: Bellingham, WA, USA, 2001; Volume 4380, pp. 328–338. [Google Scholar]
  48. Hoffman, J.R.; Mahler, R.; Zajic, T. User-defined information and scientific performance evaluation. In Signal Processing, Sensor Fusion, and Target Recognition X; International Society for Optics and Photonics: Bellingham, WA, USA, 2001; Volume 4380, pp. 300–311. [Google Scholar]
  49. Mahler, R. Scientific performance metrics for data fusion: New results. In Signal Processing, Sensor Fusion, and Target Recognition IX; International Society for Optics and Photonics: Bellingham, WA, USA, 2000; Volume 4052, pp. 172–182. [Google Scholar]
  50. Villani, C. Optimal Transport. Old and New; Springer: Berlin/Heidelberg, Germany, 2009; p. 976. [Google Scholar]
  51. Hoffman, J.R.; Mahler, R. Multitarget Miss Distance via Optimal Assignment. IEEE Trans. Syst. Man Cybern. Part Syst. Hum. 2004, 34, 327–336. [Google Scholar] [CrossRef]
  52. García-Femández, Á.F.; Svensson, L. Spooky effect in optimal OSPA estimation and how GOSPA solves it. In Proceedings of the 2019 22th International Conference on Information Fusion (FUSION), Ottawa, ON, Canada, 2–5 July 2019. [Google Scholar]
  53. Vu, T.; Evans, R. A new performance metric for multiple target tracking based on optimal subpattern assignment. In Proceedings of the 17th International Conference on Information Fusion (FUSION), Salamanca, Spain, 7–10 July 2014. [Google Scholar]
  54. Mei, L.; Li, H.; Zhou, Y.; Li, D.; Long, W.; Xing, F. Output-Only Damage Detection of Shear Building Structures Using an Autoregressive Model-Enhanced Optimal Subpattern Assignment Metric. Sensors 2020, 20, 2050. [Google Scholar] [CrossRef] [Green Version]
  55. Lian, F.; Zhang, G.H.; Duan, Z.S.; Han, C.Z. Multi-Target Joint Detection and Estimation Error Bound for the Sensor with Clutter and Missed Detection. Sensors 2016, 16, 169. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  56. Li, W.; Han, C. Dual Sensor Control Scheme for Multi-Target Tracking. Sensors 2018, 18, 1653. [Google Scholar] [CrossRef] [Green Version]
  57. Schubert, R.; Klöden, H.; Wanielik, G.; Kälberer, S. Performance evaluation of Multiple Target Tracking in the absence of reference data. In Proceedings of the 2010 13th International Conference on Information Fusion, Edinburgh, UK, 26–29 July 2010. [Google Scholar]
  58. Rahmathullah, A.S.; García-Fernández, Á.F.; Svensson, L. Generalized optimal sub-pattern assignment metric. In Proceedings of the 2017 20th International Conference on Information Fusion (Fusion), Xi’an, China, 10–13 July 2017. [Google Scholar]
  59. Xia, Y.; Granstrcom, K.; Svensson, L.; García-Fernández, A.F. Performance evaluation of multi-bernoulli conjugate priors for multitarget filtering. In Proceedings of the 2017 20th International Conference on Information Fusion (Fusion), Xi’an, China, 10–13 July 2017. [Google Scholar]
  60. Beard, M.; Vo, B.T.; Vo, B.N. OSPA(2): Using the OSPA metric to evaluate multitarget tracking performance. In Proceedings of the 2017 International Conference on Control, Automation and Information Sciences (ICCAIS), Chiang Mai, Thailand, 31 October–1 November 2017; pp. 86–91. [Google Scholar]
  61. Beard, M.; Ba, T.V.; Vo, B.N. A Solution for Large-Scale Multi-Object Tracking. IEEE Trans. Signal Process. 2020, 68, 2754–2769. [Google Scholar] [CrossRef] [Green Version]
  62. Votruba, P.; Nisley, R.; Rothrock, R.; Zombro, B. Single Integrated Air Picture (SIAP) Metrics Implementation; Technical Report; Single Integrated Air Picture System Engineering Task Force: Arlington, VA, USA, 2001. [Google Scholar]
  63. Available online: https://stonesoup.readthedocs.io/en/v0.1b7/auto_examples/Metrics.html (accessed on 16 September 2021).
  64. Shang, J.; Vargas, L. New Concepts and Applications of AHP in the Internet Era. J. Multi-Criteria Decis. Anal. 2012, 19, 1–2. [Google Scholar] [CrossRef]
  65. Cho, J.; Lee, J. Development of a new technology product evaluation model for assessing commercialization opportunities using Delphi method and fuzzy AHP approach. Expert Syst. Appl. 2013, 40, 5314–5330. [Google Scholar] [CrossRef]
  66. Song, W.; Wen, W.; Guo, Q.; Chen, H.; Zhao, J. Performance evaluation of sensor information fusion system based on cloud theory and fuzzy pattern recognition. In Proceedings of the 2020 IEEE International Conference on Information Technology, Big Data and Artificial Intelligence (ICIBA), Chongqing, China, 6–8 November 2020; Volume 1, pp. 299–303. [Google Scholar]
  67. Ma, L.Y.; Shi, Q.S. Method of Safety Assessment about the Electric Power Supply Company Based on Cloud Gravity Center Theory. Adv. Mater. Res. 2012, 354, 1149–1156. [Google Scholar] [CrossRef]
  68. Zhang, C.; Shi, Q.; Liu, T.L. Air Base Damage Evaluation Using Cloud Barycenter Evaluation Method. Adv. Mater. Res. 2012, 430, 803–807. [Google Scholar] [CrossRef]
  69. Gu, H.; Cheng, Z.; Quan, S. Equipment maintenance support capability evaluation using cloud barycenter evaluation method. Telkomnika Indones. J. Electr. Eng. 2013, 11, 599–606. [Google Scholar] [CrossRef]
  70. Liu, H.; Li, B.; Sun, Y.; Dou, X.; Zhang, Y.; Fan, X. Safety Evaluation of Large-size Transportation Bridges Based on Combination Weighting Fuzzy Comprehensive Evaluation Method. Iop Conf. Ser. Earth Environ. Sci. 2021, 787, 012194. [Google Scholar] [CrossRef]
  71. Zhang, L.; Pan, Z. Fuzzy Comprehensive Evaluation Based on Measure of Medium Truth Scale. In Proceedings of the 2009 International Conference on Artificial Intelligence and Computational Intelligence, Shanghai, China, 7–8 November 2009; Volume 2, pp. 83–87. [Google Scholar]
  72. Wang, J.; Zhang, Y.; Wang, Y.; Gu, L. Assessment of Building Energy Efficiency Standards Based on Fuzzy Evaluation Algorithm. Eng. Sustain. 2019, 173, 1–14. [Google Scholar] [CrossRef]
  73. Delgado, A.; Cuadra, D.; Simon, K.; Bonilla, K.; Lee, E. Evaluation of Water Quality in the Lower Huallaga River Watershed using the Grey Clustering Analysis Method. Int. J. Adv. Comput. Sci. Appl. 2021, 12. [Google Scholar] [CrossRef]
  74. Delgado, A.; Fernandez, A.; Chirinos, B.; Barboza, G.; Lee, E. Impact of the Mining Activity on the Water Quality in Peru Applying the Fuzzy Logic with the Grey Clustering Method. Int. J. Adv. Comput. Sci. Appl. 2021, 12. [Google Scholar] [CrossRef]
  75. Dang, Y.G.; Liu, S.F.; Liu, B. Study on the Integrated Grey Clustering Method under the Clustering Coefficient with Non-Distinguished Difference. Chin. J. Manag. Sci. 2005, 13, 69–73. [Google Scholar]
  76. Jiskani, I.M.; Han, S.; Rehman, A.U.; Shahani, N.M.; Brohi, M.A. An Integrated Entropy Weight and Grey Clustering Method-Based Evaluation to Improve Safety in Mines. Min. Metall. Explor. 2021, 38, 1773–1787. [Google Scholar] [CrossRef]
  77. Rahmathullah, A.S.; García-Fernández, Á.F.; Svensson, L. A metric on the space of finite sets of trajectories for evaluation of multitarget-tracking algorithms. IEEE Trans. Signal Process. 2016, 68, 3908–3917. [Google Scholar]
  78. Rezatofighi, H.; Nguyen, T.; Vo, B.N.; Vo, B.T.; Reid, I. How trustworthy are the existing performance evaluations for basic vision tasks? arXiv 2020, arXiv:2008.03533. [Google Scholar]
  79. Zhou, J.; Li, T.; Wang, X. State Estimation with Linear Equality Constraints Based on Trajectory Function of Time and Karush-Kuhn-Tucker Conditions. In Proceedings of the 2021 International Conference on Control, Automation and Information Sciences (ICCAIS), Xi’an, China, 14–17 October 2021; pp. 438–443. [Google Scholar]
  80. Tahk, M.; Speyer, J.L. Target tracking problems subject to kinematic constraints. IEEE Trans. Autom. Control. 1990, 35, 324–326. [Google Scholar] [CrossRef]
  81. Ko, S.; Bitmead, R.R. State estimation for linear systems with state equality constraints. Automatica 2007, 43, 1363–1368. [Google Scholar] [CrossRef]
  82. Simon, D.; Chia, T.L. Kalman filtering with state equality constraints. IEEE Trans. Aerosp. Electron. Syst. 2002, 38, 128–136. [Google Scholar] [CrossRef] [Green Version]
  83. Boyd, S.; Boyd, S.P.; Vandenberghe, L. Convex Optimization; Cambridge University Press: Cambridge, UK, 2004. [Google Scholar]
  84. Xu, L.; Li, X.R.; Duan, Z.; Lan, J. Modeling and state estimation for dynamic systems with linear equality constraints. IEEE Trans. Signal Process. 2013, 61, 2927–2939. [Google Scholar] [CrossRef]
  85. Li, T.; Chen, H.; Sun, S.; Corchado, J.M. Joint Smoothing and Tracking Based on Continuous-Time Target Trajectory Function Fitting. IEEE Trans. Autom. Sci. Eng. 2019, 16, 1476–1483. [Google Scholar] [CrossRef] [Green Version]
  86. Zhou, J.; Li, T.; Wang, X.; Zheng, L. Target Tracking With Equality/Inequality Constraints Based on Trajectory Function of Time. IEEE Signal Process. Lett. 2021, 28, 1330–1334. [Google Scholar] [CrossRef]
  87. Li, T. Single-Road-Constrained Positioning Based on Deterministic Trajectory Geometry. IEEE Comm. Lett. 2019, 23, 80–83. [Google Scholar] [CrossRef]
  88. Li, T.; Fan, H. From Target Tracking to Targeting Track: A Data-Driven Approach to Non-cooperative Target Detection and Tracking. arXiv 2021, arXiv:2104.11122. [Google Scholar]
Figure 1. Classification of representative CE metrics.
Figure 1. Classification of representative CE metrics.
Sensors 22 00793 g001
Figure 2. Mapping the tracker hypotheses to objects. In the easiest case, different associations result in evaluation metrics.
Figure 2. Mapping the tracker hypotheses to objects. In the easiest case, different associations result in evaluation metrics.
Sensors 22 00793 g002
Figure 3. Qualitative evaluation of the cloud-generator model.
Figure 3. Qualitative evaluation of the cloud-generator model.
Sensors 22 00793 g003
Figure 4. PE of cloud gravity for target tracking.
Figure 4. PE of cloud gravity for target tracking.
Sensors 22 00793 g004
Figure 5. PE of target tracking based on fuzzy CE.
Figure 5. PE of target tracking based on fuzzy CE.
Sensors 22 00793 g005
Figure 6. PE for target tracking based on grey clustering.
Figure 6. PE for target tracking based on grey clustering.
Sensors 22 00793 g006
Table 1. The key metrics of the SIAP.
Table 1. The key metrics of the SIAP.
MetricDescription
AmbiguityA measure of the number of tracks assigned to each true object
CompletenessThe percentage of live objects with tracks on them
LSThe percentage of time spent tracking true objects across the dataset
LT1/R, where R is the average number of excess tracks assigned;
the higher this value, the better
Positional AccuracyGiven by the average positional error of the track to the truth
SpuriousnessThe percentage of tracks unsigned to any object
Velocity AccuracyThe average error in the velocity of the track to the truth
Number of TargetsThe total number of targets
Number of TracksThe total number of tracks
Table 2. Number field variation interval of the comments set.
Table 2. Number field variation interval of the comments set.
CommentsExcellentGoodFairWorsePoor
Number field interval[1,c1][c1,c2][c2,c3][c3,c4][c4,0]
Table 3. Cloud model of the comments.
Table 3. Cloud model of the comments.
CommentsNumber Field IntervalNumeral Characteristics
excellent[1,0.8](0.9,0.033)
good[0.8,0.6](0.7,0.033)
fair[0.6,0.4](0.5,0.033)
worse[0.4,0.2](0.3,0.033)
poor[0.2,0](0.1,0.033)
Table 4. The cloud model of the parameter status.
Table 4. The cloud model of the parameter status.
ParameterExpectationsEntropy
C10.860.33
C20.660.33
C30.640.33
C40.640.33
C50.50.33
C60.480.33
C70.560.33
C80.460.33
C90.680.33
C100.480.33
C110.50.33
C120.70.33
C130.760.33
C140.760.33
Table 5. Metric statistics of each tracking algorithm.
Table 5. Metric statistics of each tracking algorithm.
ArithmeticTPETLTVETPDRFA
PS5.254512.10560.9580.00042
PRO4.499712.87360.9730.0085
KKT4.499723.07890.9650.014
KKT_KF3.904812.5050.9660.0007
UKF6.355112.5050.9680.001
T-FoT5.530923.07890.9610.005
Table 6. Whitenization weight functions of four grey categories.
Table 6. Whitenization weight functions of four grey categories.
Excellent Grey CategoryGood Grey Category
f 1 1 ( c 1 1 , ) = f 1 1 ( 6 , ) f 1 2 ( , c 1 2 , + ) = f 1 2 ( , 4.5 , + )
f 2 1 ( c 2 1 , ) = f 2 1 ( 1.5 , ) f 2 2 ( , c 2 2 , + ) = f 2 2 ( , 1.25 , + )
f 3 1 ( c 3 1 , ) = f 3 1 ( 3 , ) f 3 2 ( , c 3 2 , + ) = f 3 2 ( , 2.6 , + )
f 4 1 ( c 4 1 , ) = f 4 1 ( 0.965 , ) f 4 2 ( , c 4 2 , + ) = f 4 2 ( , 0.95 , + )
f 5 1 ( c 5 1 , ) = f 5 1 ( 0.01 , ) f 5 2 ( , c 5 2 , + ) = f 5 2 ( , 0.001 , + )
Medium Grey CategoryPoor Grey Category
f 1 3 ( , c 1 3 , + ) = f 1 3 ( , 3 , + ) f 1 4 ( 0 , c 1 4 ) = f 1 4 ( 0 , 1 )
f 2 3 ( , c 2 3 , + ) = f 2 3 ( , 1 , + ) f 2 4 ( 0 , c 2 4 ) = f 2 4 ( 0 , 0.5 )
f 3 3 ( , c 3 3 , + ) = f 3 3 ( , 2 , + ) f 3 4 ( 0 , c 3 4 ) = f 3 4 ( 0 , 1 )
f 4 3 ( , c 4 3 , + ) = f 4 3 ( , 0.5 , + ) f 4 4 ( 0 , c 4 4 ) = f 4 4 ( 0 , 0.5 )
f 5 3 ( , c 5 3 , + ) = f 5 3 ( , 0.0005 , + ) f 5 4 ( 0 , c 5 4 ) = f 5 4 ( 0 , 0.0001 )
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Song, Y.; Hu, Z.; Li, T.; Fan, H. Performance Evaluation Metrics and Approaches for Target Tracking: A Survey. Sensors 2022, 22, 793. https://doi.org/10.3390/s22030793

AMA Style

Song Y, Hu Z, Li T, Fan H. Performance Evaluation Metrics and Approaches for Target Tracking: A Survey. Sensors. 2022; 22(3):793. https://doi.org/10.3390/s22030793

Chicago/Turabian Style

Song, Yan, Zheng Hu, Tiancheng Li, and Hongqi Fan. 2022. "Performance Evaluation Metrics and Approaches for Target Tracking: A Survey" Sensors 22, no. 3: 793. https://doi.org/10.3390/s22030793

APA Style

Song, Y., Hu, Z., Li, T., & Fan, H. (2022). Performance Evaluation Metrics and Approaches for Target Tracking: A Survey. Sensors, 22(3), 793. https://doi.org/10.3390/s22030793

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop