Next Article in Journal
Stochastic Capsule Endoscopy Image Enhancement
Previous Article in Journal
Efficient Implementation of Gaussian and Laplacian Kernels for Feature Extraction from IP Fisheye Cameras
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Review of Supervised Edge Detection Evaluation Methods and an Objective Comparison of Filtering Gradient Computations Using Hysteresis Thresholds

by
Baptiste Magnier
1,*,
Hasan Abdulrahman
2 and
Philippe Montesinos
1
1
IMT Mines d’Alès, Laboratoire de Génie Informatique et d’Ingénierie de Production (LGI2P), 6. avenue de Clavières, 30100 Alès, France
2
Northern Technical University, Department of Technical Computer systems, Kirkuk 36001, Iraki
*
Author to whom correspondence should be addressed.
J. Imaging 2018, 4(6), 74; https://doi.org/10.3390/jimaging4060074
Submission received: 19 March 2018 / Revised: 9 May 2018 / Accepted: 24 May 2018 / Published: 31 May 2018

Abstract

:
Useful for human visual perception, edge detection remains a crucial stage in numerous image processing applications. One of the most challenging goals in contour detection is to operate algorithms that can process visual information as humans require. To ensure that an edge detection technique is reliable, it needs to be rigorously assessed before being used in a computer vision tool. This assessment corresponds to a supervised evaluation process to quantify differences between a reference edge map and a candidate, computed by a performance measure/criterion. To achieve this task, a supervised evaluation computes a score between a ground truth edge map and a candidate image. This paper presents a survey of supervised edge detection evaluation methods. Considering a ground truth edge map, various methods have been developed to assess a desired contour. Several techniques are based on the number of false positive, false negative, true positive and/or true negative points. Other methods strongly penalize misplaced points when they are outside a window centered on a true or false point. In addition, many approaches compute the distance from the position where a contour point should be located. Most of these edge detection assessment methods will be detailed, highlighting their drawbacks using several examples. In this study, a new supervised edge map quality measure is proposed. The new measure provides an overall evaluation of the quality of a contour map by taking into account the number of false positives and false negatives, and the degrees of shifting. Numerous examples and experiments show the importance of penalizing false negative points differently than false positive pixels because some false points may not necessarily disturb the visibility of desired objects, whereas false negative points can significantly change the aspect of an object. Finally, an objective assessment is performed by varying the hysteresis thresholds on contours of real images obtained by filtering techniques. Theoretically, by varying the hysteresis thresholds of the thin edges obtained by filtering gradient computations, the minimum score of the measure corresponds to the best edge map, compared to the ground truth. Twenty-eight measures are compared using different edge detectors that are robust or not robust regarding noise. The scores of the different measures and different edge detectors are recorded and plotted as a function of the noise level in the original image. The plotted curve of a reliable edge detection measure must increase monotonously with the noise level and a reliable edge detector must be less penalized than a poor detector. In addition, the obtained edge map tied to the minimum score of a considered measure exposes the reliability of an edge detection evaluation measure if the edge map obtained is visually closer to the ground truth or not. Hence, experiments illustrate that the desired objects are not always completely visible using ill-suited evaluation measure.

1. Introduction: Edge Detection and Hysteresis Thresholding

A digital image is a discrete representation of a real and continuous world. Each point of an image, i.e., pixel, quantifies a piece or pieces of gray-scale, brightness or color information. The transition between dark and bright pixels corresponds to contours. They are essential information for the interpretation and exploitation of images. Edge detection is an important field and one of the oldest topics in image processing because the process frequently attempts to capture the most important structures in the image [1]. Edge detection is therefore a fundamental step in computer vision approaches. Furthermore, edge detection could itself be used to qualify a region segmentation technique. Additionally, the edge detection assessment remains very useful in image segmentation, registration, reconstruction or interpretation. It is hard to design an algorithm that is able to detect the exact edge from an image with good localization and orientation. In the literature, various techniques have emerged and, due to its importance, edge detection continues to be an active research area [2]. The detection is based on the local geometric properties of the considered image by searching for intensity variation in the gradient direction [1]. There are two main approaches for contour detection: first-order derivative [3,4,5,6,7] or second-order [8]. The best-known and most useful edge detection methods are based on gradient computing first-order fixed operators [3,4]. Oriented first-order operators compute the maximum energy in an orientation [9,10,11] or two directions [12]. As illustrated in Figure 1, typically, these methods consist of three steps:
  • Computation of the gradient magnitude | I | and its orientation η , see Table 1, using a 3 × 3 templates [3], the first derivative of the filter (vertical and horizontal [4]), steerable Gaussian filters, oriented anisotropic Gaussian kernels or combination of two half Gaussian kernels.
  • Non-maximum suppression to obtain thin edges: the selected pixels are those having gradient magnitude at a local maximum along the gradient direction η , which is perpendicular to the edge orientation [4].
  • Thresholding of the thin contours to obtain an edge map.
Table 1 gives the different possibilities for gradient and its associated orientations involving several edge detection algorithms compared in this paper.
The final step remains a difficult stage in image processing, but it is a crucial operation for comparing several segmentation algorithms. Unfortunately, it is far from straightforward to choose an ideal threshold value to detect the edges of the desirable features. Usually, a threshold is fixed in a function of the objects’ contours, which must be visible, but this is not an objective segmentation for the evaluation. Otherwise, in edge detection, the hysteresis process uses the connectivity information of the pixels belonging to thin contours and thus remains a more elaborated method than binary thresholding [4]. To put it simply, this technique determines a contour image that has been thresholded at different levels (low: τ L and high: τ H ). The low threshold τ L determines which pixels are considered as edge points if at least one point higher than τ H exists in a contour chain where all the pixel values are also higher than τ L , as represented with a signal in Figure 2. Segmented real images using hysteresis thresholds are presented, later in this paper, in Figure 11. On the one hand, this algorithm is able to partly detect blurred edges of an object. On the other hand, the lower the thresholds are, the more the undesirable pixels are preserved and the problem remains that thresholds are fixed for both the segmentation and the evaluation.
In order to compare the quality of the results by different methods, they need to render binary edge maps. This normally requires a manual process of threshold selection aimed at maximizing the quality of the results by each of the contending methods. However, this assessment suffers from a major drawback: segmentations are compared using the (deliberately) chosen threshold, and this evaluation is very subjective and not reproducible. The aim is therefore to use the dissimilarity measures without any user intervention for an objective assessment. Finally, to consider a valuable edge detection assessment, the evaluation process should produce a result that correlates with the perceived quality of the edge image, which relies on human judgment [13,14,15]. In other words, a reliable edge map should characterize all the relevant structures of an image as closely as possible, without any disappearance of desired contours. In addition, a minimum of spurious pixels should be created by the edge detector, disturbing at the same time the visibility of the main/desired objects to be detected.
In this paper, a novel technique is presented to compare edge detection techniques by using hysteresis thresholds in a supervised way, consistent with the visual perception of a human being. Comparing a ground truth contour map with an ideal edge map, several assessments can be compared by varying the parameters of the hysteresis thresholds. This study shows the importance of more strongly penalizing false negative points than false positive points, leading to a new edge detection evaluation algorithm. The experiment using synthetic and real images demonstrated that the proposed method obtains contour maps closer to the ground truth without requiring tuning parameters, and objectively outperforms other assessment methods.

2. Supervised Measures for Image Contour Evaluations

In the last 40 years, several edge detectors have been developed for digital images. Depending on their applications, with different difficulties such as noise, blur or textures in images, the best edge detector must be selected for a given task. An edge detector therefore needs to be carefully tested and assessed to study the influence of the input parameters. The measurement process can be classified as either an unsupervised or a supervised evaluation criterion. The first class of methods exploits only the input contour image and gives a coherence score that qualifies the result given by the algorithm [15]. For example, two desirable qualities are measured in [16,17]: continuation and thinness of edges; for continuation, two connected pixels of a contour must have almost identical gradient direction ( η ). In addition, the connectivity, i.e., how contiguous and connected edge pixels are, is evaluated in [18]. These approaches obtain a segmentation that could generally be well interpreted in image processing tasks. Even though the segmentation includes continuous, thin and contiguous edges, it does not enable evaluation of whether the segmentation result is close to or far from a desired contour. A supervised evaluation criterion computes a dissimilarity measure between a segmentation result and a ground truth, generally obtained from synthetic data or expert judgement (i.e., manual segmentation). Pioneer works in edge detection assessments were directly applicable only to vertical edges [19,20] (examples for [19] are available in [21]). Another method [22] considers either vertical contours or closed forms, pixels of contour chains connected to the true contour. Contours inside or outside the closed form are treated differently. Alternatively, authors in [23] propose an edge detector performance evaluation method in the context of image compression according to a mean square difference between the reconstructed image and the original uncompressed one. Various supervised methods have been proposed in the literature to assess different shapes of edges [21,24,25,26], the majority are detailed in this study, and more precisely in an objective way using hysteresis thresholds. In this paper, the closer to 0 the score of the evaluation is, the more the segmentation is qualified as good. Several measures are presented with respect to this property. This work focusses on comparisons of supervised edge detection evaluations in an objective way and proposes a new measure, aimed at achieving an objective assessment.

2.1. Error Measures Involving Only Statistics

To assess an edge detector, the confusion matrix remains a cornerstone in boundary detection evaluation methods. Let G t be the reference contour map corresponding to ground truth and D c the detected contour map of an original image I. Comparing pixel per pixel G t and D c , the 1st criterion to be assessed is the common presence of edge/non-edge points. A basic evaluation is composed of statistics; to that end, G t and D c are combined. Afterwards, denoting | · | as the cardinality of a set, all points are divided into four sets (see Figure 3):
  • True Positive points (TPs), common points of G t and D c :           T P = G t D c ,
  • False Positive points (FPs), spurious detected edges of D c :         F P = ¬ G t D c ,
  • False Negative points (FNs), missing boundary points of D c :     F N = G t ¬ D c ,
  • True Negative points (TNs), common non-edge points:                T N = ¬ G t ¬ D c .
Figure 3 presents an example of G t and D c . Comparing these two images, there are 23 TPs, one FN and one FP. Other examples are presented in Figure 10 comparing different D c with the same G t .
Several edge detection evaluations involving confusion matrices are presented in Table 2. Computing only FPs and FNs or their sum enables a segmentation assessment to be performed and several edge detectors to be compared [12]. On the contrary, TPs are an indicator, as for Absotude Grading ( A G ) and S S R ; these two formulae are nearly the same, just a square root of difference, so they behave absolutely similarly. The Performance measure ( P m , also known as Jaccard coefficient [27]) or D i c e directly and simultaneously considers the three entities T P , F P and F N to assess a binary image. It decreases with improved quality of detection. Note that | G t | = T P + F N and that | D c | = T P + F P , so it is easy to observe that D i c e , A G , S S R and P m behave similarly when F N and/or F P increase (more details in [21]), as shown in the experimental results. Moreover, considering the original versions of D i c e and P m are widely utilized for medical images assessments, they are related by D i c e = 2 · | G t D c | | G t | + | D c | = 2 · | G t D c | | G t D c | / 1 + | G t D c | | G t D c | = 2 · P m P m + 1 . In addition, L o c a l i z a t i o n e r r o r ( P E ) and M i s c l a s s i f i c a t i o n E r r o r ( M E ) represent the same measurement. Indeed, as | I | = T P + T N + F P + F N , the M E measure can be rewritten as:
M E G t , D c = 1 T P + T N T N + F N + T P + F P = T N + F N + T P + F P T N + F N + T P + F P T P + T N T N + F N + T P + F P = P E .
Another way to display evaluations is to create Receiver Operating Characteristic (ROC) [40] curves, involving True Positive Rates ( T P R ) and False Positive Rates ( F P R ):
T P R = T P T P + F N and F P R = F P F P + T N .
Then, T P R is plotted versus (vs.) F P R by varying the threshold of the detector (see Figure 4 (Section 4 details filters)). The closer the area under the curve is to 1, the better the segmentation, and an area of 1 represents a perfect edge detection. Finally, the score higher than and furthest from the diagonal (i.e., line from (0, 0) to (1, 1)) of ROC is considered as the best segmentation (here, H-K in (e) in Figure 4). However, the score of S F 5 is poor, but the segmentation seems better than Canny, Sobel and H-K for this example. Thus, any edge detectors can be called the best by simply making small changes G t or the parameter set [41]. As TNs are the majority set of pixels, Precision–Recall (PR) [39,42] does not take into account the T N value by substituting F P R with a precision variable: P r e c = T P T P + F P . By using both T P R and P r e c entities, PR curves quantify more precisely than ROC curves the compromise between under-detection ( T P R value) and over-detection ( P r e c value) (see Figure 4g). An example of PR curve is available in Figure 4f. The best segmentation is tied to the curve point closest to the point situated in (1, 1). As shown in Figure 4h,j, results of Sobel and H-K for PR are similar to those obtained with ROC. These evaluation types are effective for G t having precise locations of edges, as in synthetic images [14,43], since a displacement of G t or D c points strongly penalizes the segmentation.
Derived from T P R and F P R , the three measures Φ , χ 2 and F α (detailed in Table 2) are frequently used. The complement of these measures translates a value close to 0 as a good segmentation. Among these three measures, F α remains the most stable because it does not consider the TNs, which are dominant in edge maps (see [14]). Indeed, taking into consideration T N in Φ and χ 2 influences solely the measurement (as is the case in huge images). These measures evaluate the comparison of two edge images, pixel per pixel, tending to severely penalize an (even slightly) misplaced contour, as illustrated in Figure 8.
Consequently, some evaluations resulting from the confusion matrix recommend incorporating spatial tolerance. Tolerating a distance from the true contour and integrating several TPs for one detected contour can penalize efficient edge detection methods, or, on the contrary, benefit poor ones (especially for corners or small objects). The assessment should therefore penalize a misplaced edge point proportionally to the distance from its true location. More details are given in [21,26], some examples and comparisons are shown in [21].

2.2. Assessments Involving Spacial Areas Around Edges

2.2.1. The Performance Value P v r

To judge the quality of segmentation results and the performance of algorithms, the performance value P v r in [44] combines four features: location ( L ), matching ( M ), unmatching ( U ) and spurious ( S ). In this approach, D c pixels are assimilated as TP when they belong to a disc of radius r centered on a pixel of G t , as illustrated in Figure 5; this set of pixels is denoted T P r . Thus, F N r represents the set of pixels of G t located at a distance (In our tests, the Euclidean distance is used, and the next section exposes different measures using distances of misplaced pixels.) higher than r of D c and, conversely, F P r the set of points of D c at a distance higher than r of G t . The location criteria depends on the sum of the distance between each point of T P r and G t , denoted by: p T P r d G t ( p ) . Hence, the four criteria are computed as follows:
L = p T P r d G t ( p ) | T P r | · | G t | , M = | T P r | | D c | , U = | G t | | T P r | | G t | = F N r | G t | , S = | D c | | T P r | | D c | = F P r | D c | .
Finally, the performance value P v r is obtained by:
P v r ( G t , D c ) = 1 M M + L + U + S .
The main drawback of P v r is that the term M M + L + U + S can obtain negative or huge values. This is explainable when r 1 , we can obtain | G t | < | T P r | (typically when | G t | < | D c | ). Thus, U < 0 ; so if | U | > M + L + S , P v r could be negative, as illustrated in Figure 5. Finally, when U < 0 and M + L + U + S 0 , P v r tends to ± infinity (see experiments). Moreover, as illustrated in Figure 8, P v r > 1 obtains the same measurement for two different shapes because FPs are close to the desired contour, which is not desirable for the evaluation of small objects segmentation. Note that, when r 1 , L , and P v r is equivalent to P m , since:
P v r 1 ( G t , D c ) = 1 T P | D c | · F N | G t | + F P + T P .

2.2.2. The Quality Measure R

In [45], a mixed measure of quality R W is presented. This evaluation depends on the number of FPs and FNs and the calculus focuses on a window W for each mistake (FP or FN). For each point of F N or of F P , to estimate the evaluation measure R W , several variables are computed:
  • n b , the number of FPs in W, minus the central pixel: n b = p F P W p p c , with p c = 1 if the central pixel is a FP point, or 0 otherwise,
  • n h , the number of FNs in W, minus the central pixel: n h = p F N W p p c , with p c = 1 if the central pixel is a FN point, or 0 otherwise.
  • n e , the number of edge points belonging to G t in W: n e = p G t W p ,
  • n b t , the number of FPs in direct contact (i.e., 8-connexity) with the central pixel: for a pixel p, if  p F N , n b t = p F N W 3 × 3 p 1 , with W 3 × 3 a window of size 3 × 3 centered on p,
  • n h t , the number of FNs in direct contact with the central pixel: for a pixel p, if p F P , thus n b t = p F P W 3 × 3 p 1 , with W 3 × 3 a window of size 3 × 3 centered on p.
Then, the final expression of R W is given by:
R W ( G t , D c ) = K · w · p F P 1 + b · n b 1 + p · n e + i b h · n h t + p F N 1 + h · n h 1 + c E u l e r · i h b · n b t .
Table 3 contains the (rounded) coefficients determined by a least square adjustment [45]. The computation of R W depends on the number of FP(s) and the number of FN(s) in a local window around each mistake, but not on the distances of misplaced points, as explained in the next section. Figure 6 exposes an example of the error images I R , representing R W with the coefficients available in Table 3.

2.2.3. The Failure Measure F M

The f a i l u r e m e a s u r e [46] is an extension of [20] (see beginning of Section 2). These evaluation computes four criteria, taking into account a multiple detection zone ( M D ). The detection zone of the ideal image can be represented as a dilation of G t , creating a rough edge, as illustrated in Figure 7. Then, the criteria are as follows: (1) False negative ( F N F M ), (2) False positive ( F P F M ), (3) Multiple detection ( D M F M ) and (4) Localization ( L O C F M ). They are computed by:
  • F N F M = max ( 0 , | G t | T P M D ) | G t | , where T P M D represents the number of points of D c in M D (see Figure 7 and Figure 8), green pixels,
  • F P F M = F P M D | I | | M D | , with F P M D the number of points of D c outside M D and | M D | denoting the number of pixels of M D , see Figure 7 and Figure 8, green and blue pixels,
  • D M F M = T P M D T P | M D | | G t | , T P representing the number of TPs (see above),
  • L O C F M = 1 | G t | · C · p G t max ( C , d ( p , D c ) ) , where C is a constant ( C = 5 in our experiments) and d ( p , D c ) represents the Euclidean distance between p and D c (see next section). In [20], L O C F M represents the number of rows containing a point around the vertical edge.
On end, the f a i l u r e m e a s u r e ( F M ) is defined as:
F M ( G t , D c ) = α · F N F M + β · F P F M + γ · D M F M + δ · L O C F M ,
with ( α , β , γ , δ ) four positive coefficients such that α + β + γ + δ = 1 ; in the experiments: α = 0.4 , β = 0.4 , γ = 0.1 and δ = 0.1 . Unfortunately, due to the multiple detection zone, F M  behaves like P v for the evaluation of small object segmentation, as shown in Figure 8, and F M obtains the same measurement for two different shapes.

2.3. Assessment Involving Distances of Misplaced Pixels

A reference-based edge map quality measure requires that a displaced edge should be penalized in function not only of FPs and/or FNs, but also of the distance from the position where it should be located. Table 4 reviews the most relevant measures involving distances. Thus, for a pixel p belonging to the desired contour D c , d G t ( p ) represents the minimal Euclidian distance between p and G t . If p belongs to the ground truth G t , d D c ( p ) is the minimal distance between p and D c , and Figure 9a shows the difference between d G t ( p ) and d D c ( p ) . Mathematically, denoting ( x p , y p ) and ( x t , y t ) , the pixel coordinates of two points p and t, respectively; thus, d G t ( p ) and d D c ( p ) are described by:
for p D c : d G t ( p ) = Inf ( x p x t ) 2 + ( y p y t ) 2 , t G t , for p G t : d D c ( p ) = Inf ( x p x t ) 2 + ( y p y t ) 2 , t D c .
These distance functions refer to the Euclidean distance. Figure 9d illustrates an example of d G t ( p ) and d D c ( p ) .
On the one hand, some distance measures are specified in the evaluation of over-segmentation (i.e., presence of FPs), for example: Υ , D k , Θ and Γ ; others are presented and detailed in [21,24]. On the other hand, the Ω measure assesses an edge detection by computing only under-segmentation (FNs). Other edge detection evaluation measures consider both distances of FPs and FNs [14]. A perfect segmentation using an over-segmentation measure could be an image including no edge points and an image having the most undesirable edge points (FPs) concerning under-segmentation evaluations [60], as shown in Figure 10 and Figure 11. In addition, another limitation of only over- and under-segmentation evaluations are that several binary images can produce the same result (Figure 8). Therefore, as demonstrated in [14], a complete and optimum edge detection evaluation measure should combine assessments of both over- and under-segmentation, as f 2 d 6 , S k , R D E k , Ψ and λ , illustrated in Figure 8.
Among the distance measures between two contours, one of the most popular descriptors is named the Figure of Merit ( F o M ). This distance measure has an advantage because it ranges from 0 to 1, where 0 corresponds to a perfect segmentation [47]. Nonetheless, for F o M , the distance of the FNs is not recorded and are strongly penalized as statistic measures:
F o M G t , D c = 1 1 max G t , D c · p T P 1 1 + κ · d G t 2 ( p ) 1 max G t , D c · p F P 1 1 + κ · d G t 2 ( p ) = 1 T P max G t , D c 1 max G t , D c · p F P 1 1 + κ · d G t 2 ( p ) .
For example, in Figure 10, F o M ( G t , C ) > F o M ( G t , M ) , whereas M contains both FPs and FNs and C only FNs. Furthermore, for the extreme cases, knowing that T P = | G t | F N , the F o M measures takes the following values:
  • if F P = 0 : F o M G t , D c = 1 T P | G t | = 1 | G t | F N | G t | ,
  • if F N = 0 : F o M G t , D c = 1 T P D c 1 D c · p F P 1 1 + κ · d G t 2 ( p ) .
When F N >0 and F P are constant, it behaves like matrix-based error assessments (Figure 10). Moreover, for F P >0, the F o M penalizes over-detection very lightly compared to under-detection. Several evaluation measures are derived from F o M : F, d 4 , D p , M F o M and S F o M . Contrary to F o M , the F measure computes the distances of FNs but not of the FPs, so F behaves inversely to F o M , it can be rewritten as:
F G t , D c = 1 1 G t + β · F P · p T P 1 1 + κ · d D c 2 ( p ) 1 G t + β · F P · p F N 1 1 + κ · d D c 2 ( p ) = 1 T P G t + β · F P 1 G t + β · F P · p F N 1 1 + κ · d D c 2 ( p ) .
Therefore, for the extreme cases, the F measures takes the following values:
  • if F P = 0 : F G t , D c = 1 T P G t 1 G t · p F N 1 1 + κ · d D c 2 ( p ) ,
  • if F N = 0 : F G t , D c = 1 T P G t + β · F P .
In addition, the d 4 measure depends particularly on T P , F P , F N and ≈1/4 on F o M , but d 4 penalizes FNs like the F o M measure; it is a close idea to the F M measure (Section 2.2). Otherwise, S F o M and M F o M take into account both distances of FNs and FPs, so they can compute a global evaluation of a contour image. However, M F o M does not consider FPs and FNs at the same time, contrary to S F o M . Another way to compute a global measure is presented in [50] with the edge map quality measure D p . The right term computes the distances of the FNs between the closest correctly detected edge pixel, i.e., G t D c , D p can be rewritten as:
D p ( G t , D c ) F P p F P 1 1 + κ · d G t 2 ( p ) 2 · I 2 · G t + F N p F N 1 1 + κ · d G t 2 ( p ) 2 · G t .
Finally, D p is more sensitive to FNs than FPs because of the huge coefficient 1 | I | | G t | .
A second measure widely computed in matching techniques is represented by the Hausdorff distance H, which measures the mismatch of two sets of points [52]. This measure is useful in object recognition, the algorithm aims to minimize H, which measures the mismatch of two shapes [61,62]. This max-min distance could be strongly deviated by only one pixel that can be positioned sufficiently far from the pattern (Figure 10). There are several enhancements of the Hausdorff distance presented in [24,63,64]. Furthermore, f 2 d 6 and D k are often called “Modified Hausdorff Distance” (abbreviated M H D ) in the literature. As another example, one idea to improve the measure is to compute H with a proportion of the maximum distances; let us note H 5 % —this measure for 5% of the values [52]. Nevertheless, as pointed out in [24], an average distance from the edge pixels in the candidate image to those in the ground truth is more appropriate, like S k , R D E k or Ψ . Thus, the score of the f 2 d 6 corresponds to the maximum between the over- and the under-segmentation (depending on 1 | D c | and 1 | G t | , respectively), whereas the values obtained by S k represents their mean. Moreover, S k takes small values in the presence of low level of outliers, whereas the score becomes large as the level of mistaken points increases [24,26] but is sensitive to remote misplaced points as presented in [21]. On the other hand, the Relative Distance Error ( R D E k ) computes both the over- and the under-segmentation errors separately, with the weights 1 | D c | and 1 | G t | , respectively. Otherwise, derived from H, the Delta Metric ( Δ k ) [57] intends to estimate the dissimilarity between each element of two binary images, but is highly sensitive to distances of misplaced points [14,21]. All of these edge detection evaluation measures are reviewed in [21] with their advantages and disadvantages (excepted R D E k ), and, as concluded in [21,25], a complete and optimum edge detection evaluation measure should combine assessments of both over- and under-segmentation, as f 2 d 6 , S k , H 15 % , R D E k and Ψ .
On another note, the Edge Mismatch Measure ( E M M ) depending on TPs and both d D c and d G t . In [36], this measure is combined with others (including M E and D k ) in order to compare several thresholding methods. Indeed, δ D c / G t ( p ) is a threshold distance function penalizing high distances (exceeding a value M d i s t ) and E M M is represented as follows:
E M M ( G t , D c ) = 1 T P T P + ω p F N δ D c ( p ) + ϵ · p F P δ G t ( p )
with δ D c and δ G t two cost functions of d D c and d G t respectively discarding/penalizing outliers [36]:
δ D c ( p ) = d D c ( p ) , if d D c ( p ) < M d i s t D m a x , otherwise , and δ G t ( p ) = d G t ( p ) , if d G t ( p ) < M d i s t D m a x , otherwise .
Thus, ω is the penalty weighting distance measures δ D c and δ G t , whereas ϵ represents a weight for distances of FPs only. For this purpose, the set of parameters are suggested as follows:
  • M d i s t = 0.025 · | I | ,
  • D m a x = | I | 10 ,
  • ω = 10 | I | ,
  • ϵ = 2 .
Note that the suggested parameters depend on | I | , the total number of pixels in I. Moreover, E M M computes a score different from 1 if there exists at least one TP (cf. Figure 8). Finally, when the E M M score is close to 0, the segmentation is qualified as acceptable, whereas a score close to 1 corresponds to a poor edge detection.

3. A New Objective Edge Detection Assessment Measure

3.1. Influence of the Penalization of False Negative Points in Edge Detection Evaluation

Several edge detection measures have been presented above. Clearly, taking into account both FP and FN distances is more objective for the assessment. However, there are two main problems concerning the edge detection measures involving distances. First, a single (or a few) FP point(s) at a sufficiently high distance may penalize a good detection (see Figure 12c). This is a well known problem concerning the Hausdorff distance. Thus, best scores for each measure obtained in an objective way (cf. next section) are not necessarily tied to the most efficient detector. Secondly, the edge maps associated with these scores lack many desired contours, because distances of FPs strongly penalize edge detectors evaluated by the majority of these measures. On the contrary, distances of FN points are neither recorded (as over-segmentation measures), nor penalized enough (cf. Figure 12b). In other words, FNs are, generally, as penalized as FPs. Moreover, FNs are often close to detected edges (TPs or FPs close to G t ), most error measures involving distances do not consider this particularity because p F N d D c are less important than p F P d G t . Note that R D E k computes p F N d D c and p F P d G t separately. In [59], a measure of the edge detection assessment is developed: it is denoted λ and improves the segmentation measure Ψ (see formulas in Table 4). The λ measure penalizes highly FNs compared to FPs (as a function of their mistake distances), depending on the number of TPs. Typically, contours of desired objects are in the middle of the image, but rarely on the periphery. Thus, using or f 2 d 6 , S k , Δ k , or Ψ , a missing edge in the image remains insufficiently penalized contrary to the distance of FPs, which could be too high, as presented in Figure 13, contrary to λ . Another example, in Figure 10, Ψ ( G t , C ) < Ψ ( G t , T ) , whereas C should be more penalized because of FNs that do not enable the object to be identified. The more FNs are present in D c , the more D c must be penalized as a function of d G t , because the desirable object becomes unrecognizable, as D c in Figure 11c. In addition, D c should be penalized as a function of d G t , of the F N number, as stated above. For λ , the term influencing the penalization of FN distances can be rewritten as: | G t | 2 T P 2 = F N + T P T P 2 = 1 + F N T P 2 1 , ensuring a stronger penalty for d G t 2 , compared to d D c 2 . The min function avoids the multiplication by infinity when T P = 0 . When | G t | = T P , λ is equivalent to Ψ and Γ (see Figure 10, image T). In addition, compared to Ψ , λ  penalizes more D c having FNs, than D c with only FPs, as illustrated in Figure 10 (images C and T). Finally, the weight | G t | 2 T P 2 tunes the λ measure by considering an edge map of better quality when FNs points are localized close to the desired contours D c , the red dot curve in Figure 14 represents this weight function. Hence, the λ function is able to assess images that are not too large, as in Figure 10, Figure 12 and Figure 13; however, the penalization is not enough for larger images. Indeed, the main difficulty remains the F N + F P coefficient to the left of λ ; as a result, the image in Figure 11a is considered by this measure as the best one. The solution is to separate the two entities F N and F P and insert them directly inside the root square of the measure, firstly to modulate the FPs distances and secondly to weight the FN distances. Therefore, the new edge evaluation assessment formula is given by:
Ξ ( G t , D c ) = 1 | G t | · F P · p D c d G t 2 ( p ) + f F N · p G t d D c 2 ( p )
with
f T P = l o g ( F N ) · e F N , if T P = 0 , i . e . , | G t | = F N l o g F N + 1 · e | G t | T P , elsewhere .
The f function influencing the penalization of FN distances ensures a strong penalty for d D c 2 , compared to d G t 2 (see blue curves in Figure 14). There exist several f functions than may effectively accomplish the purpose. When F N = 0 , f T P = 0 , and only the FP distances are recorded, pondered by the number of FPs. Otherwise, if T P = 0 , so | G t | = F N , thus f T P = l o g ( F N ) · e F N to avoid a division by 0, and l o g ( F N ) · e F N > l o g F N + 1 · e | G t | T P . Finally, by separating the two weights for d D c 2 and d G t 2 penalizes D c images containing FPs and/or D c images with missing edges (FNs).
The next subsection details the way to evaluate an edge detector in an objective way. Results presented in this paper show the importance to penalize false negative points more severely than false positive points because the desired objects are not always completely visible using ill-suited evaluation measure, and Ξ provides a reliable edge detection assessment.

3.2. Minimum of the Measure and Ground Truth Edge Image

Dissimilarity measures are used to assess the divergence of binary images. Instead of manually choosing a threshold to obtain a binary image (see Figure 3 in [14]), the purpose is to compute the minimal value of a dissimilarity measure by varying the thresholds (double loop: loop over τ L and loop over τ H ) of the thin edges obtained by filtering gradient computations (see Table 1). Compared to a ground truth contour map, the ideal edge map for a measure corresponds to the desired contour at which the evaluation obtains the minimum score for the considered measure among the thresholded (binary) images. Theoretically, this score corresponds to the thresholds at which the edge detection represents the best edge map, compared to the ground truth contour map [14,25,46]. Figure 15 illustrates the choice of a contour map as a function of τ L and τ H . Algorithm 1 represents this argmin function and summarizes the different steps to compute an ideal edge map concerning a chosen measure.
Algorithm 1 Calculates the minimum score and the best edge map of a given measure M e a s
Require:   | I |         : normalized thin gradient image
Require:  G t            : Ground Truth edge image
Require:  H y s t e r    : hysteresis threshold function
Require:  M e a s       : Measure computing a dissimilarity score between G t and a desired contour D c
   s t e p τ = 0 . 01               % step for the loops on thresholds 
   s c o r e L = r e a l m a x    % the largest finite floating-point number  
  for τ H = 0 : s t e p τ : 1 do
    for τ L = 0 : s t e p τ : 1 do
      if τ H τ L then
         D c H y s t e r ( | I | , τ L , τ H )
         s c o r e M e a s ( G t , D c )
        if s c o r e L s c o r e then
           s c o r e L s c o r e L    % ideal score  
           I d e a l D c D c          % ideal edge map
        end if 
      end if 
    end for 
  end for
Since low thresholds lead to heavy over-segmentation and high thresholds may create numerous false-negative pixels, the minimum score of an edge detection evaluation should be a compromise between under- and over-segmentation (detailed and illustrated in [14]).
As demonstrated in [14], the significance of the choice of ground truth map influences the dissimilarity evaluations. Indeed, if not reliable [43], a ground truth contour map that is inaccurate in terms of localization penalizes precise edge detectors and/or advantages the rough algorithms as edge maps presented in [13,15]. For these reasons, the ground truth edge map concerning the real image in our experiments is built semi-automatically, as detailed in [14].

4. Experimental Results

The aim of the experiments is to obtain the best edge map in a supervised way. The importance of an assessment penalizing false negative points more severely compared to false positive points has been shown above. In order to study the performance of the edge detection evaluation measures, the hysteresis thresholds vary and the minimum score of the studied measure corresponds to the best edge map (cf. Figure 15). The thin edges of real noisy images are computed by nine filtering edge detectors:
  • Sobel [3],
  • Shen [5],
  • Bourennane [7],
  • Deriche [6],
  • Canny [4],
  • Steerable filter of order 1 ( S F 1 ) [9],
  • Steerable filter of order 5 ( S F 5 ) [10],
  • Anisotropic Gaussian Kernels (AGK) [11,65,66],
  • Half Gaussian Kernels (H-K) [12,56].
The kernels of these methods are size-adaptable, except for the Sobel operator that corresponds to a 3 × 3 mask. The parameters of the filters are chosen to keep the same spatial support for the derivative information, e.g., σ = 1.5 for Gaussians (details of these filters are available in [56]). Ground truth images ( G t ) are shown in Figure 16, whereas corrupted and original images are presented in Figure 17. The scores of the different measures are recorded by varying the thresholds of the normalized thin edges computed by an edge detector and plotted as a function of the noise level in the original image, as presented in Figure 18 and Figure 19. A plotted curve should increase monotonously with noise level (Gaussian noise), represented by Peak Signal to Noise Ratio (PSNR) values (from 17 dB to 10 dB). Among all the edge detectors, box (Sobel [3]) and exponential (Shen [5], Bourennane [7] filters do not delocalize contour points [67], whereas they are sensitive to noise (i.e., addition of FPs). The Deriche [6] and Gaussian filters [4] are less sensitive to noise, but suffer from rounding corners and junctions (see [67,68]) as the oriented filters S F 1 [9], S F 5 [10] and AGK [11], but the more the 2D filter is elongated, the more the segmentation remains robust against noise. Finally, as a compromise, H-K correctly detects contour points that have corners and is robust against noise [12]. Consequently, the scores of the evaluation measures for the first three filters must be lower than the three last ones, and, Canny, Deriche and S F 1 scores must be situated between these two sets of assessments. Furthermore, as S F 5 , AGK and H-K are less sensitive to noise than other filters, the ideal segmented image for these three algorithms should be visually closer to G t . The presented segmentations correspond to the original image for a PSNR = 14 dB. Therefore, on the one hand, considered segmentations must be tied to the robustness of the detector. On the other hand, the scores must increase monotonously, with an increasing order as a function of the edge detector quality. Note that the matlab code of F o M , D k , S k and Δ k measures are available at http://kermitimagetoolkit.net/library/code/. The matlab code of several other measures are available on MathWorks: https://fr.mathworks.com/matlabcentral/fileexchange/63326-objective-supervised-edge-detection-evaluation-by-varying-thresholds-of-the-thin-edges.
Firstly, the segmented images tied to corrupted images with PSNR = 14 dB, representing the best edge quality map for 28 different measures, are presented in Figure 20, Figure 21, Figure 22, Figure 23, Figure 24, Figure 25, Figure 26, Figure 27, Figure 28, Figure 29, Figure 30, Figure 31, Figure 32, Figure 33, Figure 34, Figure 35, Figure 36, Figure 37, Figure 38, Figure 39, Figure 40, Figure 41, Figure 42, Figure 43, Figure 44, Figure 45, Figure 46 and Figure 47. The results concerning over- and under-segmentation measures (cf. Section 2.3) are not reported because the score will always attain 0 for the best edge map which are either full of FPs or devoid of any contour point [14]. The edge map obtained using the Sobel filter is complicated; indeed, this filter is very sensitive to the noise in the image, so only few edge points will be correctly detected, the rest being FPs. Furthermore, thin edges (before thresholding) obtained using Shen, Bourennane and Deriche filters are not reliable, and it is difficult to choose/compute correct thresholds in order to visualize continuous objects’ contours. Segmentations obtained by D i c e , P m and F α are overall visible, with a little too many FPs, except for AGK and H-K, which are correctly segmented. On the contrary, contour points concerning S S R and χ 2 are less corrupted by FPs, but true edges are missing; in addition, edge maps concerning P E are worse. Edge maps tied to Φ , F M W and D p are hugely corrupted by FPs, since most of the object contours remain unidentifiable. Concerning P v , either edges are missing, when r = 2 , or too many FPs appear, when r = 4 . Edge maps obtained by R W evaluation measures are adequate, even though object contours are not really visible concerning Shen, Bourennane and Deriche filters and some spurious pixels appear concerning AGK and H-K (cf. parkingmeter image). The Hausdorff distance H and Δ k measures are not reliable because edge maps tied to these evaluations are either too noisy, or most edges are missing (except for H-K). The edge maps associated with H 5 % , f 2 d 6 , S k = 2 k and ψ are similar: not too many FPs, but edges with Shen, Bourennane and Deriche filters are not continuous. However, edges obtained using f 2 d 6 are too noisy with AGK and H-K (cf. “parkingmeter” image), and the same remark applies to S k = 1 k for AGK. Concerning S k , note that, when k = 1 , edges are more easily visible than using k = 2 because the distance measure score expands rapidly for a missing point far from its true position (demonstrated in [21]). For image 109, edges obtained by E M M are not really continuous with the Shen, Bourennane, Deriche and Canny filters, whereas spurious pixels appears for the edges of the “parkingmeter” image. The edge maps obtained using minimum score of F o M are heavily corrupted by continuous FPs, like hanging objects. This phenomenon is always present, but less pronounced, with d 4 . Edge maps are too corrupted by FPs with M F o M and S F o M , even though objects are visible, whereas FPs remains less present using F and λ . The segmentations tied to R D E are reliable, not too many FPs, although some edges are missing. Lastly, the edges maps using the proposed measure Ξ are not corrupted by noise, the objects are visible, even with the Shen, Bourennane and Deriche filters. In addition, edge maps for Canny and S F 1 are particularly well segmented.
Secondly, the plotted curves available in Figure 18 and Figure 19 evolve as a function of the noise level (Gaussian noise). The noise level is represented by PSNR values: from 17 dB to 10 dB. Consequently, the measure scores and the noise level must increase simultaneously. Moreover, scores of the evaluation measures associated with the Sobel filter, which is sensitive to noise, must be higher than other measures; scores concerning Shen and Bourennane filters must be situated just bellow. Finally, measure scores tied to S F 5 , AGK and H-K must be plotted at the bottom, and, scores associated with Canny, Deriche and S F 1 filters must be situated above, but below the Shen and Bourennane filters. Now, scores of D i c e , P m , S S R , χ 2 and F α measures increase monotonously, but these scores are not consistent with the computed edge maps. Indeed, considering the segmented images presented with PSNR = 14 dB, scores concerning Canny and S F 1 filters are better than H-K, whereas the H-K segmentation is of higher quality than others (continuous contours, less spurious pixels). Concerning P E , in particular, this measure qualifies the Sobel, Shen and Bourennane filters better than H-K. Similarly, Φ , F M W , D p and Ψ qualify H-K and AGK as the worse edge detectors. Concerning P v , either curves are confused, when r = 2 , or scores are negative, when r = 4 . By contrast, H, H 5 % and Δ k scores have a random behavior, even though H 5 % seems better, but not reliable (see H-K or Sobel scores as examples). The curves for F o M , F, d 4 and M F o M are mixed and confused, F qualifies Shen and Bourennane filters as the best edge detectors, whereas H-K and AGK are qualified as the worst. The Deriche filter appears as the best edge detector for d 4 , although the segmentation using H-K is clearly better. Curves are mixed using S F o M for the “parkingmeter” image. These plotted scores are consistent with the images of segmentation, which are heavily corrupted by FPs. No filter can be really qualified as better than the others. It is also a similar case for λ , where the scores are confused, except with the Sobel filter. Concerning S k = 2 k , the plotted scores remain unreliable, cf. AGK scores. When k = 1 , S k scores evolve properly, even though S F 5 is penalized as strongly as the Canny filter. Therefore, the measures having the correct evolution with the correct filter qualification are E M M , R W , R D E , f 2 d 6 and S k = 1 k . The scores obtained by Ξ are presented in Figure 48 and Figure 49, where FP and FN distances are also reported. Although these distances do not evolve monotonously, the final score remains monotonous and the qualifications of the filters are reliable. Actually, the weights concerning FN distances allow a reliable final computation of Ξ scores. Finally, the results gathering reliability of the segmentation, curve evolution and filter qualification for each edge detection evaluation are summarized in Table 5.

5. Conclusions

This study presents a survey of supervised edge detection evaluation methods. Several techniques are based on the number of false positive, false negative, true positive and/or true negative points. Other methods strongly penalize misplaced points when they are outside a window centered on a true point. In addition, many approaches compute the distance from the position where a contour point should be located. Most of these edge detection assessment methods are presented here, with many examples, exposing the drawbacks for different comparisons of edges with different shapes. Measures involving only statistics fail to assess objectively when there are no common edge points between the ground truth ( G t ) and the desired contour ( D c ). On the contrary, assessments involving spatial areas around edges (i.e., windows around a point) remain unreliable if several points are detected for one contour point. Moreover, these techniques depend strongly on the window size, which enables misplaced points outside the considered window to be severely penalized. Among assessments involving spacial areas around edges, only the R W measure is suitable. Therefore, assessment involving distances of the misplaced pixels can evaluate a desired edge as a function of the distances between the ground truth edges and each point of D c . There exist different implementations to assess edges using distances (Note that different strategies exist containing some operators other than confusion matrices of distances to assess edge detectors, they are referenced in [69].). On the one hand, some methods record only distances of false positive points, or only distances of false negative points. On the other hand, some assessment techniques are based on both distances of false positives (FPs) and false negative points (FNs). Among the more prominent measures, the Figure of Merit ( F o M ) remains the most widely used. The main drawback of this technique is that is does not consider distances of false negative points, i.e., false negative points are strongly penalized without considering their distances; consequently, two different desired contours can obtain the same evaluation, even if one of them if visually closer to the true edge. Consequently, several edge evaluation methods are derived from the Hausdorff distance, they compute both distances of FPs and FNs. The main differences between these edge detection evaluation measures are the weights for the FP and/or FN distances and the power tied to the distance computations. As FNs are often close to detected edges (TPs or FPs close to G t ), most error measures involving distances do not consider this particularity and are not sufficiently penalized. Distances of FPs strongly penalize edge detectors evaluated by the majority of these measures. Only R D E k computes the distances of FPs and FNs separately.
In order to objectively compare all these supervised edge detection assessment methods in an objective way, based on the theory of the dissimilarity evaluation measures, the objective evaluation assessed nine 1st-order edge detectors involving the minimum score of the considered measures by varying the parameters of the hysteresis. The segmentation that obtains the minimum score of a measure is considered as the best one. The scores of the different measures and different edge detectors are recorded and plotted as a function of the noise level in the original image. A plotted curve must increase monotonously with the noise level (Gaussian noise), represented by PSNR values (from 17 dB to 10 dB). It is proved that some edge detectors are better than others. The experiments show the importance of the growing increase of the noise level: a given edge evaluation measure can qualify an edge detector as low for a given noise level, whereas, for a higher noise level, the same edge detector obtains a better score. Consequently, mixing the results of curve evolution (monotonic or not), filter qualification (poor edge detector penalized stronger than robust edge detector) and the obtained edge map tied to the minimum score of a considered measure, a credible evaluation is obtained concerning the studied measures. These experiments exhibit the importance of dissociating both distances of FPs and FNs. A minimum of measures involving only statistics can be tied to correct segmented images, but the evolution of the scores is not reliable as a function of the edge detector robustness. On the contrary, edge maps are visually closer to the ground truth by considering the distance of false negative points tuned by a weighting. The same applies to the score evolution, and remains significant for edge detector qualification. The results gathering reliability of the segmentation, curve evolution and filter qualification for each edge detection evaluation are summarized in Table 5. Thus, the edge detection evaluations that are objectively suitable are the Relative Distance Error ( R D E k = 1 ) and the new proposed measure Ξ . The main difference between R D E and Ξ is that R D E separates the computations of distances of FPs and FNs as a function of the number of points in D c and G t , respectively, whereas Ξ gives a strong weight concerning distances of FNs. This weight depends on the number of false negative points: the more there are, the more the segmentation is penalized. This enables an edge map to be obtained objectively containing the main structures, similar to the ground truth, concerning a reliable edge detector, and a contour map where the main structures of the image are noticeable. Finally, the computation of the minimum score of a measure does not require tuning parameters, which is a huge advantage. The open problem remains the normalization of the distance measures, which could qualify a good segmentation and a poor edge detection close to 0 and 1, respectively. Another open problem concerns the choice of the hysteresis thresholds in the absence of a ground truth edge map, where the selection of thresholds may be learned thanks to a reliable edge detection evaluation measure.

Author Contributions

The majority of the measures and edge detectors were coded by Baptiste Magnier in MATLAB. The objective comparison of filtering gradient computations using hysteresis threshold experiments was carried out by Hassan Abdulrahman. The figures were created by Baptiste Magnier. Finally, the text was written by Baptiste Magnier.

Acknowledgments

Special thanks to Adam Clark for the English enhancement.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

| I | Gradient magnitude of an image I
η gradient orientation
T P set of True Positive pixels
F P set of False Positive pixels
F N set of False Negative pixels
T N set of True Negative pixels
G t Ground truth contour map
D c Detected contour map
D i c e D i c e measure
P m Performance measure
S S R Segmentation Success Ratio
P E Localization-error
M E Misclassiffication Error
Φ Φ measure
χ 2 χ 2 measure
F α F α measure, with α [ 0 , 1 ]
P v Performance value
R W Quality Measure R W focussing on a window W
F M Failure measure F M
F o M Pratt’s Figure of Merit
FFigure of Merit revisited
d 4 Combination of Figure of Merit and statistics
D p Edge map quality measure
S F o M Symmetric Figure of Merit
M F o M Maximum Figure of Merit
Υ Yasnoff measure
HHausdorff distance
f 2 d 6 Maximum distance measure
D k Distance to ground truth, with k a real positive
Θ Over-segmentation measure Θ
Ω Under-segmentation measure Ω
R D E k R e l a t i v e D i s t a n c e E r r o r , with k a real positive
S k Symmetric distance measure, with k a real positive
Δ k Baddeley’s Delta Metric
Γ Over-segmentation measure Γ
Ψ Complete distance measure
λ λ measure
Ξ Ξ measure
d G t ( p ) minimal Euclidian distance between a pixel p and G t
d D c ( p ) minimal Euclidian distance between a pixel p and D c
SobelSobel edge detection method
ShenShen edge detection method
BourennaneBourennane edge detection method
DericheDeriche edge detection method
CannyCanny edge detection method
S F 1 Steerable filter of order 1
S F 5 Steerable filter of order 5
AGKAnisotropic Gaussian Kernels
H-KHalf Gaussian Kernels

References

  1. Ziou, D.; Tabbone, S. Edge detection techniques: An overview. Int. J. on Patt. Rec. and Image Anal. 1998, 8, 537–559. [Google Scholar]
  2. Arbelaez, P.; Maire, M.; Fowlkes, C.; Malik, J. Contour detection and hierarchical image segmentation. IEEE TPAMI 2011, 33, 898–916. [Google Scholar] [CrossRef] [PubMed]
  3. Sobel, I. Camera Models and Machine Perception. Ph.D. Thesis, Stanford University, Stanford, CA, USA, 1970. [Google Scholar]
  4. Canny, J. A computational approach to edge detection. IEEE TPAMI 1986, 6, 679–698. [Google Scholar] [CrossRef]
  5. Shen, J.; Castan, S. An optimal linear operator for step edge detection. CVGIP 1992, 54, 112–133. [Google Scholar] [CrossRef]
  6. Deriche, R. Using Canny’s criteria to derive a recursively implemented optimal edge detector. IJCV 1987, 1, 167–187. [Google Scholar] [CrossRef]
  7. Bourennane, E.; Gouton, P.; Paindavoine, M.; Truchetet, F. Generalization of Canny-Deriche filter for detection of noisy exponential edge. Signal Proces. 2002, 82, 1317–1328. [Google Scholar] [CrossRef]
  8. Marr, D.; Hildreth, E. Theory of edge detection. Proc. R. Soc. Lond. Ser. B Biol. Sci. 1980, 207, 187. [Google Scholar] [CrossRef]
  9. Freeman, W.T.; Adelson, E.H. The Design and Use of Steerable Filters. IEEE TPAMI 1991, 13, 891–906. [Google Scholar] [CrossRef]
  10. Jacob, M.; Unser, M. Design of steerable filters for feature detection using Canny-like criteria. IEEE TPAMI 2004, 26, 1007–1019. [Google Scholar] [CrossRef] [PubMed]
  11. Geusebroek, J.; Smeulders, A.; van de Weijer, J. Fast anisotropic Gauss filtering. In European Conference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2002; pp. 99–112. [Google Scholar]
  12. Magnier, B.; Montesinos, P.; Diep, D. Fast anisotropic edge detection using Gamma correction in color images. In Proceedings of the 7th International Symposium on Image and Signal Processing and Analysis (ISPA 2011), Dubrovnik, Croatia, 4–6 September 2011; pp. 212–217. [Google Scholar]
  13. Martin, D.; Fowlkes, C.; Tal, D.; Malik, J. A database of human segmented natural images and its application to evaluating segmentation algorithms and measuring ecological statistics. In Proceedings of the Eighth IEEE International Conference on Computer Vision, Vancouver, BC, Canada, 7–14 July 2001; Volume 2, pp. 416–423. [Google Scholar]
  14. Abdulrahman, H.; Magnier, B.; Montesinos, P. From contours to ground truth: How to evaluate edge detectors by filtering. J. WSCG 2017, 25, 133–142. [Google Scholar]
  15. Heath, M.D.; Sarkar, S.; Sanocki, T.; Bowyer, K.W. A robust visual method for assessing the relative performance of edge-detection algorithms. IEEE TPAMI 1997, 19, 1338–1359. [Google Scholar] [CrossRef]
  16. Kitchen, L.; Rosenfeld, A. Edge evaluation using local edge coherence. IEEE Trans. Syst. Man Cybern. 1981, 11, 597–605. [Google Scholar] [CrossRef]
  17. Haralick, R.M.; Lee, J.S. Context dependent edge detection and evaluation. Pattern Recognit. 1990, 23, 1–19. [Google Scholar] [CrossRef]
  18. Zhu, Q. Efficient evaluations of edge connectivity and width uniformity. Image Vis. Comput. 1996, 14, 21–34. [Google Scholar] [CrossRef]
  19. Deutsch, E.S.; Fram, J.R. A quantitative study of the orientation bias of some edge detector schemes. IEEE Trans. Comput. 1978, 3, 205–213. [Google Scholar] [CrossRef]
  20. Venkatesh, S.; Kitchen, L.J. Edge evaluation using necessary components. CVGIP Graph. Models Image Process. 1992, 54, 23–30. [Google Scholar] [CrossRef]
  21. Magnier, B. Edge detection: A review of dissimilarity evaluations and a proposed normalized measure. Multimed. Tools Appl. 2017, 77, 1–45. [Google Scholar] [CrossRef]
  22. Strickland, R.N.; Chang, D.K. An adaptable edge quality metric. Optical Eng. 1993, 32, 944–952. [Google Scholar] [CrossRef]
  23. Nguyen, T.B.; Ziou, D. Contextual and non-contextual performance evaluation of edge detectors. Pattern Recognit. Lett. 2000, 21, 805–816. [Google Scholar] [CrossRef]
  24. Dubuisson, M.P.; Jain, A.K. A modified Hausdorff distance for object matching. IEEE ICPR 1994, 1, 566–568. [Google Scholar]
  25. Chabrier, S.; Laurent, H.; Rosenberger, C.; Emile, B. Comparative study of contour detection evaluation criteria based on dissimilarity measures. EURASIP J. Image Video Process. 2008, 2008, 2. [Google Scholar] [CrossRef]
  26. Lopez-Molina, C.; De Baets, B.; Bustince, H. Quantitative error measures for edge detection. Pattern Recognit. 2013, 46, 1125–1139. [Google Scholar] [CrossRef]
  27. Jaccard, P. Nouvelles recherches sur la distribution florale. Bulletin de la Societe Vaudoise des Sciences Naturelles 1908, 44, 223–270. [Google Scholar]
  28. Dice, L.R. Measures of the amount of ecologic association between species. Ecology 1945, 26, 297–302. [Google Scholar] [CrossRef]
  29. Sneath, P.; Sokal, R. Numerical Taxonomy. The Principles and Practice of Numerical Classification; The University of Chicago Press: Chicago, IL, USA, 1973. [Google Scholar]
  30. Duda, R.; Hart, P.; Stork, D. Pattern Classification and Scene Analysis, 2nd ed.; Wiley Interscience: New York, NY, USA, 1995. [Google Scholar]
  31. Grigorescu, C.; Petkov, N.; Westenberg, M. Contour detection based on nonclassical receptive field inhibition. IEEE TIP 2003, 12, 729–739. [Google Scholar] [CrossRef] [PubMed]
  32. Wang, S.; Ge, F.; Liu, T. Evaluating edge detection through boundary detection. EURASIP J. Appl. Signal Process. 2006, 2006, 076278. [Google Scholar] [CrossRef]
  33. Bryant, D.; Bouldin, D. Evaluation of edge operators using relative and absolute grading. In Proceedings of the Conference on Pattern Recognition and Image Processing, Chicago, IL, USA, 6–8 August 1979; pp. 138–145. [Google Scholar]
  34. Usamentiaga, R.; García, D.F.; López, C.; González, D. A method for assessment of segmentation success considering uncertainty in the edge positions. EURASIP J. Appl. Signal Proc. 2006, 2006, 021746. [Google Scholar] [CrossRef]
  35. Lee, S.U.; Chung, S.Y.; Park, R.H. A comparative performance study of several global thresholding techniques for segmentation. CVGIP 1990, 52, 171–190. [Google Scholar] [CrossRef]
  36. Sezgin, M.; Sankur, B. Survey over image thresholding techniques and quantitative performance evaluation. J. Electron. Imaging 2004, 13, 146–166. [Google Scholar]
  37. Venkatesh, S.; Rosin, P.L. Dynamic threshold determination by local and global edge evaluation. CVGIP 1995, 57, 146–160. [Google Scholar] [CrossRef]
  38. Yitzhaky, Y.; Peli, E. A method for objective edge detection evaluation and detector parameter selection. IEEE TPAMI 2003, 25, 1027–1033. [Google Scholar] [CrossRef]
  39. Martin, D.R.; Fowlkes, C.C.; Malik, J. Learning to detect natural image boundaries using local brightness, color, and texture cues. IEEE TPAMI 2004, 26, 530–549. [Google Scholar] [CrossRef] [PubMed]
  40. Bowyer, K.; Kranenburg, C.; Dougherty, S. Edge detector evaluation using empirical ROC curves. Comput. Vis. Image Underst. 2001, 84, 77–103. [Google Scholar] [CrossRef]
  41. Forbes, L.A.; Draper, B.A. Inconsistencies in edge detector evaluation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Hilton Head Island, SC, USA, 15 June 2000; Volume 2, pp. 398–404. [Google Scholar]
  42. Davis, J.; Goadrich, M. The relationship between Precision–Recall and ROC curves. In Proceedings of the 23rd International Conference on Machine Learning, Pittsburgh, PA, USA, 25–29 June 2006; pp. 233–240. [Google Scholar]
  43. Hou, X.; Yuille, A.; Koch, C. Boundary detection benchmarking: Beyond F-measures. In 2013 IEEE Conference on Computer Vision and Pattern Recognition (CVPR); IEEE: Piscataway Township, NJ, USA, 2013; pp. 2123–2130. [Google Scholar]
  44. Valverde, F.L.; Guil, N.; Munoz, J.; Nishikawa, R.; Doi, K. An evaluation criterion for edge detection techniques in noisy images. In Proceedings of the International Conference on Image Processing, Thessaloniki, Greece, 7–10 October 2001; Volume 1, pp. 766–769. [Google Scholar]
  45. Román-Roldán, R.; Gómez-Lopera, J.F.; Atae-Allah, C.; Martınez-Aroza, J.; Luque-Escamilla, P. A measure of quality for evaluating methods of segmentation and edge detection. Pattern Recognit. 2001, 34, 969–980. [Google Scholar] [CrossRef]
  46. Fernández-Garcıa, N.; Medina-Carnicer, R.; Carmona-Poyato, A.; Madrid-Cuevas, F.; Prieto-Villegas, M. Characterization of empirical discrepancy evaluation measures. Pattern Recognit. Lett. 2004, 25, 35–47. [Google Scholar] [CrossRef]
  47. Abdou, I.E.; Pratt, W.K. Quantitative design and evaluation of enhancement/thresholding edge detectors. Proc. IEEE 1979, 67, 753–763. [Google Scholar] [CrossRef]
  48. Pinho, A.J.; Almeida, L.B. Edge detection filters based on artificial neural networks. In ICIAP; Springer: Berlin/Heidelberg, Germany, 1995; pp. 159–164. [Google Scholar]
  49. Boaventura, A.G.; Gonzaga, A. Method to evaluate the performance of edge detector. In Brazlian Symp. on Comput. Graph. Image Process; Citeseer: State College, PA, USA, 2006; pp. 234–236. [Google Scholar]
  50. Panetta, K.; Gao, C.; Agaian, S.; Nercessian, S. A New Reference-Based Edge Map Quality Measure. IEEE Trans. Syst. Man Cybern. Syst. 2016, 46, 1505–1517. [Google Scholar] [CrossRef]
  51. Yasnoff, W.; Galbraith, W.; Bacus, J. Error measures for objective assessment of scene segmentation algorithms. Anal. Quant. Cytol. 1978, 1, 107–121. [Google Scholar]
  52. Huttenlocher, D.; Rucklidge, W. A multi-resolution technique for comparing images using the hausdorff distance. In Proceedings of the Computer Vision and Pattern Recognition (IEEE CVPR), New York, NY, USA, 15–17 June 1993; pp. 705–706. [Google Scholar]
  53. Peli, T.; Malah, D. A Study of Edge Detection Algorithms; CGIP: Indianapolis, IN, USA, 1982; Volume 20, pp. 1–21. [Google Scholar]
  54. Odet, C.; Belaroussi, B.; Benoit-Cattin, H. Scalable discrepancy measures for segmentation evaluation. In Proceedings of the 2002 International Conference on Image Processing, Rochester, NY, USA, 22–25 September 2002; Volume 1, pp. 785–788. [Google Scholar]
  55. Yang-Mao, S.F.; Chan, Y.K.; Chu, Y.P. Edge enhancement nucleus and cytoplast contour detector of cervical smear images. IEEE Trans. Syst. Man Cybern. Part B 2008, 38, 353–366. [Google Scholar] [CrossRef] [PubMed]
  56. Magnier, B. An objective evaluation of edge detection methods based on oriented half kernels. In Proceedings of the Illinois Consortium for International Studies and Programs (ICISP), Normandy, France, 2–4 July 2018. [Google Scholar]
  57. Baddeley, A.J. An error metric for binary images. In Robust Computer Vision: Quality of Vision Algorithms; Wichmann: Bonn, Germany, 1992; pp. 59–78. [Google Scholar]
  58. Magnier, B.; Le, A.; Zogo, A. A Quantitative Error Measure for the Evaluation of Roof Edge Detectors. In Proceedings of the 2016 IEEE International Conference on Imaging Systems and Techniques (IST), Chania, Greece, 4–6 October 2016; pp. 429–434. [Google Scholar]
  59. Abdulrahman, H.; Magnier, B.; Montesinos, P. A New Objective Supervised Edge Detection Assessment using Hysteresis Thresholds. In International Conference on Image Analysis and Processing; Springer: Cham, Switzerland, 2017. [Google Scholar]
  60. Chabrier, S.; Laurent, H.; Emile, B.; Rosenberger, C.; Marche, P. A comparative study of supervised evaluation criteria for image segmentation. In Proceedings of the European Signal Processing Conference, Vienna, Austria, 6–10 September 2004; pp. 1143–1146. [Google Scholar]
  61. Hemery, B.; Laurent, H.; Emile, B.; Rosenberger, C. Comparative study of localization metrics for the evaluation of image interpretation systems. J. Electron. Imaging 2010, 19, 023017. [Google Scholar]
  62. Paumard, J. Robust comparison of binary images. Pattern Recognit. Lett. 1997, 18, 1057–1063. [Google Scholar] [CrossRef]
  63. Zhao, C.; Shi, W.; Deng, Y. A new Hausdorff distance for image matching. Pattern Recognit. Lett. 2005, 26, 581–586. [Google Scholar] [CrossRef]
  64. Baudrier, É.; Nicolier, F.; Millon, G.; Ruan, S. Binary-image comparison with local-dissimilarity quantification. Pattern Recognit. 2008, 41, 1461–1478. [Google Scholar] [CrossRef]
  65. Perona, P. Steerable-scalable kernels for edge detection and junction analysis. In European Conference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 1992; Volume 10, pp. 3–18. [Google Scholar]
  66. Shui, P.L.; Zhang, W.C. Noise-robust edge detector combining isotropic and anisotropic Gaussian kernels. Pattern Recognit. 2012, 45, 806–820. [Google Scholar] [CrossRef]
  67. Laligant, O.; Truchetet, F.; Meriaudeau, F. Regularization preserving localization of close edges. IEEE Signal Process. Lett. 2007, 14, 185–188. [Google Scholar] [CrossRef]
  68. De Micheli, E.; Caprile, B.; Ottonello, P.; Torre, V. Localization and noise in edge detection. IEEE Trans. Pattern Anal. Mach. Intell. 1989, 11, 1106–1117. [Google Scholar] [CrossRef]
  69. Lopez-Molina, C.; Bustince, H.; De Baets, B. Separability criteria for the evaluation of boundary detection benchmarks. IEEE Trans. Image Process. 2016, 25, 1047–1055. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Example of edge detection on an image. In (c), arrows representing η are pondered by | I | .
Figure 1. Example of edge detection on an image. In (c), arrows representing η are pondered by | I | .
Jimaging 04 00074 g001
Figure 2. Example of hysteresis threshold applied along a contour chain.
Figure 2. Example of hysteresis threshold applied along a contour chain.
Jimaging 04 00074 g002
Figure 3. Example of ground truth ( G t ) versus (vs.) a desired contour ( D c ).
Figure 3. Example of ground truth ( G t ) versus (vs.) a desired contour ( D c ).
Jimaging 04 00074 g003
Figure 4. Receiver Operating Characteristic (ROC) and Precision–Recall (PR) curves for several edge detectors. Images in (be) and (hj) represent the best segmentation for each indicated detector tied to ROC curves and PR curves, respectively. The ground truth image (parkingmeter) is available in Figure 16 and the original image in Figure 17 (Peak Signal to Noise Ratio: PSNR = 14 dB).
Figure 4. Receiver Operating Characteristic (ROC) and Precision–Recall (PR) curves for several edge detectors. Images in (be) and (hj) represent the best segmentation for each indicated detector tied to ROC curves and PR curves, respectively. The ground truth image (parkingmeter) is available in Figure 16 and the original image in Figure 17 (Peak Signal to Noise Ratio: PSNR = 14 dB).
Jimaging 04 00074 g004
Figure 5. P v evaluation depends on the r parameter and can produce a negative evaluation. The variable r is represented by the radius of the circle in (ce). The higher the value of r, the higher L and M are and the smaller U and S are (or can become negative for U ).
Figure 5. P v evaluation depends on the r parameter and can produce a negative evaluation. The variable r is represented by the radius of the circle in (ce). The higher the value of r, the higher L and M are and the smaller U and S are (or can become negative for U ).
Jimaging 04 00074 g005
Figure 6. R evaluation depends on mistake distances but depends on an area contained in a window around a mistake point. I R corresponds to the value of R W for each pixel. Images representing I R in (d,e) correspond to inverse images. Here, the window size around each mistake is of size W = 3 × 3 .
Figure 6. R evaluation depends on mistake distances but depends on an area contained in a window around a mistake point. I R corresponds to the value of R W for each pixel. Images representing I R in (d,e) correspond to inverse images. Here, the window size around each mistake is of size W = 3 × 3 .
Jimaging 04 00074 g006
Figure 7. Failure Measure ( F M ) evaluation with two different Multiple Detection ( M D ) zones: in (c), dilation of G t with structuring element 3 × 3, and 5 × 5 in (d). The greater the M D area, the lower the F M error.
Figure 7. Failure Measure ( F M ) evaluation with two different Multiple Detection ( M D ) zones: in (c), dilation of G t with structuring element 3 × 3, and 5 × 5 in (d). The greater the M D area, the lower the F M error.
Jimaging 04 00074 g007
Figure 8. Different D c : number of false positive points ( F P ) and false negative points ( F N ) are the same for D 1 and for D 2 but the distances of FNs and the shapes of the two D c are different. The legend for (d,e) is available in Figure 7.
Figure 8. Different D c : number of false positive points ( F P ) and false negative points ( F N ) are the same for D 1 and for D 2 but the distances of FNs and the shapes of the two D c are different. The legend for (d,e) is available in Figure 7.
Jimaging 04 00074 g008
Figure 9. Example of ground truth ( G t ) versus (vs.) a desired contour ( D c ).
Figure 9. Example of ground truth ( G t ) versus (vs.) a desired contour ( D c ).
Jimaging 04 00074 g009
Figure 10. Evaluation measure results for different D c images in (be) using the same G t in (a).
Figure 10. Evaluation measure results for different D c images in (be) using the same G t in (a).
Jimaging 04 00074 g010
Figure 11. Evaluation measure results for a real image segmented [4] at different hysteresis thresholds.
Figure 11. Evaluation measure results for a real image segmented [4] at different hysteresis thresholds.
Jimaging 04 00074 g011
Figure 12. A single (or a few) FP point(s) at a sufficiently high distance may penalize a good detection. S b represents G t in (a) with only five FPs that penalize the shape using several edge detection evaluation functions.
Figure 12. A single (or a few) FP point(s) at a sufficiently high distance may penalize a good detection. S b represents G t in (a) with only five FPs that penalize the shape using several edge detection evaluation functions.
Jimaging 04 00074 g012
Figure 13. Edge detection evaluations must be more sensitive to F N distances than F P distances. In (b), | D c | = 61 , so there are 20 FPs, whereas, in (d), | D c | = 14 , so there are 27 FNs; so F P < F N .
Figure 13. Edge detection evaluations must be more sensitive to F N distances than F P distances. In (b), | D c | = 61 , so there are 20 FPs, whereas, in (d), | D c | = 14 , so there are 27 FNs; so F P < F N .
Jimaging 04 00074 g013
Figure 14. Several examples of f function evolution as a function of the FN number.
Figure 14. Several examples of f function evolution as a function of the FN number.
Jimaging 04 00074 g014
Figure 15. Example of computation of a minimum score for a given measure.
Figure 15. Example of computation of a minimum score for a given measure.
Jimaging 04 00074 g015
Figure 16. Ground truth edge images tied to original images available in Figure 17 used in the presented experiments. These G t images are available in [14].
Figure 16. Ground truth edge images tied to original images available in Figure 17 used in the presented experiments. These G t images are available in [14].
Jimaging 04 00074 g016
Figure 17. Image 109 (top) and image “parkingmeter” (bottom) at different levels of noise (PSNR).
Figure 17. Image 109 (top) and image “parkingmeter” (bottom) at different levels of noise (PSNR).
Jimaging 04 00074 g017
Figure 18. Image 109: Comparison of edge detection evaluation evolution as a function of PSNR values.
Figure 18. Image 109: Comparison of edge detection evaluation evolution as a function of PSNR values.
Jimaging 04 00074 g018
Figure 19. Image parkingmeter: Comparison of edge detection evaluation evolutions.
Figure 19. Image parkingmeter: Comparison of edge detection evaluation evolutions.
Jimaging 04 00074 g019
Figure 20. Ideal segmentations for several edge detectors on image 109, PSNR = 14 dB.
Figure 20. Ideal segmentations for several edge detectors on image 109, PSNR = 14 dB.
Jimaging 04 00074 g020
Figure 21. Ideal segmentations for several edge detectors on image 109, PSNR = 14 dB.
Figure 21. Ideal segmentations for several edge detectors on image 109, PSNR = 14 dB.
Jimaging 04 00074 g021
Figure 22. Ideal segmentations for several edge detectors on image 109, PSNR = 14 dB.
Figure 22. Ideal segmentations for several edge detectors on image 109, PSNR = 14 dB.
Jimaging 04 00074 g022
Figure 23. Ideal segmentations for several edge detectors on image 109, PSNR = 14 dB.
Figure 23. Ideal segmentations for several edge detectors on image 109, PSNR = 14 dB.
Jimaging 04 00074 g023
Figure 24. Ideal segmentations for several edge detectors on image 109, PSNR = 14 dB.
Figure 24. Ideal segmentations for several edge detectors on image 109, PSNR = 14 dB.
Jimaging 04 00074 g024
Figure 25. Ideal segmentations for several edge detectors on image 109, PSNR = 14 dB.
Figure 25. Ideal segmentations for several edge detectors on image 109, PSNR = 14 dB.
Jimaging 04 00074 g025
Figure 26. Ideal segmentations for several edge detectors on image 109, PSNR = 14 dB.
Figure 26. Ideal segmentations for several edge detectors on image 109, PSNR = 14 dB.
Jimaging 04 00074 g026
Figure 27. Ideal segmentations for several edge detectors on image 109, PSNR = 14 dB.
Figure 27. Ideal segmentations for several edge detectors on image 109, PSNR = 14 dB.
Jimaging 04 00074 g027
Figure 28. Ideal segmentations for several edge detectors on image 109, PSNR = 14 dB.
Figure 28. Ideal segmentations for several edge detectors on image 109, PSNR = 14 dB.
Jimaging 04 00074 g028
Figure 29. Ideal segmentations for several edge detectors on image 109, PSNR = 14 dB.
Figure 29. Ideal segmentations for several edge detectors on image 109, PSNR = 14 dB.
Jimaging 04 00074 g029
Figure 30. Ideal segmentations for several edge detectors on image 109, PSNR = 14 dB.
Figure 30. Ideal segmentations for several edge detectors on image 109, PSNR = 14 dB.
Jimaging 04 00074 g030
Figure 31. Ideal segmentations for several edge detectors on image 109, PSNR = 14 dB.
Figure 31. Ideal segmentations for several edge detectors on image 109, PSNR = 14 dB.
Jimaging 04 00074 g031
Figure 32. Ideal segmentations for several edge detectors on image 109, PSNR = 14 dB.
Figure 32. Ideal segmentations for several edge detectors on image 109, PSNR = 14 dB.
Jimaging 04 00074 g032
Figure 33. Ideal segmentations for several edge detectors on image 109, PSNR = 14 dB.
Figure 33. Ideal segmentations for several edge detectors on image 109, PSNR = 14 dB.
Jimaging 04 00074 g033
Figure 34. Ideal segmentations for several edge detectors on image parkingmeter, PSNR = 14 dB.
Figure 34. Ideal segmentations for several edge detectors on image parkingmeter, PSNR = 14 dB.
Jimaging 04 00074 g034
Figure 35. Ideal segmentations for several edge detectors on image parkingmeter, PSNR = 14 dB.
Figure 35. Ideal segmentations for several edge detectors on image parkingmeter, PSNR = 14 dB.
Jimaging 04 00074 g035
Figure 36. Ideal segmentations for several edge detectors on image parkingmeter, PSNR = 14 dB.
Figure 36. Ideal segmentations for several edge detectors on image parkingmeter, PSNR = 14 dB.
Jimaging 04 00074 g036
Figure 37. Ideal segmentations for several edge detectors on image parkingmeter, PSNR = 14 dB.
Figure 37. Ideal segmentations for several edge detectors on image parkingmeter, PSNR = 14 dB.
Jimaging 04 00074 g037
Figure 38. Ideal segmentations for several edge detectors on image parkingmeter, PSNR = 14 dB.
Figure 38. Ideal segmentations for several edge detectors on image parkingmeter, PSNR = 14 dB.
Jimaging 04 00074 g038
Figure 39. Ideal segmentations for several edge detectors on image parkingmeter, PSNR = 14 dB.
Figure 39. Ideal segmentations for several edge detectors on image parkingmeter, PSNR = 14 dB.
Jimaging 04 00074 g039
Figure 40. Ideal segmentations for several edge detectors on image parkingmeter, PSNR = 14 dB.
Figure 40. Ideal segmentations for several edge detectors on image parkingmeter, PSNR = 14 dB.
Jimaging 04 00074 g040
Figure 41. Ideal segmentations for several edge detectors on image parkingmeter, PSNR = 14 dB.
Figure 41. Ideal segmentations for several edge detectors on image parkingmeter, PSNR = 14 dB.
Jimaging 04 00074 g041
Figure 42. Ideal segmentations for several edge detectors on image parkingmeter, PSNR = 14 dB.
Figure 42. Ideal segmentations for several edge detectors on image parkingmeter, PSNR = 14 dB.
Jimaging 04 00074 g042
Figure 43. Ideal segmentations for several edge detectors on image parkingmeter, PSNR = 14 dB.
Figure 43. Ideal segmentations for several edge detectors on image parkingmeter, PSNR = 14 dB.
Jimaging 04 00074 g043
Figure 44. Ideal segmentations for several edge detectors on image parkingmeter, PSNR = 14 dB.
Figure 44. Ideal segmentations for several edge detectors on image parkingmeter, PSNR = 14 dB.
Jimaging 04 00074 g044
Figure 45. Ideal segmentations for several edge detectors on image parkingmeter, PSNR = 14 dB.
Figure 45. Ideal segmentations for several edge detectors on image parkingmeter, PSNR = 14 dB.
Jimaging 04 00074 g045
Figure 46. Ideal segmentations for several edge detectors on image parkingmeter, PSNR = 14 dB.
Figure 46. Ideal segmentations for several edge detectors on image parkingmeter, PSNR = 14 dB.
Jimaging 04 00074 g046
Figure 47. Ideal segmentations for several edge detectors on image parkingmeter, PSNR = 14 dB.
Figure 47. Ideal segmentations for several edge detectors on image parkingmeter, PSNR = 14 dB.
Jimaging 04 00074 g047
Figure 48. Segmentations and scores concerning Ξ measure and image 109.
Figure 48. Segmentations and scores concerning Ξ measure and image 109.
Jimaging 04 00074 g048
Figure 49. Segmentations and scores concerning Ξ measure and image parkingmeter.
Figure 49. Segmentations and scores concerning Ξ measure and image parkingmeter.
Jimaging 04 00074 g049
Table 1. Gradient magnitude and orientation computation for a scalar image I, where I θ represents the image derivative using a first-order filter at the θ orientation (in radians).
Table 1. Gradient magnitude and orientation computation for a scalar image I, where I θ represents the image derivative using a first-order filter at the θ orientation (in radians).
Type of OperatorFixed Operator [3,4,5,6,7]Oriented Filters [9,10,11]Half Gaussian Kernels [12]
Gradient magnitude | I | = I 0 2 + I π / 2 2 | I | = max θ 0 , π | I θ | | I | = max θ 0 , 2 π I θ min θ 0 , 2 π I θ
Gradient direction η = arctan I π / 2 I 0 η = arg max θ 0 , π | I θ | + π 2 η = arg max θ 0 , 2 π I θ + arg min θ 0 , 2 π I θ / 2
Table 2. List of error measures involving only statistics.
Table 2. List of error measures involving only statistics.
Complemented D i c e measure [28] D i c e = 1 2 · T P 2 · T P + F N + F P
Complemented P e r f o r m a n c e m e a s u r e [29,30,31,32] P m G t , D c = 1 T P | G t D c | = 1 T P T P + F P + F N
Complemented Absolute Grading [33] A G = 1 T P G t · D c = 1 T P T P + F N · T P + F P
Complemented Segmentation Success Ratio [34] S S R = 1 T P 2 G t · D c = 1 T P 2 T P + F N · T P + F P
L o c a l i z a t i o n e r r o r [35] P E G t , D c = F P + F N | I |
M i s c l a s s i f i c a t i o n E r r o r [36] M E G t , D c = 1 T P + T N T N + F N + T P + F P
Complemented Φ measure [37] Φ G t , D c = 1 T P R · T N T N + F P
Complemented χ 2 measure [38] χ 2 G t , D c = 1 T P R T P F P 1 T P F P · T P + F P + F P R T P + F P
Complemented F α measure [39] F α G t , D c = 1 P REC · T P R α · T P R + ( 1 α ) · P REC , with α ] 0 ; 1 ]
Table 3. Coefficients of Equation (4) determined by a least square adjustment [45].
Table 3. Coefficients of Equation (4) determined by a least square adjustment [45].
Kwbp i bh h i hb c Euler
1.71.10.0130.154.50.370.0868.9
Table 4. List of error measures involving distances, generally: k = 1 or k = 2 , and, κ = 0.1 or κ = 1 / 9 .
Table 4. List of error measures involving distances, generally: k = 1 or k = 2 , and, κ = 0.1 or κ = 1 / 9 .
Error Measure NameFormulationParameters
Pratt’s Figure of Merit (FoM) [47] F o M G t , D c = 1 1 max G t , D c · p D c 1 1 + κ · d G t 2 ( p ) κ 0 ; 1
F o M revisited [48] F G t , D c = 1 1 G t + β · F P · p G t 1 1 + κ · d D c 2 ( p ) κ 0 ; 1 and β R +
Combination of F o M and statistics [49] d 4 G t , D c = 1 2 · T P max G t , D c 2 + F N 2 + F P 2 max G t , D c 2 + F o M 2 G t , D c κ 0 ; 1 and β R +
Edge map quality measure [50] D p G t , D c = 1 / 2 I G t · p F P 1 1 1 + κ · d G t 2 ( p ) + 1 / 2 G t · p F N 1 1 1 + κ · d T P 2 ( p ) κ 0 ; 1
Symmetric FoM [21] S F o M G t , D c = 1 2 · F o M G t , D c + 1 2 · F o M D c , G t κ 0 ; 1
Maximum FoM [21] M F o M G t , D c = max F o M G t , D c , F o M D c , G t κ 0 ; 1
Yasnoff measure [51] Υ G t , D c = 100 I · p D c d G t 2 p None
Hausdorff distance [52] H G t , D c = max max p D c ( d G t ( p ) ) , max p G t ( d D c ( p ) ) None
Maximum distance [24] f 2 d 6 G t , D c = max 1 D c · p D c d G t p , 1 G t · p G t d D c p None
Distance to G t [24,26,53] D k G t , D c = 1 D c · p D c d G t k p k , k = 1 for [24,53] k R +
Oversegmentation [54] Θ G t , D c = 1 F P · p D c d G t p δ T H k for [54]: k R + and δ T H R +
Under-segmentation [54] Ω G t , D c = 1 F N · p G t d D c p δ T H k for [54]: k R + and δ T H R +
R e l a t i v e D i s t a n c e E r r o r [24,55,56] R D E k G t , D c = 1 D c · p D c d G t k p k + 1 G t · p G t d D c k p k , k R + , k = 1 for [24], k = 2 for [55,56]
Symmetric distance [24,26] S k G t , D c = p D c d G t k p ) + p G t d D c k p D c G t k ,     k = 1 for [24] k R +
Baddeley’s Delta Metric [57] Δ k ( G t , D c ) = 1 | I | · p I | w ( d G t ( p ) ) w ( d D c ( p ) ) | k k k R + and a convex function w : R R
Magnier et al. measure [58] Γ ( G t , D c ) = F P + F N | G t | 2 · p D c d G t 2 ( p ) None
Complete distance measure [21] Ψ ( G t , D c ) = F P + F N | G t | 2 · p G t d D c 2 ( p ) + p D c d G t 2 ( p ) None
λ measure [59] λ ( G t , D c ) = F P + F N | G t | 2 · p D c d G t 2 ( p ) + min | G t | 2 , | G t | 2 T P 2 · p G t d D c 2 ( p ) None
Table 5. Reliability of the reviewed edge detection evaluation measures.
Table 5. Reliability of the reviewed edge detection evaluation measures.
MeasureSegmentation ReliabilityMonotonic CurvesFilter Qualification
D i c e
P m
S S R
P E
χ 2
Φ
F α
P v r = 2
P v r = 4
R W
F M W
H
Δ k
H 5 %
F o M
F
d 4
S F o M
M F o M
D p
E M M
f 2 d 6
R D E k = 1
R D E k = 2
S k = 1 k
S k = 2 k
Ψ
λ
Ξ

Share and Cite

MDPI and ACS Style

Magnier, B.; Abdulrahman, H.; Montesinos, P. A Review of Supervised Edge Detection Evaluation Methods and an Objective Comparison of Filtering Gradient Computations Using Hysteresis Thresholds. J. Imaging 2018, 4, 74. https://doi.org/10.3390/jimaging4060074

AMA Style

Magnier B, Abdulrahman H, Montesinos P. A Review of Supervised Edge Detection Evaluation Methods and an Objective Comparison of Filtering Gradient Computations Using Hysteresis Thresholds. Journal of Imaging. 2018; 4(6):74. https://doi.org/10.3390/jimaging4060074

Chicago/Turabian Style

Magnier, Baptiste, Hasan Abdulrahman, and Philippe Montesinos. 2018. "A Review of Supervised Edge Detection Evaluation Methods and an Objective Comparison of Filtering Gradient Computations Using Hysteresis Thresholds" Journal of Imaging 4, no. 6: 74. https://doi.org/10.3390/jimaging4060074

APA Style

Magnier, B., Abdulrahman, H., & Montesinos, P. (2018). A Review of Supervised Edge Detection Evaluation Methods and an Objective Comparison of Filtering Gradient Computations Using Hysteresis Thresholds. Journal of Imaging, 4(6), 74. https://doi.org/10.3390/jimaging4060074

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop