Next Article in Journal
Guest Editors’ Note—Special Issue on Spatial Audio
Previous Article in Journal
An Efficient Synthesis of Novel Pyrazole-Based Heterocycles as Potential Antitumor Agents
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Feature-Based Structural Measure: An Image Similarity Measure for Face Recognition

by
Noor Abdalrazak Shnain
1,
Zahir M. Hussain
2,3,* and
Song Feng Lu
1
1
School of Computer Science and Technology, Huazhong University of Science and Technology, Wuhan 430074, China
2
Faculty of Computer Science & Mathematics, University of Kufa, Najaf 54001, Iraq
3
School of Engineering, Edith Cowan University, Perth, WA 6027, Australia
*
Author to whom correspondence should be addressed.
Appl. Sci. 2017, 7(8), 786; https://doi.org/10.3390/app7080786
Submission received: 4 July 2017 / Revised: 25 July 2017 / Accepted: 31 July 2017 / Published: 3 August 2017

Abstract

:

Featured Application

This work integrates image structural features into a new similarity measure suitable for face recognition with the new merit of recognition confidence.

Abstract

Facial recognition is one of the most challenging and interesting problems within the field of computer vision and pattern recognition. During the last few years, it has gained special attention due to its importance in relation to current issues such as security, surveillance systems and forensics analysis. Despite this high level of attention to facial recognition, the success is still limited by certain conditions; there is no method which gives reliable results in all situations. In this paper, we propose an efficient similarity index that resolves the shortcomings of the existing measures of feature and structural similarity. This measure, called the Feature-Based Structural Measure (FSM), combines the best features of the well-known SSIM (structural similarity index measure) and FSIM (feature similarity index measure) approaches, striking a balance between performance for similar and dissimilar images of human faces. In addition to the statistical structural properties provided by SSIM, edge detection is incorporated in FSM as a distinctive structural feature. Its performance is tested for a wide range of PSNR (peak signal-to-noise ratio), using ORL (Olivetti Research Laboratory, now AT&T Laboratory Cambridge) and FEI (Faculty of Industrial Engineering, São Bernardo do Campo, São Paulo, Brazil) databases. The proposed measure is tested under conditions of Gaussian noise; simulation results show that the proposed FSM outperforms the well-known SSIM and FSIM approaches in its efficiency of similarity detection and recognition of human faces.

Graphical Abstract

1. Introduction

In view of the current issues of terrorism, security in the western world has been heightened, and is becoming an important and challenging task. Surveillance cameras are now common in airports, hospitals, universities, ATMs, banks and anywhere else with a security system. There is an urgent need to find an accurate and reliable automated system which is able to recognize human faces, in order to identify suspicious persons using cameras [1]. Facial recognition has therefore become the focus of current research, and has received significant attention from academics; many techniques and methods for recognizing human faces have been proposed and applied. However, until now there has been no method which offers the required accuracy in all situations. In other words, existing systems still fall far short of the abilities of a human perception system. This challenge is mainly due to factors that affect the features of an image, such as changes in illumination, background, facial expression and head pose. The task of automatic face recognition is therefore a very complex problem; see [2,3,4] for surveys.
In general, there are two classes of face recognition methods: structural and statistical approaches. Structural approaches are based on the extraction of the local features of a face, such as the shapes of the eyes, mouth and nose. These methods suffer from the unpredictability in facial appearance and environmental conditions, e.g., head pose, when a profile image is compared with a frontal image [5]. Conversely, in statistical approaches the complete face region is treated as input data; in other words, global features are extracted from the whole image. The region surrounding the face, including the hair, shoulders and background, is considered to be irrelevant data and may adversely affect the recognition results [6]. Statistical approaches for feature extraction which are based on statistical moment have been used for classification and recognition applications due to their invariance properties [7].
In most facial recognition applications, the similarity between two images is calculated, i.e., using a reference image and a training database image. Several matching algorithms for face and object recognition systems have been designed based on image similarity measurements, such as the structural similarity measure (SSIM), which won an Emmy award for its prediction of the perceived quality of digital images and videos [8].
The main aim of this work is to design an efficient similarity index to resolve the shortcomings of the existing measures in giving high confidence in recognition. The proposed measure combines the best features of the well-known SSIM and FSIM approaches, striking a balance between performance for similar and dissimilar images of human faces. In addition to the statistical structural properties provided by SSIM, edge detection is incorporated as a distinctive structural feature. In addition to confidence, the proposed measure (FSM) gives a reliable similarity between any two images even under noise, in other words, FSM produces maximal similarity when the images are similar, while giving near-zero similarity when the images are dissimilar. These properties gave the proposed measure high ability to recognize face images under noisy conditions, different facial expressions and pose variations. Such properties are highly needed in security applications while checking the identity of a specific face image in a big database.

2. Background

During the last few years, many studies, methods and approaches to similarity and recognition of human faces have been presented to achieve high success rates for accuracy, confidence in identification and authentication in security systems. A significant step in image similarity was presented by Wang and Bovik in 2004 [9], when they proposed the SSIM. This measure is based on the statistical similarity between two images. However, SSIM can give ambiguous results for two different poses, and confusing results in certain cases where it indicates a high similarity for two images which are in fact dissimilar [10].
In an attempt to develop the similarity measures and using them for face recognition; Hu et al. proposed a similarity measure based on Hausdorff distance for face recognition. This measure can provide similarity and dissimilarity information of two objects to compare them such as faces with different illumination condition and facial expression. The measure has a better performance than the measures based on conventional Hausdorff distance and the eigenface approaches [11].
Xu et al. [12] proposed two new operators for face recognition; to find the pixel intensity variation information of overlapping blocks resulted from the original face image. Two factors let the proposed operators work better than the conventional interest operator: the first one is that by taking the relative as the feature of an image block. The second one is that the scheme to partition an image into overlapping allows the proposed operators to give more representation information for the face image.
In 2011, an improved version of SSIM was proposed by Zhang et al. [13]. This measure was called feature similarity index for image quality assessment (FSIM) in an effort to emphasize its use in recognition; this method depends on the low-level features of images. In this measure, the dimensionless and gradient magnitude are the main features, which play complementary roles in characterizing the image. However, FSIM also produces confusing results in certain cases; for example, it may detect a similarity between dissimilar images or a non-trivial amount of similarity between two different poses.
Singh et al. [14] presented a new similarity approach and applied it in face recognition; by combining phase information of wavelet moments and use the real and imaginary components of wavelet moments to describe an image and develop a similarity measure which is invariant under image rotation.
In 2016 Li et al. presented a novel approach for solving the problem of automatic face recognition when the human faces are taken from frontal views with varying illumination and disguise. The approach is based on extracting dynamic subspace of images and obtaining the discriminative parts in each individual to represent the characteristics of discriminative components and give a recognition protocol to classify face images [15].
The topic of 3D face recognition is getting attention. In 2008 a novel deformation invariant image for robust 3D face recognition was proposed by Li et al. [16]. This work is divided into three phases; firstly getting the depth and the intensity of an image from the original 3D facial data. Secondly, geodesic level curves are created by constructing radial geodesic distance image from the depth image. Finally, deformation invariant image is created by evenly sampling points from the selected geodesic level curves in the intensity image. In 2017 Marcolin et al. presented 105 new geometrical descriptors for 3D face analysis. These descriptors created by composing primary geometrical descriptors (for instance mean and Gaussian) and the coefficients of the fundamental forms, and by applying trigonometric functions (for instance sine and cosine) and logarithm to them [17]. Another approach for 3D feature extraction via geometrical analysis is presented in [18]. To the best of our knowledge, 3D face recognition methods haven’t addressed 3D image similarity so far. It is the intention of the Authors to handle this topic soon.
All of these works didn’t address the confidence in recognition. Only the recognition rate has been used as performance measure. Suppose that two measures ( M 1 and M 2 ) recognized a face image with 100% similarity. This recognition does not mean that the two measures have the same performance. For example, if M 1 produces significant similarities between the test image and another group of K database images, then the suspected faces would be not only the one with maximal similarity, but 1 + K of possible faces. Now if M 2 produces significant similarity with one correct face while giving trivial similarities with the rest of the database, then M 2 is giving the correct answer with more confidence than M 1 which gave several confusing faces in addition to the best match. The proposed measure outperforms existing measures by far in giving less confusion while recognizing the best match for a given face image.
The remainder of this paper is arranged as follows: Section 3 describes the design of the proposed FSM approach for use in face recognition. Section 4 presents the experimental results, performance analysis and discussion. Section 5 presents the conclusions.

3. Similarity Measures

Similarity is the strength of the relationship between two images, while a similarity measure is defined as the distance (based on a specific norm) between various data points. The performance of any similarity measure is heavily dependent on the choice of an appropriate distance function for a given dataset. In this work, an efficient measure is proposed; put simply, it combines the essential elements of both the feature similarity and structural similarity measures. Its performance is analyzed in comparison with both the SSIM and FSIM approaches; the results of the proposed measure were better in all cases, even under conditions of noise.

3.1. Overview of SSIM: A Structural Similarity Measure

Automatic detection of the similarity between images is a significant issue; it plays an essential role in image processing due to its importance in many image processing applications such as enhancement, compression and identity checks. The most important step in the development of image similarity techniques was taken in 2004 by Wang and Bovik [9] when they designed the structural similarity measure (SSIM), which considered image distortion as a combination of three factors: correlation, luminance and contrast. This measure is used extensively in image quality assessment and in many image processing algorithms. The basis of this measure is a set of statistical measurements such as mean, variance and co-variance, used to develop a definition for a distance function that can calculate the structural similarity between a test image and training images. The SSIM measure is expressed by the following equation:
ρ ( x , y ) = ( 2 μ x μ y + c 1 ) ( 2 σ x y + c 2 ) ( μ x 2 + μ y 2 + c 1 ) ( σ x 2 + σ y 2 + c 2 ) ,
where ρ ( x , y ) represents the similarity between two images x and y ; x is the reference image, while y is usually a corrupted version of x ; and μ x , μ y , σ x and σ y are the statistical means and variances of images x and y , respectively. The variable σ x y is the statistical co-variance between the pixels in images x and y . The constants c 1 and c 2 are given by c 1 = ( K 1 L ) 2 and c 2 = ( K 2 L ) 2 , where K 1 and K 2 are small constants and L = 255 (maximum pixel value). Researchers have confirmed that the values of these constants have little effect on the performance of SSIM [19].

3.2. Overview of FSIM: A Feature Similarity Measure

The feature similarity measure (FSIM) is based on the fact that human visual perception recognizes an image according to its low-level features. FSIM is divided into two parts: the first is the phase congruency (PC), which is a dimensionless measure for the significance of a local structure, while the second is the gradient magnitude (GM). PC and GM play complementary roles in image characterization [13]. The FSIM index between two images f 1 and f 2 is defined as follows:
F S I M { f 1 ( x ) , f 2 ( x ) } = Φ { f 1 ( x ) , f 2 ( x ) } = x Ω S L ( x ) × P C ( x ) x Ω P C m ( x ) ,
where Ω represents the whole spatial domain for image pixels, P C is the phase congruency and S L is a similarity resulting from combined similarity measure for phase congruency S P C ( x ) and similarity measure for gradient S G ( x ) given by the formula:
S L ( x ) = [ S P C ( x ) ] α × [ S G ( x ) ] β ,
where α and β are parameters used to adjust the relative importance of phase congruency (PC) and gradient magnitude (GM) features, here taken as α = β = 1 ; and
S P C ( x ) = 2 P C 1 ( x ) × P C 2 ( x ) + T 1 P C 1 2 ( x ) × P C 2 2 ( x ) + T 1 ,
where T 1 is a positive constant, inserted to increase the stability of S P C (such a consideration was also included in SSIM), and
S G ( x ) = 2 G 1 ( x ) × G 2 ( x ) + T 1 G 1 2 ( x ) × G 2 2 ( x ) + T 2 ,
is the gradient similarity, where G = G x 2 ( x ) + G y 2 ( x ) is the gradient magnitude; G x and G y are partial derivatives of image f ( x ) . The phase congruency PC is defined as follows:
PC ( x ) = E ( x ) ϵ + n A n ( x ) ,
where ϵ is a small positive constant, and
E ( x ) = K 2 ( x ) + H 2 ( x ) ,
where H ( x ) = n o n ( x ) and K ( x ) = n e n ( x ) , o n ( x ) = ξ ( x ) * M n e ; e n ( x ) = ξ ( x ) * M n o , noting that M n e and M n o are even and odd symmetric filters on scale n , and ‘*’ denotes convolution. The function ξ ( x ) is a 1D signal obtained after arranging pixels in different orientations. The local amplitudes A n ( x ) are defined as:
A n ( x ) = e n 2 ( x ) + o n 2 ( x )   ,
where x is the position on scale n .

3.3. The Proposed Measure (FSM)

As discussed above, there are certain shortcomings in the performance of the feature similarity and structural similarity measures, particularly since they produce confusing results in some cases. Both FSIM and SSIM may detect a non-trivial amount of similarity when two images are dissimilar and can give low recognition confidence. Both measures fail to detect similarity under conditions of significant noise.
A new similarity index is proposed here to resolve the shortcomings of SSIM and FSM and to give a more reliable measure of similarity, especially for human face images. This new measure is called the Feature-Based Structural Measure (FSM). On the other front, a new measure for confidence in face recognition is proposed.
FSM combines the best features of the well-known SSIM and FSIM approaches, striking a balance between their performances for similar and dissimilar images. The inclusion of edge detection in FSM presents a distinctive structural feature, in which two binary edge images, g x and g y , are obtained after edge detection of the original images x and y using Canny’s method [20]. The FSM is given by a rational function, as follows:
( x , y ) = ( a + c ) · R ( g x ,   g y ) · Φ ( g x ,   g y ) + b · R ( g x ,   g y ) + e ( a + b ) · ρ ( g x ,   g y ) + c · R ( g x ,   g y ) · Φ ( g x ,   g y ) + e ,
where, ( x , y ) represents the proposed similarity between two images x and y ; usually, x represents the reference image and y represents a corrupted version of x ; Φ represents the feature similarity measure (FSIM); and ρ represents the structural similarity measure (SSIM). The constants are chosen as a = 5, b = 3 and c = 7, while e = 0.01 is added to balance the quotient and avoid division by zero. The function R ( x ,   y ) refers to the global 2D correlation between the images, as follows:
R ( x , y ) = i j ( x i j x o ) ( y i j y o ) [ i j ( x i j x o ) 2 ] [ i j ( y i j y o ) 2 ] ,
where x o and y o are the image global means. This function is applied here as R ( g x ,   g y ) on the new images obtained from the application of edge detection to the original images x and y. The simulation results show that the proposed FSM technique gives better performance than SSIM and FSIM for facial recognition, even under conditions of noise. This excellent performance is due to the rational combination of different approaches (hence different features) with proper weights in Equation (9). The proper weights can control the good/bad features provided by the combined measures.
The only limitation for the proposed measure is the complexity of Equation (9) as compared with the complexity of Equations (1) and (2), where it requires more time. This issue can be alleviated by high-performance computing. However, this is a cheap price for correct recognition, especially in security applications.

3.4. A Measure for Face Recognition Confidence

The quality of recognition varies from one measure to another; hence, a measure is necessary to evaluate its performance. Here, a measure for the quality of recognition is designed based on the confidence provided by the face recognition method. This confidence indicates the nearness of the similarities to a test image given for the recognized face and the second most likely face from the database.
The similarity between a test image and a database of N face images will produce N similarity numbers, ranging between 0 and 1. These numbers are method-dependent. A cluster of numbers within a small range indicates a reduced confidence. The most important similarities (that should be well apart) are the maximum similarity number (which gives the best match, or recognition result) and the second maximum number, which is associated with confusion. If normalization is applied then the maximum is always 1, while the difference m with the second maximum would vary according to the degree of confusion. The confidence in face recognition at level k of similarity with a specific test image is defined in this work as follows:
C k = ( 1 k ) n k ,
where n k is the number of persons with k -level of similarity s k > k .

3.5. Discussion

Image similarity measures (for face recognition) didn’t address the issue of confidence in recognition. Only the recognition rate has been used as performance measure. Suppose that two measures ( M 1 and M 2 ) have recognized a face image (inside a database) with 100% similarity to a reference face image. This recognition does not mean that the two measures have the same performance. For example, if M 1 produces significant similarities between the test image and another group of K (a positive integer of) database images, then the suspected faces would be not only the one with maximal similarity, but a number 1 + K of possible faces. Now if M 2 produces significant similarity with one correct face while giving trivial similarities with the rest of the database images, then M 2 is giving the correct answer with more confidence than M 1 which gave several confusing faces in addition to the best match. We will see that the proposed measure outperforms existing measures by far in giving less confusion while recognizing the best match (in a database) for a given face image.

4. Experimental Results and Performance

4.1. Image Database

Most automatic face recognition systems are dependent on a comparison between a given face image (a reference image) and images saved in memory. The face images in memory are represented by a training set saved as a database. In this paper, our training sets are:
  • ORL faces images database: AT&T (American Telephone & Telegraph, Inc; New York City, NY, USA) or ORL database this is a well-known database which is widely used for testing face recognition systems. It consists of 40 individuals; each with 10 different images (poses). These images are taken with different levels of illumination, rotation, facial expression and facial details such as glasses. The size of each image is 92 × 112 pixels, with 256 grey levels per pixel [21]. See Figure 1.
  • FEI faces images database: FEI database or Brazilian database, it consists of 200 individuals; each with 14 different images (poses). These images are taken at the artificial intelligence laboratory of FEI in Brazil; represented by students and staff at FEI, between 19 and 40 years old with distinct appearance, hairstyle, and adorns. The images in FEI database are colorful and size of each image is 640 × 480 pixels; in our experiment we used 700 images from FEI database for 50 individuals each with 14 poses [22]. See Figure 2.

4.2. Testing

The proposed FSM approach is tested using MATLAB R2015 (from Mathworks Inc., Natick, MA, USA), with a focus on its performance in detecting similarity under noisy conditions. Specifically, the corruption of images under Gaussian noise is considered, due to the fact that this is the main source of noise in many image processing systems. Peak signal-to-noise ratio (PSNR) was used in this test, as follows:
P S N R = L 2 p n ,
where L is the maximum level of illumination and p n is the power of the Gaussian noise [23].

4.3. Results and Discussion

The new FSM technique was implemented as per Equation (9), with the well-known measures SSIM as per Equation (1) and FSIM as per Equation (2). These three methods were implemented simultaneously on 1100 images from two datasets; all images of AT&T database and 700 images of FEI database to compare their performance. The distance between the first and second maxima of the similarity curve (peaks) is used for all recognition methods. A greater distance means more confidence in the facial recognition decision. Almost for all cases, the proposed FSM keeps larger difference between the test image and other individuals; better confidence in the recognition process is obtained. Figure 3 shows the poses of the fourth person in the AT&T database, used for testing; Figure 4 shows the poses of the eighth person, also used for testing. Figure 5 shows the poses of the fifth person in FEI database and Figure 6 shows the poses of the fifteenth person in FEI database also used for testing. The reference pose, used as the test image, is indicated. A pose with no significant angle is chosen for the purposes of testing and comparison, although eye-glasses are considered in the first test (see Figure 3), which increases the difficulty of face recognition.
In Figure 7, Figure 8, Figure 9 and Figure 10, the proposed FSM gives better performance in terms of recognition confidence. In spite of the fact that the other measures SSIM and FSIM correctly decide the proper person with maximum similarity, they give low confidence in their decision because there are many cases of distrust (big similarities with wrong persons) in their decisions (similarities); this is a big challenge when we employ these measures in security recognition tasks. FSM gives more confidence to decide the proper person from a database.
Figure 11 shows the performance of similarity measures using person No. 15 in FEI database. Pose 8 is the Reference Image, while a different pose (pose 10) is used for testing for all database images (including the target person). FSIM fails, while confidence of SSIM, and FSM is 0.0364, 0.12 respectively.
Figure 12, Figure 13, Figure 14 and Figure 15 we simulated the confidence measure as per Equation (11) to show the efficiency of the proposed measure (FSM) in giving the highest confidence versus the well-known SSIM and FSIM.
In Figure 16, Figure 17, Figure 18 and Figure 19, the proposed measure FSM produces near-one (maximal) similarity for similar images under Gaussian noise, while giving near-zero similarity when the images are dissimilar. The other measures SSIM and FSIM gave non-trivial amounts of similarity when the two images are quite different, which indicates a shortcoming in these measures.
Figure 7 and Figure 8 show the results of the similarity test on person No. 4 and person No. 8 in AT&T database. The similarity test between the reference image and other images in database, where for the 10 poses for each individual the maximum similarity between the rest of the poses is considered.
Figure 9 and Figure 10 show the results of the similarity test on person No. 5 and person No. 15 in FEI database. The similarity test between the reference image and other images in database, where for the 14 poses for each individual the maximum similarity between the rest of the poses is considered.
The performance of the proposed measure was also tested when the image was corrupted with Gaussian noise. Peak signal-to-noise ratio (PSNR) was utilized in this case, as discussed above for Equation (12); this was then compared with the structural similarity measure (SSIM) and feature similarity measure (FSIM). The results show the effectiveness of FSM in similarity and face recognition. Figure 16 and Figure 17 show the performance of FSM for similar images under Gaussian noise, while Figure 18 and Figure 19 show the performance of FSM for dissimilar images under Gaussian noise. Similar images from AT&T database corrupted with Gaussian noise were tested; these are shown in Figure 16a. The performance of FSM (represented by Equation (9)) as compared to that of SSIM and FSIM is shown in Figure 16b.
Similar images from FEI database corrupted with Gaussian noise were tested; these are shown in Figure 17a. The performance of FSM (represented by Equation (9)) as compared to that of SSIM and FSIM is shown in Figure 17b.
The results of comparing another face image are shown in Figure 18a. Another face image was used from AT&T database as the dissimilar image. Figure 18b shows the performance of FSM (represented by Equation (9)) as compared to SSIM and FSIM for conditions of Gaussian noise.
In [24] it is shown that higher-order correlations can lead to better performance. It is the intention of the authors to extend this work in the near future towards a higher-order correlative measure, analogous to that in Equation (9).
The results of comparing another face image are shown in Figure 19a. Another face image was used from FEI database as the dissimilar image. Figure 19b shows the performance of FSM [represented by Equation (9)] as compared to SSIM and FSIM under Gaussian noise.

5. Conclusions

An efficient similarity measure known as the feature-based structural measure (FSM) is proposed here, based on a rational functional and combining the well-known algorithms of the structural similarity measure (SSIM) and feature similarity measure (FSIM) with Canny edge-detected versions of two images. This measure enhances the best features of both SSIM and FSIM in addition to the distinctive structural feature provided by Canny’s approach. It is shown here that FSM outperforms the conventional SSIM and FSIM in its ability to detect the similarity (that is, high similarity results) among similar faces, even under significantly noisy conditions; it also avoids the detection of similarity in the case of dissimilar faces, which is a major shortcoming of the existing FSIM and SSIM approaches. A measure for confidence in face recognition has also been proposed as a performance measure for face recognition methods. In facial recognition tasks, FSM gives higher recognition confidence than existing measures. In addition to confidence, the proposed measure (FSM) gives a reliable similarity between any two images even under noise, in other words, FSM produces maximal similarity when the images are similar, while giving near-zero similarity when the images are dissimilar. These properties gave the proposed measure high ability to recognize face images under noisy conditions, different facial expressions and pose variations. Such properties are highly needed in security applications while checking the identity of a specific face image in a big database. In this work, global analysis of facial images has been used, in which the complete facial region is taken as input data and treated simultaneously. Local face analysis has been shown to play an important role in improving face recognition performance. The authors intend to extend their previous work on local analysis to improve the performance of the above measure. In addition, the Authors suggested a call for 3D image similarity, a new topic that they will handle soon.

Acknowledgments

The Authors would like to thank Huazhong University of Science and Technology (China), Edith Cowan University (Australia), and Chinese Scholarship Council for financial support of this project. Many thanks go to the (anonymous) Reviewers of this work for their deep remarks and constructive comments. Without those valuable comments this paper would never be in this readable form.

Author Contributions

All authors extensively discussed the contents of this paper and contributed to its modelling and preparation. Noor Abdalrazak Shnain and Zahir M. Hussain proposed the model, drafted the manuscript, performed experiments, and analyzed results. The revisions of mathematical modelling and simulations were done by Song Feng Lu.

Conflicts of Interest

The authors declare no conflict of interest.

Research Ethics

The authors declare that there are no ethical issues associated with this work.

References

  1. Parmar, D.N.; Mehta, B.B. Face recognition methods & applications. arXiv, 2014; arXiv:1403.0485. [Google Scholar]
  2. Zhao, W.; Chellappa, R.; Phillips, P.J.; Rosenfeld, A. Face recognition: A literature survey. ACM Comput. Surv. (CSUR) 2003, 35, 399–458. [Google Scholar] [CrossRef]
  3. Barrett, W.A. A survey of face recognition algorithms and testing results. In Proceedings of the Conference Record of the Thirty-First Asilomar Conference on Signals, Systems & Computers, Pacific Grove, CA, USA, 2–5 November 1997. [Google Scholar]
  4. Fromherz, T. Face Recognition: A Summary of 1995–1997; Technical Report ICSI TR-98-027; International Computer Science Institute: Berkeley, CA, USA, 1998. [Google Scholar]
  5. Wiskott, L.; Fellous, J.M.; Krüger, N.; Malsburg, C.V.D. Face recognition by elastic bunch graph matching. IEEE Trans. Pattern Anal. Mach. Intell. 1997, 19, 775–779. [Google Scholar] [CrossRef]
  6. Turk, M.A.; Pentland, A.P. Face recognition using eigenfaces. In Proceedings of the CVPR’91 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Washington, DC, USA, 19–23 June 1991. [Google Scholar]
  7. Belkasim, S.O.; Shridhar, M.; Ahmadi, M. Pattern recognition with moment invariants: A comparative study and new results. Pattern Recognit. 1991, 24, 1117–1138. [Google Scholar] [CrossRef]
  8. Taj-Eddin, I.A.; Afifi, M.; Korashy, M.; Hamdy, D.; Nasser, M.; Derbaz, S. A new compression technique for surveillance videos: Evaluation using new dataset. In Proceedings of the Sixth International Conference on Digital Information and Communication Technology and its Applications (DICTAP), Konya, Turkey, 21–23 July 2016. [Google Scholar]
  9. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed]
  10. Hashim, A.N.; Hussain, Z.M. Novel imagedependent quality assessment measures. J. Comput. Sci. 2014, 10, 1548–1560. [Google Scholar] [CrossRef]
  11. Hu, Y.; Wang, Z. A similarity measure based on Hausdorff distance for human face recognition. In Proceedings of the 18th International Conference on Pattern Recognition (ICPR 2006), Hong Kong, China, 20–24 August 2006. [Google Scholar]
  12. Xu, Y.; Yao, L.; Zhang, D.; Yang, J.Y. Improving the interest operator for face recognition. Exp. Syst. Appl. 2009, 36, 9719–9728. [Google Scholar] [CrossRef]
  13. Zhang, L.; Zhang, L.; Mou, X.; Zhang, D. FSIM: A feature similarity index for image quality assessment. IEEE Trans. Image Process. 2011, 20, 2378–2386. [Google Scholar] [CrossRef] [PubMed]
  14. Singh, C.; Sahan, A.M. Face recognition using complex wavelet moments. Opt. Laser Technol. 2013, 47, 256–267. [Google Scholar] [CrossRef]
  15. Li, H.; Suen, C.Y. Robust face recognition based on dynamic rank representation. Pattern Recognit. 2016, 60, 13–24. [Google Scholar] [CrossRef]
  16. Li, L.; Xu, C.; Tang, W.; Zhong, C. 3D face recognition by constructing deformation invariant image. Pattern Recognit. Lett. 2008, 29, 1596–1602. [Google Scholar] [CrossRef]
  17. Marcolin, F.; Vezzetti, E. Novel descriptors for geometrical 3D face analysis. Multimedia Tools Appl. 2017, 76, 13805–13834. [Google Scholar] [CrossRef]
  18. Moos, S.; Marcolin, F.; Tornincasa, S.; Vezzetti, E.; Violante, M.G.; Fracastoro, G.; Speranza, D.; Padula, F. Cleft lip pathology diagnosis and foetal landmark extraction via 3D geometrical analysis. Int. J. Interact. Des. Manuf. (IJIDeM) 2017, 11, 1–18. [Google Scholar] [CrossRef]
  19. Hassan, A.F.; Hussain, Z.M.; Cai-lin, D. An Information-Theoretic Measure for Face Recognition: Comparison with Structural Similarity. Int. J. Adv. Res. Artif. Intell. 2014, 3, 7–13. [Google Scholar]
  20. Canny, J. A computational approach to edge detection. IEEE Trans. Pattern Anal. Mach. Intell. 1986, 8, 679–698. [Google Scholar] [CrossRef] [PubMed]
  21. AT&T Laboratories Cambridge. The Database of Faces. Available online: http://www.cl.cam.ac.uk/research/dtg/attarchive/facedatabase.html (accessed on 28 April 2017).
  22. Thomaz, C.E. (Caru). FEI Face Database. Available online: http://fei.edu.br/~cet/facedatabase.html (accessed on 15 May 2017).
  23. Hassan, A.F.; Cai-lin, D.; Hussain, Z.M. An information-theoretic image quality measure: Comparison with statistical similarity. J. Comput. Sci. 2014, 10, 2269–2283. [Google Scholar] [CrossRef]
  24. Lajevardi, S.M.; Hussain, Z.M. Novel higher-order local autocorrelation-like feature extraction methodology for facial expression recognition. IET Image Process. 2010, 4, 114–119. [Google Scholar] [CrossRef]
Figure 1. Various face poses for a single person from The AT&T (American Telephone & Telegraph) Face Database.
Figure 1. Various face poses for a single person from The AT&T (American Telephone & Telegraph) Face Database.
Applsci 07 00786 g001
Figure 2. Various face poses for a single person from The FEI (Brazilian) Face Database.
Figure 2. Various face poses for a single person from The FEI (Brazilian) Face Database.
Applsci 07 00786 g002
Figure 3. Poses for person No. 4 in AT&T database.
Figure 3. Poses for person No. 4 in AT&T database.
Applsci 07 00786 g003
Figure 4. Poses for person No. 8 in AT&T database.
Figure 4. Poses for person No. 8 in AT&T database.
Applsci 07 00786 g004
Figure 5. Poses for person No. 5 in FEI database.
Figure 5. Poses for person No. 5 in FEI database.
Applsci 07 00786 g005
Figure 6. Poses for person No. 15 in FEI database.
Figure 6. Poses for person No. 15 in FEI database.
Applsci 07 00786 g006
Figure 7. Performance of similarity measures using person No. 4 (pose 10) in the AT&T database. Confidence in recognition for SSIM, FSIM and FSM is 0.5693, 0.2597 and 0.8556 respectively.
Figure 7. Performance of similarity measures using person No. 4 (pose 10) in the AT&T database. Confidence in recognition for SSIM, FSIM and FSM is 0.5693, 0.2597 and 0.8556 respectively.
Applsci 07 00786 g007
Figure 8. Performance of similarity measures using person No. 8 (pose 10) in the AT&T database. Confidence in recognition for SSIM, FSIM and FSM is 0.5773, 0.2803 and 0.8661 respectively.
Figure 8. Performance of similarity measures using person No. 8 (pose 10) in the AT&T database. Confidence in recognition for SSIM, FSIM and FSM is 0.5773, 0.2803 and 0.8661 respectively.
Applsci 07 00786 g008
Figure 9. Performance of similarity measures using person No. 5 (pose 7) in the FEI database. Confidence in recognition for SSIM, FSIM and FSM is 0.2354, 0.2103 and 0.8791 respectively.
Figure 9. Performance of similarity measures using person No. 5 (pose 7) in the FEI database. Confidence in recognition for SSIM, FSIM and FSM is 0.2354, 0.2103 and 0.8791 respectively.
Applsci 07 00786 g009
Figure 10. Performance of similarity measures using person No. 15 (pose 8) in the FEI database. Confidence in recognition for SSIM, FSIM and FSM is 0.2430, 0.1916 and 0.8697 respectively.
Figure 10. Performance of similarity measures using person No. 15 (pose 8) in the FEI database. Confidence in recognition for SSIM, FSIM and FSM is 0.2430, 0.1916 and 0.8697 respectively.
Applsci 07 00786 g010
Figure 11. Performance of similarity measures using person No. 15 in the FEI database. Pose 8 is the Reference Image, while pose 10 is used for testing for all database images (including the target person). FSIM fails, while confidence in recognition for SSIM, and FSM is 0.0364, 0.12 respectively.
Figure 11. Performance of similarity measures using person No. 15 in the FEI database. Pose 8 is the Reference Image, while pose 10 is used for testing for all database images (including the target person). FSIM fails, while confidence in recognition for SSIM, and FSM is 0.0364, 0.12 respectively.
Applsci 07 00786 g011
Figure 12. Recognition confidence for similarity measures SSIM, FSIM and FSM using person No. 4 (pose 10) in the AT&T database.
Figure 12. Recognition confidence for similarity measures SSIM, FSIM and FSM using person No. 4 (pose 10) in the AT&T database.
Applsci 07 00786 g012
Figure 13. Recognition confidence for similarity measures SSIM, FSIM and FSM using person No. 8 (pose 10) in the AT&T database.
Figure 13. Recognition confidence for similarity measures SSIM, FSIM and FSM using person No. 8 (pose 10) in the AT&T database.
Applsci 07 00786 g013
Figure 14. Recognition confidence for similarity measures SSIM, FSIM and FSM using person No. 5 (pose 7) in FEI database.
Figure 14. Recognition confidence for similarity measures SSIM, FSIM and FSM using person No. 5 (pose 7) in FEI database.
Applsci 07 00786 g014
Figure 15. Recognition confidence for similarity measures SSIM, FSIM and FSM using person No. 15 (pose 7) in FEI database.
Figure 15. Recognition confidence for similarity measures SSIM, FSIM and FSM using person No. 15 (pose 7) in FEI database.
Applsci 07 00786 g015
Figure 16. Performance of similarity measures using similar images under Gaussian noise; (a) The test images; (b) Performance comparison among (SSIM, FSM and FSIM) under Gaussian noise.
Figure 16. Performance of similarity measures using similar images under Gaussian noise; (a) The test images; (b) Performance comparison among (SSIM, FSM and FSIM) under Gaussian noise.
Applsci 07 00786 g016
Figure 17. Performance of similarity measures using similar images under Gaussian noise; (a) Above: The test images; (b) Performance comparison among (SSIM, FSM and FSIM) under Gaussian noise.
Figure 17. Performance of similarity measures using similar images under Gaussian noise; (a) Above: The test images; (b) Performance comparison among (SSIM, FSM and FSIM) under Gaussian noise.
Applsci 07 00786 g017
Figure 18. Performance of similarity measures using dissimilar images under Gaussian noise; (a) Above: The test images; (b) Performance of SSIM, FSM and FSIM under Gaussian noise.
Figure 18. Performance of similarity measures using dissimilar images under Gaussian noise; (a) Above: The test images; (b) Performance of SSIM, FSM and FSIM under Gaussian noise.
Applsci 07 00786 g018
Figure 19. Performance of similarity measures using dissimilar images under Gaussian noise; (a) The test images; (b) Performance comparison among (SSIM, FSM and FSIM) under Gaussian noise.
Figure 19. Performance of similarity measures using dissimilar images under Gaussian noise; (a) The test images; (b) Performance comparison among (SSIM, FSM and FSIM) under Gaussian noise.
Applsci 07 00786 g019

Share and Cite

MDPI and ACS Style

Shnain, N.A.; Hussain, Z.M.; Lu, S.F. A Feature-Based Structural Measure: An Image Similarity Measure for Face Recognition. Appl. Sci. 2017, 7, 786. https://doi.org/10.3390/app7080786

AMA Style

Shnain NA, Hussain ZM, Lu SF. A Feature-Based Structural Measure: An Image Similarity Measure for Face Recognition. Applied Sciences. 2017; 7(8):786. https://doi.org/10.3390/app7080786

Chicago/Turabian Style

Shnain, Noor Abdalrazak, Zahir M. Hussain, and Song Feng Lu. 2017. "A Feature-Based Structural Measure: An Image Similarity Measure for Face Recognition" Applied Sciences 7, no. 8: 786. https://doi.org/10.3390/app7080786

APA Style

Shnain, N. A., Hussain, Z. M., & Lu, S. F. (2017). A Feature-Based Structural Measure: An Image Similarity Measure for Face Recognition. Applied Sciences, 7(8), 786. https://doi.org/10.3390/app7080786

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop