Next Article in Journal
Dual-Scale Doppler Attention for Human Identification
Next Article in Special Issue
Individual Identification by Late Information Fusion of EmgCNN and EmgLSTM from Electromyogram Signals
Previous Article in Journal
A Probability-Based Models Ranking Approach: An Alternative Method of Machine-Learning Model Performance Assessment
Previous Article in Special Issue
Augmented PIN Authentication through Behavioral Biometrics
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Biometrics: Going 3D

by
Gerasimos G. Samatas
and
George A. Papakostas
*
MLV Research Group, Department of Computer Science, International Hellenic University, 65404 Kavala, Greece
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(17), 6364; https://doi.org/10.3390/s22176364
Submission received: 3 July 2022 / Revised: 12 August 2022 / Accepted: 20 August 2022 / Published: 24 August 2022
(This article belongs to the Special Issue Biometric Systems for Personal Human Recognition)

Abstract

:
Biometrics have been used to identify humans since the 19th century. Over time, these biometrics became 3D. The main reason for this was the growing need for more features in the images to create more reliable identification models. This work is a comprehensive review of 3D biometrics since 2011 and presents the related work, the hardware used and the datasets available. The first taxonomy of 3D biometrics is also presented. The research was conducted using the Scopus database. Three main categories of 3D biometrics were identified. These were face, hand and gait. The corresponding percentages for these categories were 74.07%, 20.37% and 5.56%, respectively. The face is further categorized into facial, ear, iris and skull, while the hand is divided into fingerprint, finger vein and palm. In each category, facial and fingerprint were predominant, and their respective percentages were 80% and 54.55%. The use of the 3D reconstruction algorithms was also determined. These were stereo vision, structure-from-silhouette (SfS), structure-from-motion (SfM), structured light, time-of-flight (ToF), photometric stereo and tomography. Stereo vision and SfS were the most commonly used algorithms with a combined percentage of 51%. The state of the art for each category and the available datasets are also presented. Finally, multimodal biometrics, generalization of 3D reconstruction algorithms and anti-spoofing metrics are the three areas that should attract scientific interest for further research. In addition, the development of devices with 2D/3D capabilities and more publicly available datasets are suggested for further research.

1. Introduction

Biometrics are unique body characteristics used to identify people. They were first used at the end of the 19th century with the well-known and globally used fingerprints. According to Jain et al. [1], the most commonly used biometric methods are DNA, ear, facial, hand and finger veins, fingerprint, gait, hand geometry, iris, palmprint, retina, signature and voice. To perform the identification, a device is used to capture the biometric data. Most often, this device captures images, and the quality of the captured images therefore affects the performance of the model. The first identifications were completed manually by experts, but in some cases, the results were controversial due to the human factor. Later, the use of technology for identification has evolved in the form of image processing methods and matching techniques, creating tremendous identification models that take advantage of biometrics. Over the years, these technologies achieved great performance and thus became more popular. Moreover, biometrics for identifying people is nowadays not only used in forensics but also to gain access to certain places or to log in to some smart devices.
As technology has advanced, a biometric security problem has emerged. The technology became very familiar and very vulnerable to malicious acts. This has had a major impact on various security protocols as biometrics have become a part of our daily lives. To counter this, a few approaches have been developed in this area. One of them is to increase the robustness of the selected biometric category. Bear in mind that the traditional methods are referred to 2D, and some scientific approaches have led to the development of 3D biometrics. Of course, this was not just about the fancy addition of the third dimension but mainly about increasing the extracted features and creating more efficient systems. These additional features are the key to the desired performance improvement.
Computer vision has always been linked to biometrics as it provides the necessary tools for identification through 3D image analysis. In addition, the technological advancement of computer vision using state-of-the-art Artificial Intelligence methods to achieve the above benefits has led to the need to apply it to identification systems. As the demand for robust models has increased, the transition from 2D to 3D biometric methods has been a one-way street.
The core element for 3D reconstruction is depth information. Various algorithms have been developed to extract the relevant information. The first work published for 3D biometrics in general was from David Zhang and Guangming Lu in 2013 [2]. In their book, they described the image acquisition methods and categorized them into two major categories, the single and multi-view approaches. Another approach is to categorize them into active and passive methods. In active methods, the light source is directed to the desired surface as an essential step for 3D reconstruction and is characterized by low computational cost. Passive methods, on the other hand, are usually very computationally intensive and use ambient light conditions [3]. Furthermore, active methods can be categorized into structured light, time of flight (ToF), photometric stereo and tomography. Passive methods include stereo vision, structure-from-silhouette (SfS), structure-from-texture (SfT) and structure-from-motion (SfM). All of the above methods are presented in the form of a taxonomy in Figure 1. The taxonomy shows that the two main approaches, active and passive, have four methods at once. With further approaches, it is possible to create a third category. This category could be semi-active, semi-passive, or even a combination of passive and active methods. Such multimodal approaches should lead to new 3D reconstruction algorithms.
Structured light produces a beam from a light source onto the surface of the object. The wavelength of the light can be in the visible range, the infrared (IR) or even the near infrared (NIR). The calculations from the reflection of the beam provide depth information. Secondly, the ToF method takes into account the reflection time between the surface and a reference point, while the photometric stereo method uses different lights to create different shades of the object and the model is created by combining these lights. Finally, tomography can be either optical coherence tomography or computed tomography (CT). In both cases, the 3D models are created by multiple scans.
In addition, stereo vision creates depth information by comparing image information of the same spot from two different viewpoints. SfS uses images taken from different angles to form the silhouette of the object. In addition, the SfT is applied when the surface has a homogeneous texture and then uses the different orientation of the images to create the depth details. Finally, the SfM uses a video as input, which is usually consisting of frames that capture an object from different angles.
An important factor for successful reconstruction is the sensor used. The RGB-D camera is commonly used in various 3D reconstruction applications. This particular type can provide color and depth information by combining the three primary colors (red–green–blue) and calculating the relative distance to the sensor accordingly. According to an analysis of RGB-D camera technologies for face recognition, Urlich et al. [4] found that stereoscopy (active and passive), followed by structured light, produced the best results. The importance of these cameras was also emphasised by Zollhöfer et al. [5]. In the research, they presented the different approaches and the great performance of these sensors for all aspects of 3D reconstructions, including biometric elements such as the face, etc.
Furthermore, the above eight different methods are used throughout the literature without any correlation between the categories of 3D biometrics, as each category has been studied separately so far. Although some reviews refer to a group of categories, such as facial and ears, the vast majority refer explicitly to one category. This creates additional barriers to the extraction of cross-biometric information, such as common methods between categories or even similarities in the state of the art. Furthermore, there is no literature review in the field that examines all categories of 3D biometrics at once.
This review examines 3D biometrics through a literature review of scientific work in the field, focusing on 3D reconstruction algorithms and methods. Both are essential for successful identification with 3D images. Furthermore, the contribution of this work is to provide statistical data on 3D biometrics so that a qualitative and quantitative comparison of each biometric category and its 3D reconstruction methods is possible. The result is the first taxonomy of 3D biometrics and a correlation analysis between 3D reconstruction methods and biometric category. The analysis shows that face as a 3D biometric is over-saturated with a variety of approaches in this direction. Moreover, stereo vision and SfS are the most commonly used methods among biometric categories. The analysis also presents various available datasets and the state of the art in each category. Finally, it highlights the challenges that have arisen and need to be addressed in the transition to 3D.
The paper is organized as follows: Section 2 contains the related work on reviews about the 3D biometrics, Section 3 presents the literature search protocol, which is vital prior to gathering firm results through literature searching. Section 4 presents the 3D reconstruction approaches for the three main biometric categories of face, hand and gait. The categories and their subcategories were extracted through the literature search. Section 5 contains the results of the research with the available datasets, state of the art and statistics. Section 6 contains a discussion of the work, indicating the need for further research and suggestions. Finally, Section 7 contains the conclusion of the paper, summarizing all the information provided in the previous sections.

2. Related Work

In order to document the related work, a preliminary research was conducted using the trusted platform Scopus [6]. A customized search query was applied. More specifically, the word 3D biometric was selected, and the results were limited to certain types of publications (reviews, survey etc.) The research revealed that the published 3D biometrics reviews consist of two main categories: the face and fingerprints.

2.1. Face

The first and most popular category was the face recognition, with the facial and ear being two popular subsections. In some cases, the ear is part of a multimodal approach or a standalone biometric. In 2011, Yuan et al. [7] studied ear recognition. In the article, the author had also included the recognition process of 3D images for the first time. The ear has a unique shape and also shows minor deformations over the years. According to the article, the main reconstruction methods are SfS, SfM and Stereo Vision, with the last method being the most effective. Yaun et al. [7] also conclude that the accuracy and robustness of the system are greatly improved when ear features are used in combination with face features.
The first review for facial biometric was by Islam et al. in 2012 [8], which presented a 3D facial pipeline in four stages. These stages were 3D data acquisition, detection, representation and recognition. The authors provide some details about the data acquisition techniques. For face recognition, there are two main categories: the use of 2D or 3D images. The state of the art for using 2D images was the Support Vector Machine (SVM) with 97.1% on 313 faces [9] and for 3D images was the Point Distribution Model (PDM), which achieved 99.6% on 827 images [10]. For the representation of the face, the balloon image [11] and iso-countours [12] were the most advanced models. When used to reconstruct a face, they achieved 99.6% and 91.4% accuracy in face recognition, respectively. The final step was the recognition. Since the face often changes during emotional expressions, the author presented two main categories: rigid and non-rigid, depending on whether the model is considered rigid or not. Although the percentages for the rigid approach were high (the Iterative Closest Point (ICP) algorithm [13] reached 98.31%), some samples were rejected by the algorithm due to different expressions. On the other hand, rigid approaches had similar performance but increased computational cost. For ear reconstruction, the author proposed three different approaches. The first was to use landmarks and a 3D mask, but this approach depends on manual intervention. For the second approach, the use of 3D template matching was proposed, with an accuracy of 91.5%. This approach had better performance than the previous one but a higher error rate. The last and most efficient method was the Ear Shape Model, proposed by Chen and Bhanu [14] in 2005, which achieved an accuracy of 92.5% with an average processing time of 6.5 s.
The use of ToF methods in 3D face recognition was reviewed by Zhang and Lu in 2013 [2], presenting two main approaches for ToF applications. The first is image capture with incoherent light, and the second is based on optical shutter technology. Both use the reflection of NIR light. They also pointed out the disadvantages of these devices, namely the high error rates due to the generation of low-resolution images by the different devices. Of course, they also believe that hardware will be able to support a higher resolution biometric system in the near future. The main advantage of the Tof is that it can provide real-time results, which are very important for biometrics.
In 2014, Subban and Mankame [15] wrote a review paper focusing on 3D face recognition methods and proposing two different approaches. The first extracts features from the different facial attributes (nose, lips, eyes, etc.), and the second assumes the face as a whole entity. Furthermore, the authors presented the methods that had the best performance based on recognition rate (RR). In particular, a combination of geometric recognition and local hybrid matching [16] achieved 98.4%, which was followed by the method of local shape descriptor with almost the same performance (98.35%) [17]. The remaining methods 3D morphing [18] and multiple nose region [19] were equally efficient with 97% and 96.6%, respectively. In the same year, Alyuz et al. [20] described the phenomenon of difficulty in identifying 3D faces in the presence of occlusions. These occlusions can be accessories such as hats or sunglasses or even a finger in front of the face. However, Alyuz proposed a method consisting of removing occlusions and then restoring the missing part of the face, achieving a high identification accuracy (93.18%).
In the following year, Balaban et al. [21] reviewed deep learning approaches for facial recognition. The author emphasized that as deep learning models evolve, better datasets are needed. In fact, the state-of-the-art Google FaceNet CNN model [22] had an accuracy of 99.63% in the Labeled Faces in the Wild (LFW) dataset [23]. This very high accuracy somehow shows that the scientific community should create datasets with a lot of additional images. Balaban also believes that such a dataset will be revolutionary and compares it to the transition from Caltech 101 to Imagenet datasets. The next year, in 2016, Liu et al. [24] presented the weaknesses of facial recognition systems when the input images are multimodal. These can be IR, 3D, low-resolution or thermal images, which are also known as heterogeneous data. Liu also suggested that future recognition algorithms should be robust to multimodal scenarios in order to be successfully used in live recognition scenarios. The authors also highlight the fact that humans can easily perform face recognition with multimodal images. In order to mimic this behaviour, one approach is to make the different models exposed to long-term learning procedures.
Furthermore, Bagga et al. introduced a review of anti-spoofing methods in face recognition, including 3D approaches [25]. In face spoofing, the “attacker” creates fake evidence to trick a biometric system. In 3D, these proofs are fake masks created from real faces. Four techniques are proposed: motion, texture, vital sign and optical flow-based analysis. Motion-based analysis involves analyzing the movement of different parts of the face, such as the chin or forehead, so that any attempt at forgery can be detected. The best approach is a calculation estimate based on illumination invariance by Klaus Kollreider et al. [26]. The second technique extracts the texture and frequency information using Local Binary Pattern (LBP), which is followed by histogram generation. Finally, an SVM classifier classifies whether the face is real or fake. The third method is vital sign recognition analysis. This can be performed either by user interaction such as following some simple commands (e.g., head movements, etc.) or in a passive way, i.e., by detecting mouth and/or eye movements. Lastly, the optical flow-based analysis, proposed by Wei et al. [27], is primarily used to distinguish a fake 2D image from a 3D face.
In 2017, Mahmood et al. [28] presented the state of the art in face recognition, using 3D images as input. According to their review, five recognition algorithms stood out for their high performance. These are ICP [29], Adaptively Selected Model Based (ASMB) [30], Iterative Closest Normal Point (ICNP) [31], Portrait Image Pairs (PIP) [32] and Perceived Facial Images (PFI) [33] with 97%, 95.04%, 99.6%, 98.75% and 98%, respectively. The authors concluded that despite the high accuracy achieved by applying state-of-the-art algorithms, the models are not yet reliable enough to process data in real-time. Furthermore, they underlined that the performance varies greatly on different datasets, and further research should be conducted in this direction.
Later, in 2019, Albakri and Alghowinem [34] reviewed various anti-spoofing methods. According to their study, pulse estimation, image quality and texture analysis are the most efficient methods. Meanwhile, they proposed a proof-of-concept study in which the fake input is detected by computing the depth data. Using four different types of attacks (flat image, 3D mask, on-screen image and video) on three different devices, iFace 800, iPhone X and Kinect, they managed to detect the fake attack with very high accuracy (92%, 100% and 100%, respectively). The last two reviews were also related to anti-spoofing approaches. Wu et al. [35] focused on applications in China and also presented three trends for protecting against facial forgery. The first was a full 3D reconstruction of the face, whose main drawback is low performance. The next was a multimodal fusion approach consisting of a combination of visible and IR light generated by binocular cameras. Finally, the generative model with a proposed new noise model [36] was the state of the art. Finally, Liu et al. [37] reviewed the results of the Chalearn Face Anti-Spoofing Attack Detection Challenge at CVPR2020. Eleven teams participated using the metric ACER standardized on ISO and achieved great performance with the best percentages of 36.62% and 2.71% for single and multimodal approaches, respectively.

2.2. Fingerprint

The 3D fingerprint was first reviewed in 2013 by Zhang and Lu [2]. First, they present the general flowchart of 3D fingerprint reconstruction from images acquired by touchless devices. The first step is to calibrate the camera, which is a very common procedure for reconstruction applications. The next step is to determine the three different correspondences. These are based on the SIFT feature, the ridge map and the minutiae. After these correspondences, the coordinates are created based on the generated matching points. The final step is to produce an estimation of the shape of the human finger. The author concludes that the most difficult process in the whole reconstruction is the establishment of the different correspondences.
In 2014, Labati et al. [38] presented a review on contactless fingerprinting as a biometric. The paper described the advantages of capturing the corresponding images without contact between the finger and the capture device. The author presented three main approaches to reconstruct the fingerprint: a multiview technique, followed by structured light and stereophotometry. In the multiview technique, several cameras are used to obtain the images. More specifically, two, three and five cameras have been proposed. It should be noted here that although a large number of cameras means higher accuracy, it also increases the computational cost. Therefore, the optimal way to design a system is to find a compromise between accuracy and computational cost. Using a structured light provides an estimate of the ridges and valleys of the fingerprint in addition to texture estimation. The light is projected onto the finger several times in the form of a wave to capture the corresponding image. The last method was photometric stereo imaging. Using a single camera and several LED lights as illuminators, the fingerprint was reconstructed using the photometric stereo technique. The above approaches were promising and opened up new scientific fields. The author also emphasized that there is no work yet on the 3D reconstruction of fingerprints that can be used as a biometric.
Later, Jung et al. [39] suggested ultrasonic transducers as a possible way to find depth information. More specifically, the author reviewed the various ultrasonic sensors used in microelectromechanical systems (MEMS) and their applications. One of them was to acquire appropriate depth information of a fingerprint and its use as a biometric. With the improvement of MEMS technology, ultrasonic sensors are improving in terms of accuracy and robustness. Moreover, this type of sensor outperforms the more traditional ones, such as optical or capacitive, achieving better performance. For this reason, Jung underlined that the above sensors have great scientific potential for 3D biometrics. The use of ultrasound devices was further highlighted in 2019 by Lula et al. [40]. The article described the use of ultrasound devices as a suitable method for capturing 3D fingerprints that can be used in biometric systems. The main problem with this approach was the long acquisition time of the images. This was overcome by using additional probes, which are usually arranged cylindrically. It should also be noted that the frequency bandwidth varies between 7.5 and 50 MHz and depends on the probe model and circumference chosen. A high ultrasound frequency offers high skin penetration but low resolution, while a lower frequency has the opposite effect.
Finally, Yu et al. [41] presented a literature review on 3D fingerprints as a biometric using optical coherence tomography (OCT) as an acquisition device. To calculate the depth information, the light beam was directed at the subject and a mirror. The light penetrated the finger and the system correlated the reflection of the finger’s and the mirror’s beam. Through this process, the system calculated the depth information and then reconstructed the 3D representation. The light penetration also provided the inner fingerprint, which is unaffected by environmental influences. Sweat pores and glands are also visible through this approach. These additional elements provide more features, and the biometric system has become more robust as a result. Despite the fact that OCT has the above advantages, there are also some disadvantages. These include latency, cost, mounting limitations and low resolution
Related work shows that only two 3D biometric categories have been reviewed: the face and the fingerprint. This is probably because these biometrics are very common as 2D biometrics, especially facial. This led to the fact that the other biometric categories such as iris or finger vein were not reviewed. In order to present their state of the art, additional reviews about them should be conducted. In this paper, the taxonomy of 3D biometrics is first created, and then, its applications are examined. As a result, the state of the art and common 3D reconstruction methods between the categories can be presented.

3. Literature Search

In order to obtain additional information on further advances in 3D biometrics, a literature analysis was conducted. This section contains that analysis with two subsections: the search protocol and the initial statistical results. It should be mentioned here that the search was conducted according to strictly defined criteria so that everyone can reproduce the methodology used and obtain the same results.

3.1. Search Protocol

In order to obtain usable results, a search query was carefully designed and also applied to Scopus. The following query was applied simultaneously to the abstract, title and keywords of the publication:
3D* W/2 reconstruction*
AND
biometric*
AND
Publication Year > 2011
The term “W/2” was inserted to ensure that the “3D” and the “reconstruction” were close in the paragraph (two words). The “*” ensures the singular or plural form of the words “biometric” and “reconstruction”. The total number of articles researched was 96. The first problem that arose early on was correspondence with biometrics in general. More specifically, many medical publications did not refer to biometrics despite references to 3D reconstruction algorithms. This was due to the misleading double meaning of the word “biometric”. To counteract this, a quality criterion was applied that excluded the above-mentioned publications. As a result, 31 papers were removed. In addition, 11 articles could not be found or the language was not English, and one article was a duplicate entry. The total number of papers that were thoroughly reviewed was finally 53.

3.2. Statistical Results

At first, the most important result extracted was the 3D biometrics categories. Three main categories were extracted: face, hand and gait. The first two had further subcategories. More specifically, for the face, there was the facial, ear, iris and skull, and for the hand, the categories were fingerprints, finger veins and palm. Considering that one paper experimented with two different biometrics, the total number of approaches was 54. The following Figure 2 shows a tree diagram with the number of applications of all the categories of 3D biometrics described above, while Table 1 shows the main category, percentage and corresponding citations for each biometric feature. In addition, Figure 3 shows a pie chart with the number of applications of the three main categories. Two other pie charts (Figure 4 and Figure 5) also contain the number of applications for the face and hand category, respectively.

4. Three-Dimensional (3D) Reconstruction

This section presents the applications obtained from the literature analysis, focusing on the 3D reconstruction approaches for each 3D biometric category. Since the face subcategory contains the most applications and is saturated and over-described as 3D biometric, it is not presented in as much detail as the other categories (relative to the number of applications) in this paper. As mentioned earlier, there are three main categories of 3D biometrics: the face, the hand and the gait. Figure 6 shows the taxonomy of these categories with their subcategories. The taxonomy, which is the first attempt at 3D biometrics, shows that not all biometrics are 3D. Some other biometrics such as voice, heartbeat or DNA should be further explored to determine if they benefit from becoming 3D. A general flowchart of 3D biometrics is proposed in Figure 7. The selection of the appropriate biometric is followed by image acquisition. This is followed by 3D reconstruction using one of the available methods. The system compares the 3D model with the data in the database and decides whether the identification of the person is valid or not.

4.1. Face

Face is the first and most popular category. Its subcategories are the facial, the ear, the iris and the skull.

4.1.1. Facial

Facial recognition is the predominant method for 3D biometric applications according to the above research. To be used as a biometric, a 3D model must first be created, which is followed by feature extraction. This can be completed using computer vision techniques. The way the images are captured determines the reconstruction method. Of the total 32 papers, more than half (18) used an active method. This corresponds to 56.25%. The remaining 14 are divided as follows: 10 of them (31.25%) used passive methods, three had used a public dataset (9.38%), and one did not refer to the method used at all. Analyzing the passive methods further, the results show that three subcategories are related to facial recognition: stereo vision, SfS and SfM. The percentages were almost equal as seven applications used SfS, six used stereo vision and five used SfM. The majority of active approaches (9) used reflection from a light source, and only one used transmissive computerized tomography. In addition, the reflective methods can be further divided into two subcategories: structured light and time-of-flight. These percentages were also similar, with five using structured light and the rest (four) using time-of-flight.
Another feature of 3D reconstruction is the way in which depth information is represented. The vast majority of applications used a 3D Morphable Model (3DMM) or point clouds. The total percentage was 78.1% for both, which was split into 40.6% and 37.5% for the point cloud and 3DMM, respectively. The remaining 25% were distributed as follows. Four used third-party software, two used unique approaches, and one was without any information. The approach using the 3DMM used a pre-created 3D model, specifically a face, which is very common in the scientific community. A very robust and effective model is the 3DMM proposed by Chu et al. [92]. These models were created by averaging values of different facial features. In most cases, these values were manually inserted, while the final 3D model is created by fine-tuning the above values. The third party software was thatsMyFace.com [42], Amira version 5.2.2, Visage Imaging, San Diego, CA, USA [43] and MeshLAb [59].
In addition, Dou et al. [52] have proposed Deep Recurrent 3D FAce Reconstruction (DRFAR). The network combines a set of Deep Convolutional Neural Network (DCNN) and Recurrent Neural Network (RNN). Prior to the training, the authors transferred the weights of the pre-trained VGG-Face [93]. This neural network architecture was able to achieve good results after experiments in different databases with three to six face images. In addition, Kneis et al. [67] in 2020 presented an alternative method to synthesize a 3D image from four different images of the same subject with different illumination and achieved excellent performance.
Finally, Crispim et al. [59] used a state-of-the-art approach to verify kinship between two faces. They used the SfM approach. Video input was captured with a smartphone camera in an uncontrolled environment. Then, the images were aligned and cropped to 64 × 64 pixels. Feature detection and extraction were followed by feature matching and geometric verification, and finally, projection of the model into a point cloud. The final part was classification, which relied on a two-layer CNN with the ReLU activation function. The highest score achieved with the comparison between the daughter and mother with the accuracy was 95%. To make the dataset more robust, the authors synthesized non-relative faces through generative adversarial networks by combining two different faces. As mentioned in the previous sections, facial is overused. Due to the high number of applications, a lot of effort was put into it using new approaches. CNNs are already used for classification purposes. This makes it necessary to modify the quality criteria of the proposed datasets, as this type of model is usually designed to perform by using a large number of images.

4.1.2. Ear

The next subcategory is the ear. This category shares characteristics with facial recognition. Although ear recognition is strong enough to be used as a stand-alone biometric, some researchers use ears as a complement to facial recognition [94]. One such approach is by Raghavendra et al. [45], who used facial recognition along with features extracted from the ear. Actually, the authors used the SfM approach prior to the entire face reconstruction. Then, the ear is detected from the reconstructed model and used for identification. The authors also proposed an ear detector using OpenCV 2.4. For training the model, they used 2000 positive and 300 negative samples from existing databases. The results show that the identification rate with the facial as a standalone biometric was 82.35%, which decreases slightly to 80% when fused with the 3D ear. The 3D ear as a stand-alone biometric had a rate of 62.5%, which increased to 80% after fusion.
Furthermore, Siu-Yeung Cho proposed a recognition approach using 3D ear models under different illumination parameters in his paper [82]. More specifically, by applying a generalized neural reflectance (GNR) model, they managed to synthesize ear images. Twenty photographs of each ear were taken from 85 different individuals, and after applying the GNR, 40 additional images were created from each individual image. In this way, a 3D ear database was created. The algorithm chosen was SfS, as the models were created from a single image (synthetic or real). A total of 12 experiments were conducted, and the performance proves that Cho’s GNR model performs significantly better when GNR is applied. Finally, the author suggested the use of principal component analysis (PCA) [95] and Fisher’s discriminant analysis (FDA) [96,97] for classification.
The state of the art is considered to be the approach of Chen Li et al. [98], in which they proposed a novel method for 3D ear reconstruction. First, a 3D Ear Morphable Model (3DEMM) was developed based on the general morphable model [99] containing 180 different ear images. Then, 179 of them were used for training and the last one was used for fitting. In addition, they proposed a novel method for feature correspondence, the Triangle Mesh Hierarchical Growth, which is based on physical features common to each ear. The constructed ear had 10,000 points, which is a higher point density than the SfM and SfS algorithms. The point density as mentioned above can be compared to the use of a laser scanner. It is quite old work, and the lack of modern approaches means that 3D ear biometrics is likely to become extinct in the near future. To avoid this, a new application with new methods and data sets should be carried out. This will certainly bring the ear again in the field.

4.1.3. Iris

The next subcategory is the iris. It owes its biometric properties to the unique surface made of two muscular fibers [100]. In general, there is a lot of scientific research in the field of the iris as a biometric but in two dimensions. Bastias et al. [84] claim that their work is the first approach to create a 3D iris model. In their work, they propose the use of near-infrared (NIR) images taken with a Raspberry Pi v2.1 camera. In addition, infrared light was generated from four NIR LED. To keep the pupil relatively small, two white LED were used simultaneously to create a bright environment around the pupil. This is important because prior to the 3D reconstruction, the pupil is removed, and the depth information is extracted from the rest of the eye. The Raspberry Pi 3 Model B single-board computer was used to control the camera and store the images temporally. All the above components were mounted on a VR glass set. The camera was also on a proposed mount that allowed it to move along an arc of 40 . The system was designed to capture a total of 17 iris images, with an angular difference of 2.5 between the two images. In addition, the authors used the Python photogrammetry toolbox [101] to create the 3D model. The toolbox is open-source and creates the model using various images from different angles. It is also noted that it was difficult to evaluate the results and methods of the work due to the lack of reports on metrics and results.
In contrast to the previous approach, Benalcazta et al. [85,86] chose visible light as the light source. They have developed a novel system that illuminates the iris with lateral and frontal visible light (LFVL). The light is produced by six LEDs in the front and two more on the side of each eye. This different angle of the light source creates the shadows and depth information of the iris. The dataset was created from 120 different subjects by recording each eye for three seconds. The total number of images was 26,520. In addition, the authors synthesized an iris dataset consisting of 100 virtual models. To do this, they used Blender, which is an open-source application for creating various 3D shapes. They also augmented the models by generating 720 images per virtual iris, bringing the total number of synthetic images to 72,000. To capture the depth information, they trained a neural network to predict the depth information based on the shadows. More specifically, they proposed irisDepth, which is a combination of the T2Net [102] and DenseDepth [103] networks.
The above model has performed excellently compared to other CNN or SfM approaches and is considered state of the art. In fact, the reconstruction accuracy was as high as the reconstruction using OCT. More specifically, when used the irisDepth, they managed to achieve an accuracy of 99.78%. Despite the fact that 3D iris can achieve high accuracy, there are not many applications in this field. This is probably due to the difficult nature of image acquisition. It is not convenient enough to have someone look through a capture device for several seconds to gain access to a location. In the near future, new devices should be developed that allow for easier and more user-friendly image capture so that a 3D iris can be used in real-world problems for recognition.

4.1.4. Skull

The skull, in a biometric sense, refers to cranio facial reconstruction. It has a slightly different scope from the previous one, because it usually involves cadavers, which are difficult to recognize. To aid recognition, a reconstruction of the face is done based on the skull. These algorithms are known as CFR (cranio facial reconstruction), and forensic science is the science that deals with them. A crucial point for successful reconstruction is the facial features introduced by Whitaker and Linton [104]. Their application to cranial reconstruction is used to reconstruct the face of the deceased by combining skull and tissue reconstruction to identify it. Data acquisition also presents some challenges. Early applications used 3D laser scanners, but their performance was limited [105]. Nowadays, therefore, computed tomography (CT) is used. This method gives tremendous results, and the disadvantages are limited to the increased radiation. Therefore, it is preferably applied to non-living organisms [106]. In addition, Vezzetti et al. [107] report that the above landmarks were always set manually.
First, Lorkiewicz-Muszyńska et al. [90] introduced a procedure to identify an unknown deceased person. One way of post-mortem identification is to examine DNA for a match or possible relationship. Another is to examine the dental status of the person. In the case of the paper, neither of the above methods could provide reliable results. More specifically, the DNA examination did not reveal a match with a possible brother of the person, and there were no dental records, either. An identification using a reconstructed skull was chosen to match the person with a photo of a missing person. The SOMATOM Sensation 64 from Siemens was chosen. The In Space software was used for the reconstruction, and the evaluation was done manually by applying the image of the missing person as a surface layer on the 3D reconstructed skull. Finally, the deceased was matched with a missing person photo through the procedure described above. Furthermore, it later turned out that the possible brother had a different father, which is why the DNA examination could not link the deceased to his brother.
The state of the art is indeed skull reconstruction from living humans [43]. Experiments were performed on three volunteers who had not had facial plastic surgery, dental transplantation or general facial deformity. A cone beam computed tomography (CBCT) scanner was used to take several images of the volunteers’ skulls. The model was the Alphard Vega from Asahi Roentgen Co., Kyoto, Japan. Reconstruction was performed using third-party software, FreeForm Modelling PlusTM, based on a pre-model database consisting of facial muscles and parotid glands. Finally, model evaluation was performed using deviation maps generated using Geomagic Qualify (Geomagic™ Qualify Version 10; Geomagic, Morrisville, NC) and Rapidform (Rapidform, Seoul, Korea) software. In terms of performance of the proposed method, it achieved the best score on subject B with an average error of 0.31 mm. The other two had 0.46 mm and 0.49 mm for the A and C subjects, accordingly.
The 3D skull biometric is not a very common recognition method. Its uniqueness creates a lot of obstacles. It is used for specific purposes, mostly on cadavers and with an auxiliary function. There is a possibility of creating a framework for skull recognition as a complementary method in forensics in the near future. However, this is not possible today, as more experiments should be conducted to create models specifically for skull reconstruction. Moreover, there is no available dataset, and therefore, future research should be accompanied by the creation of a dataset.

4.2. Hand

The second category is the hand, which contains the fingerprint, finger vein and palm.

4.2.1. Fingerprint

Second in order is the most commonly used biometric method in general, the fingerprint. As fingerprints became a biometric identification method used on a daily basis, the need arose to increase the features. This was achieved by replicating the shape of the finger and using the fingerprint as a surface imprint. The 3D fingerprinting methods are also referred to as non-contact methods in the literature [108]. This is very important, because the 2D version of the fingerprint was correlated with the mandatory contact between the finger and the capture device. It is also noted here that much of the problem with 2D fingerprint recognition is related to contact with the device, as there is often dirt, moisture, etc. on the surface of a finger, making identification difficult. A total of six works were found during the research. To obtain the appropriate depth information, four of them used the stereo vision method and the other two used photometric stereo methods and structured light.
Chaterjee et al. [76] proposed a method to reconstruct the fingerprint using a structured light technique. An LED as a light source is projected onto the finger, while a CCD camera records the reflection. More specifically, the projection unit generates sinusoidal fringe patterns, and the frequency selective (FTM) algorithm [109] was selected to obtain the depth information. In order to create a more reliable system, the authors also incorporated biospeckle analysis as an anti-spoofing method. This was achieved by implementing the visual speckle-based MSF algorithm [110] using a laser as the light source. For the experiments, the laser and the LED were mounted in the same device so that both processes (reconstruction and anti-spoofing) could be performed simultaneously. The authors also found that 3D reconstruction outperformed other techniques, and anti-spoofing via blood flow was successful. Consequently, their proposal was simple, fast, cost-effective, and ready to be used as a commercial product.
Another approach to 3D fingerprinting came from Kumar and Kwong [72]. Their approach focused on creating a cost-effective and accurate system. The need to reduce cost stems from the fact that some approaches use multiple or extremely accurate cameras. To overcome this hurdle, the authors used the photometric stereo technique and a digital camera. The finger was placed 10 cm from the camera, and seven LEDs were symmetrically placed. After acquisition, the images were resized from 2592 × 1944 to 500 × 350 pixels. A total of 10,080 2D images of 240 subjects were captured. Seven 2D images were used for each 3D image. In addition, six fingers were reconstructed from each subject. The results show that the 3D approach outperforms 2D finger recognition. The authors also noted that a multimodal approach, combining 2D and 3D features, could provide even better results.
Liu et al. [74] used three JAI CA-A50 cameras along with four blue LEDs. The three cameras were placed below the finger: one in the middle and the other two on each side. The finger was placed on a fixed base in order to be stable throughout the image acquisition process. The reconstruction process was based on stereo vision. In addition, only the two side cameras were used for reconstruction. The central camera is only used for reference and the texture of the fingerprint. Furthermore, the appropriate fingerprint features are based on SIFT [111], ridge maps, and minutiae. SIFT is a popular approach that is robust to deformation variations and low-quality images. The ridge map is an image where the ridge has a value of 1 and the background has a value of 0. Three steps are required to create the ridge map. These are preprocessing, enhancement and post-processing. Preprocessing involves extracting the region of interest (ROI) and normalizing the image. In addition, the author used Gabor filtering, orientation estimation, and ridge frequency estimation to enhance the images. Finally, binarization and thinning of the image are applied as post-process procedures before the ridge map is created. Minutiae, the most common feature of fingerprints, were extracted from the ridge map.
The final result of all the above methods was finalized by the random sample consensus (RANSAC) algorithm [112]. In addition, the author proposed a precomputed 3D finger model. In the absence of ready-made 3D finger shapes, they designed a new model using the active structured light method to create it. Four hundred and forty images of 220 different fingers were collected. According to Liu, the best performance was achieved when all of the above feature extraction methods were used, which also highlights the importance of making the fingerprint 3D.
Labati et al. [73] proposed a 3D reconstruction approach that is also based on the stereo vision technique. Two charge-coupled device (CCD ) cameras were placed underneath the finger. An LED array emitting green light was placed between the finger and the camera. The system also had a photocell that triggers the light to reduce the blurring effect. The system was designed to be unaffected by the rotation of the finger. In addition, the prepossessing procedure also included enhancement and 2D mapping. The device was used in 30 volunteers aged 10 to 64 years. The authors created five different databases. Each database had its own unique characteristics and purpose. In addition, after calculating the depth value using the triangulation method, the 3D point cloud was created. The final step of the process was to apply the fingerprint as a texture to the 3D shape. Finally, Labati used the Neurotechnology Verifinger software and obtained similar results compared to touch-based approaches. As a result, the author stated that more effort for non-contact systems will soon give better results.
Furthermore, two other approaches [75,77] also used stereo vision. The first is by Xu and Hu, who proposed a 3D reconstruction of minutiae using a differential evolution algorithm [113,114]. The authors believe that SIFT is not powerful enough for 3D fingerprints. They also used the Verifinger software to extract the minutiae features. The experiment was conducted with 150 volunteers (10 fingers per volunteer). Three cameras were also used to take pictures from the left, front and right. Although Xu proved from images that their approach found more corresponding points compared to SIFT, the reconstructed 3D minutiae are far from reliable, and further research should be conducted.
The second approach of Yin et al. [77] is the state of the art. In their work, they proposed a very robust system that also had better accuracy and shorter processing times than other modern approaches. The reconstruction was based on the ridge–valley contrast. First, the depth information was extracted using stereo vision. Then, the fingertip regions were extracted using a pre-trained CNN [115], and the images were subsequently rectified. The SIFT method [116] was applied to the images, which was followed by the creation of the ridge maps. Using the disparity map, the 3D depth information was calculated to reconstruct the fingerprint. The authors also proposed feature extraction from the minutiae and 3D topology polymer (TTP), which is a novel method. For their experiment, they used 60 samples from 24 different fingers. To evaluate the performance of the model, they conducted several experiments with different databases and methods, and the model of Yin et al. [77] still achieved better results. More specifically, the average accuracy on both databases, DB1 and DB2, was 98.05%. This result was the best compared to other approaches. In summary, 3D fingerprinting appears to be prosperous, with databases already available to train new models. Future researchers will take advantage of the already existing datasets and implement modern methods, and algorithms and will surely achieve great results. A big advantage of fingerprints is that they are already known to users. Once the hardware improves for better processing time and the accuracy is better than 2D, 3D fingerprints will prevail over 2D.

4.2.2. Finger Vein

The finger vein is an alternative biometric method and has some unique features. First of all, it is robust to spoofing attacks due to the obligatory existence of an IR light source. The difference between the hemoglobin and the skin’s absorption of the IR light creates the shape of the veins, and their unique shape provides the information for identification [117]. Two-dimensional (2D) finger vein identification provides promising results, such as the approach of Zhang and Wang [118], which achieved a recognition rate of over 99% for the USM database [119]. Despite the very high accuracy of 2D recognition, some researchers have already moved to 3D finger vein recognition, taking advantage of the additional features of the third dimension. In contrast to the numerous research papers for 2D finger veins, there is not much related work for 3D finger veins.
First, Ma et al. [79] used the stereo vision method for 3D reconstruction. Two simple CCD cameras, sensitive to IR light, were mounted in opposite directions from an array of NIR LEDs. Then, the captured images were processed using the contrast-limited adaptive histogram equalization (CLAHE) [120], a 5 × 5 denoising median filter and a linear gray transformation. Sequentially, the finger veins and contours were extracted by applying masks and an adaptive threshold algorithm. The stereo vision algorithm generated the appropriate 3D cloud points, and the ICP algorithm was used for matching. An error threshold generated by the least mean square error was made adjustable for different sensitivity requirements. The computational cost depended on the number of iterations before identification. As the authors pointed out, more iterations meant more computational time and more effort could be put into this.
Veldhuis et al. [80] also used the stereo vision approach for reconstruction. In their preliminary study on 3D finger veins, they also used three cameras. In contrast to the approach of Kang et al., the cameras were placed under the finger. Several NIR LED strips used for IR illumination were positioned in the opposite direction to the cameras. A Raspberry Pi single-board computer was also installed to control the different illumination options. Their research focused mainly on 2D finger veins, which is not covered in this article.
Kang et al. [78], reconstructed the finger vein along with the shape of the finger. First, they proposed a prototype acquisition device. Three simple USB cameras were mounted symmetrically to form an equilateral triangle. There were also three NIR LEDs (850 nm wavelength) between the cameras. The function of these cameras was controlled by a Light Control Unit (LCU). The author proposed a 3D reconstruction of the finger shape and used the vein information as texture. The reconstruction was based on a novel 3D finger model (3DFM) developed by the research team. Furthermore, the features were extracted from the vein texture and finger geometry by a specially developed CNN, and after a fusion process, the model generated a matching score. Despite the promising accuracy of the model, the authors emphasized that future research should be conducted to reduce the high processing time and create a more robust model.
Their research continued in 2022; they proposed a state-of-the-art approach to the same problem [81]. The hardware was the same: three cameras next to NIR LEDs to create a finger and apply the veins as texture. The main difference was the reconstruction algorithm, as SfS was used instead of a stereo vision approach. This approach proved to be more efficient and requires less computation time, so the model is able to solve problems in real-time. They also proposed a new detection method called 3DFVSNet. The experiments were conducted with 905 different fingers. Fourteen images were taken of each finger, combining three different rotations. The total number of different images was thus 12,670. The results show that the 3DFVSNet performs better compared to other known approaches, achieving higher accuracy and lower computational costs. The metrics ER and HR were used by the authors. On both databases, the model manages to achieve on both metrics the best result: 2.61% (ER) and 5.07% (HR) for the SCUT-3DFV-V1 and 2.81% (ER) and 4.49% (HR) for the LFMB-3DPVFV. Finger veins have just started to become 3D. The great advantage of this category will become apparent as more research is conducted. Machine learning is already being used. Three-dimensional (3D) finger veins will be evolved when a variety of publicly available datasets and new devices are created that are accurate and fast enough to be used for real-world problems.

4.2.3. Palm

The last biometric hand category is the palm. Svoboda et al. [91] used a not so common 2D biometric, the palm. They also explored the identification performance of the palm (as a biometric) after it became 3D. The structured light technique was used to capture the biometric information from the surface of three palms. The light source was four 10 mW lasers with a wavelength of 532 nm, each of which could produce a single line beam. There were also three green lenses for illumination. Images were captured by two webcams (Microsoft Lifecam HD 3000) mounted on each side. All the above components were controlled by a microcontroller (MC9S08JS16CWJ). In addition, the system was designed to use only a single image from each laser reflection. The depth information was also generated by the triangulation principle, and the root mean square error (RMSE) was used to evaluate the reconstruction, achieving a merged score of 0.0018 on models 2 and 3. The authors emphasize that the best performance was obtained when the information from all four lasers was combined. Finally, they are reluctant to conduct further experiments in this direction to create a more stable model with better accuracy—this further research may or may not reveal the importance of this biometric, because it mainly depends on the convenience of the hardware and the performance.

4.3. Gait

Although gait is often used to describe a person’s condition (psychological or physical), it is also used as a unique biometric [121]. As a category, it has no subcategory because it is unique.
Fernandez et al. [87] proposed an approach to 3D gait identification. For their experiments, they used two databases, each with six and 16 cameras. The silhouettes were obtained using the Horprasert algorithm [122], and then, the 3D model was reconstructed using the SfS algorithm. Furthermore, the models were further processed to reduce the features by applying PCA and linear discriminant analysis (LDA). For classification, the authors used SVM. The method gave promising results because a majority-based voting system was applied to achieve perfect accuracy, so further research should be conducted.
The same team [88] proposed a new method a year later to solve the same problem. The main difference between the two approaches was the type of 3D reconstruction. Although the algorithms were the same (Horprasert and SfS), the model was represented as a stack of voxels simulating the person’s gait. The size of the voxels was 0.27 × 10−4 m3 and was considered by the authors to be sufficient for 3D gait reconstruction. Three gait morphological descriptors were also proposed: cover by rectangles (CR), cover by rectangles projection (CRP) and cover by cubes (CC). The combination of the aforementioned innovations provided more accuracy and robustness for all metrics in both databases.
Finally, Imote et al. [89] presented a state-of-the-art method for 3D gait reconstruction based on the height–constraint assumption, which assumes that the height of body points changes slightly during walking/running. For the experiments, 14 cameras were used with various resolutions (from 960 × 540 to 350 × 240); two of them were controlled by a Raspberry Pi single board computer, while the remaining 12 were controlled by three CCTV recorders. The model had great performance compared to various datasets. The model achieved a better identification rate of almost 95% on the KY 4D database, which was better than other approaches. Moreover, the authors noted the relationship between the number of height points (Np) and the normalized square errors. The greater the number of Np, the fewer the number of errors, with the optimum number of Np being 240. Despite the great performance of the model, the authors mentioned that future 3D gait reconstruction should be conducted on three axes. The first is to make a greater comparison with various approaches, the second is to expand the applicability of the method, making the process more automated, and the third is to make the model more capable of supporting new human activities. Three-dimensional (3D) gait reconstruction will soon be more popular. The great advantage is the simple acquisition method. A few frames are enough to proceed to recognition. The transition to 3D is probably essential because the models should be very robust and benefit from the additional features of the third dimension.

5. Results

After reviewing the above papers, a lot of useful information was obtained. The first statistics extracted after the initial literature review were the 3D biometric categories. The most predominant category is facial, which includes 59.28% of the applications. The second category, fingerprint, has a share of 11% and a finger vein of 7.41%. Ear, iris and gait have 5.56%, while skull and palm account for 3.7% and 1.85%, respectively. There are some additional data to the three main categories and Figure 3, Figure 4 and Figure 5. More specifically, the percentages of the three main categories are as follows: face →74.07%, hand →20.37% and gait →5.56%. In the face category, the percentages are as follows: facial →80%, ear →7.50%, iris →7.50% and skull →5%. Correspondingly, for the hand category: fingerprint →54.55%, finger vein →36.36% and palm →9.09%.
Moreover, all datasets used for the above 3D biometrics applications were compiled in Table 2. It consists of 26 different databases grouped according to their biometric category, also indicating the year of publication, the number of images included, and the corresponding citation. It should be noted that the number of images for gait biometric and the Youtube Faces database refers to videos and not images. Table 2 shows that there is an abundance of facial datasets, with the most recent being from 2019, so it is fairly recent. Ear datasets appear to be older. The most recent is from 2012, showing that 3D ears are becoming less common. Although the gait is not very common, there are already five databases. The rather simple way to create such a database has probably led to this high number of databases (related to the number of applications). For the others finger vein, iris and fingerprint, there is one dataset each, but the 3D iris dataset is quite old (2007), and new databases should be created for further development.
In addition, after combining the results of the above literature review in each category, according to the impact on each field alongside citation score and year of publication, a state-of-the-art table is presented as Table 3 grouped by biometric category. It contains also information on the year of publication, biometric category, and correspondence reference.
Additional information on 3D capture methods was also extracted. Forty-nine applications referred to the selected method for 3D reconstruction. In summary, the results show that 65.3% used the passive method and the remaining 34.7% used the active method, as shown in Figure 8. Further analysis of the percentages per biometric feature is shown in Figure 9. Accordingly, for the facial and fingerprint biometrics, both methods were used, with a slight preference for the passive method. In contrast, only passive methods were used for finger vein, ear, and gait, and active methods were used for iris, skull, and palm.
More interesting data can be obtained from further analysis of the methods. Figure 10 shows the percentages of each method in the number of applications. Here, it can be seen that stereo vision and SfS are the dominant methods with 27% and 25%, respectively. The combination of these two methods exceeds 50%. These two methods were used for facial, fingerprint, finger vein, ear and gait approaches. The exact number of applications is shown in Figure 11. The diagram shows that the facial uses both methods, as does the finger vein. In contrast, the gait and ear use SfS and fingerprints use stereo vision accordingly. To explore the relationship between the methods and each biometrics, a diagram is presented in Figure 12. Accordingly, gait, iris, skull and palm are associated with a method as follows: gait →SfS, iris →photometric stereo, skull →tomography, and palm →structured light. The most complicated of all is the facial, because the percentages are almost equal. To illustrate this, Figure 13 shows only the 3D facial biometric methods. According to the pie chart, although stereo vision and SfS predominate with a combined percentage of 48%, the other structured light, SfM, and ToF methods have similar numbers (19%, 18%, and 15%, respectively), and it is difficult to describe a pattern.

6. Discussion

This research has explored many aspects of 3D biometrics. First of all, facial recognition was by far the dominant 3D biometric. The percentage was 59.26% far enough from the others. New approaches are applied with different methods, from some traditional computer vision algorithms to state-of-the-art machine/deep learning methods. This particular biometric is almost saturated, and many experiments are being conducted. The percentages are already high, and the models are reliable and robust. On the other hand, there are also some weak points. The most recent one was exploited with COVID-19, where the face mask was an incomparable barrier. Any device with a built-in face recognition mechanism could not identify the correct person when the test subjects were wearing a mask. These types of reasons led to the need for alternative biometric methods. Such an example is the recent software upgrade of Apple’s Face ID, which is a built-in face recognition biometric mechanism. The user selects the option to activate the face recognition with a mask, and the software achieves it by focusing on different face parts such as the eyes and the area around them.
In addition, fusion techniques were proposed in some publications. More specifically, the performance was better when fusion was applied. Fusion can be applied in several ways. It can be applied between 2D and 3D images and has been shown to provide better results. In addition, it can be combined with different biometric features and opens a wide field for further research. There are a variety of combinations of biometric features that are practically limited to the capabilities of the capturing device that should capture the data simultaneously. Fusion can also be used for primary or supplemental use. The additional option can be activated when the results exceed a certain threshold. Such an addition to further research will provide a great framework for robust 3D biometrics.
In addition, the appropriate hardware plays an important role in achieving effective fusion. This literature analysis shows that some biometrics lack datasets, and to overcome this hurdle, researchers need to manufacture new experimental devices to capture the images. The choice of hardware depends on the choice of biometric. This causes additional effort if the experiment needs to combine several different biometrics. Another issue that needs to be pointed out is the lack of performance comparison between 2D and 3D biometrics under the same experimental conditions in the reviewed publications. Despite the potentially better performance of 3D biometrics models due to the larger number of features, they need to be compared to their 2D counterparts to confirm the need for the 3D transition.
To achieve this, future researchers should consider making their own hardware to allow for the simultaneous acquisition of 2D and 3D images. This will ensure that the images produced were created under the same conditions, allowing for a more reliable accuracy comparison. This multi-dimensional capability of the proposed device, in addition to the ability to compare, also provides the ability to merge 2D and 3D images. As mentioned earlier, this will lead to better results. There is another area that could benefit from this capability, which is the anti-spoofing.
The research area of anti-spoofing has a great impact on biometrics. This is another reason why researchers are moving biometrics to 3D, expanding its capabilities. According to the above analysis, some publications are choosing a 3D approach to defend against these types of attacks. Adding a 3D biometric module to a 2D device, or vice versa, offers the possibility of using that module as protection against spoofing. Further research could be conducted here to investigate how vulnerable each 3D biometric category is to spoofing and how each category can help protect against spoofing.
In summary, facial and fingerprints are the categories that benefit most from going 3D. Three-dimensional (3D) facial recognition is already being used on several devices for real-time 3D recognition. It has been proven that the additional features generated from depth information add robustness to the system. Its use in everyday devices, such as smartphones, shows that hardware development is on the cutting edge to support these technologies. On the other hand, despite the fact that there are many approaches to 3D fingerprinting and the additional data have led to great results, the 3D fingerprint system is still far from being used commercially. There are limitations in the capture devices, as fingerprints are already captured with small devices. However, the development of contactless methods is likely to solve this problem. Although it is not obvious that going 3D will automatically give better results than 2D approaches, advanced 3D reconstruction algorithms and hardware should be used to achieve great performance. Therefore, it is worth investing in new algorithms and improving hardware so that the other biometric (other than facial and fingerprint) also provide reliable recognition systems.
Finally, as mentioned earlier, there are datasets for some 3D biometrics and not for others, as there are some ethical issues regarding the release of these data. These data are very sensitive, and in some countries, their publication is prohibited by local laws. Some research teams have developed a dataset but have not shared it with the scientific community. This is an obstacle for prospective research, which needs to create the dataset to conduct the desired experiments. To overcome this hurdle, more efforts should be made to publish the appropriate datasets. Finally, there has not been enough research effort to generalize the 3D reconstruction algorithms, and more research should be conducted to share the state-of-the-art algorithms among different 3D biometrics.

7. Conclusions

This paper is a literature review of 53 papers that attempt to present the various aspects of the transition to 3D biometrics. The papers were selected based on strict criteria, which are described in Section 3. The search targeted 3D biometrics under 3D reconstruction since 2011.
After a comprehensive review of the selected publications, the initial statistical results showed that facial recognition was the most commonly used 3D biometric in this particular research area. The other seven categories were fingerprint, finger vein, ear, iris, gait, skull and palm. These can be classified into three main categories: face, hand and gait. Related work was also presented in Section 2, showing that some of the biometrics have not yet been reviewed, and further research needs to be conducted.
Then, in Section 5, the results of the conducted analysis are presented. The available 3D biometrics databases and the state of the art per category are presented in Table 2 and Table 3, respectively. Summarizing the results, it becomes clear in which area more research should be conducted. This is the creation of multimodal biometrics that combine 2D with 3D and mixing the different categories such as the Tharewal et al. multimodal approach with face and ear [148]. In addition, further research should be conducted on how resistant each category is to spoofing and how they offer new methods to deal with it. Finally, further research should be conducted on how the 3D reconstruction algorithms can be generalized to 3D biometrics.
As for future work, a framework should be described in which each new 3D biometric research should include metrics about the corresponding 2D images. To this end, devices should include 2D capabilities. In addition, more datasets for less common biometric features should be created to make it easier for future researchers to conduct experiments. These points are crucial and will be the key factor for further improvements so that the necessity for going 3D can be proven.

Author Contributions

Conceptualization, G.A.P.; methodology, G.G.S. and G.A.P.; formal analysis, G.G.S.; investigation, G.G.S.; resources, G.G.S.; data curation, G.G.S.; writing—original draft preparation, G.G.S.; writing—review and editing, G.G.S. and G.A.P.; visualization, G.G.S.; supervision, G.A.P. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable, the study does not report any data.

Acknowledgments

This work was supported by the MPhil program “Advanced Technologies in Informatics and Computers”, hosted by the Department of Computer Science, International Hellenic University, Kavala, Greece.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Jain, A.K.; Ross, A.; Prabhakar, S. An introduction to biometric recognition. IEEE Trans. Circuits Syst. Video Technol. 2004, 14, 4–20. [Google Scholar] [CrossRef]
  2. Zhang, D.; Lu, G. 3D Biometrics; Springer: New York, NY, USA, 2013. [Google Scholar]
  3. Moons, T.; Van Gool, L.; Vergauwen, M. 3D reconstruction from multiple images part 1: Principles. Found. Trends® Comput. Graph. Vis. 2010, 4, 287–404. [Google Scholar] [CrossRef]
  4. Ulrich, L.; Vezzetti, E.; Moos, S.; Marcolin, F. Analysis of RGB-D camera technologies for supporting different facial usage scenarios. Multimed. Tools Appl. 2020, 79, 29375–29398. [Google Scholar] [CrossRef]
  5. Zollhöfer, M.; Stotko, P.; Görlitz, A.; Theobalt, C.; Nießner, M.; Klein, R.; Kolb, A. State of the art on 3D reconstruction with RGB-D cameras. Comput. Graph. Forum 2018, 37, 625–652. [Google Scholar] [CrossRef]
  6. Martín-Martín, A.; Orduna-Malea, E.; Thelwall, M.; López-Cózar, E.D. Google Scholar, Web of Science, and Scopus: A Systematic Comparison of Citations in 252 Subject Categories. J. Inf. 2018, 12, 1160–1177. [Google Scholar] [CrossRef]
  7. Yuan, L.; Mu, Z.C.; Yang, F. A review of recent advances in ear recognition. In Chinese Conference on Biometric Recognition; Springer: Berlin/Heidelberg, Germany, 2011; pp. 252–259. [Google Scholar]
  8. Islam, S.M.; Bennamoun, M.; Owens, R.A.; Davies, R. A review of recent advances in 3D ear-and expression-invariant face biometrics. ACM Comput. Surv. (CSUR) 2012, 44, 1–34. [Google Scholar] [CrossRef]
  9. Osuna, E.; Freund, R.; Girosit, F. Training support vector machines: An application to face detection. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Juan, PR, USA, 17–19 June 1997; pp. 130–136. [Google Scholar]
  10. Nair, P.; Cavallaro, A. 3-D face detection, landmark localization, and registration using a point distribution model. IEEE Trans. Multimed. 2009, 11, 611–623. [Google Scholar] [CrossRef]
  11. Pears, N. RBF shape histograms and their application to 3D face processing. In Proceedings of the 2008 8th IEEE International Conference on Automatic Face & Gesture Recognition, Amsterdam, The Netherlands, 17–19 September 2008; pp. 1–8. [Google Scholar]
  12. Mpiperis, I.; Malasiotis, S.; Strintzis, M.G. 3D face recognition by point signatures and iso-contours. In Proceedings of the Fourth IASTED International Conference on Signal Processing, Pattern Recognition, and Applications, Anaheim, CA, USA, 14–16 February 2007. [Google Scholar]
  13. Chen, Y.; Medioni, G. Object modelling by registration of multiple range images. Image Vis. Comput. 1992, 10, 145–155. [Google Scholar] [CrossRef]
  14. Chen, H.; Bhanu, B. Shape model-based 3D ear detection from side face range images. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05)-Workshops, San Diego, CA, USA, 21–23 September 2005; p. 122. [Google Scholar]
  15. Subban, R.; Mankame, D.P. Human face recognition biometric techniques: Analysis and review. In Recent Advances in Intelligent Informatics; Springer: Cham, Switzerland, 2014; pp. 455–463. [Google Scholar]
  16. Huang, D.; Ardabilian, M.; Wang, Y.; Chen, L. 3-D face recognition using eLBP-based facial description and local feature hybrid matching. IEEE Trans. Inf. Forensics Secur. 2012, 7, 1551–1565. [Google Scholar] [CrossRef]
  17. Inan, T.; Halici, U. 3-D face recognition with local shape descriptors. IEEE Trans. Inf. Forensics Secur. 2012, 7, 577–587. [Google Scholar] [CrossRef]
  18. Zhang, C.; Cohen, F.S. 3-D face structure extraction and recognition from images using 3-D morphing and distance mapping. IEEE Trans. Image Process. 2002, 11, 1249–1259. [Google Scholar] [CrossRef] [PubMed]
  19. Chang, K.; Bowyer, K.W.; Sarkar, S.; Victor, B. Comparison and combination of ear and face images in appearance-based biometrics. IEEE Trans. Pattern Anal. Mach. Intell. 2003, 25, 1160–1165. [Google Scholar] [CrossRef]
  20. Alyuz, N.; Gokberk, B.; Akarun, L. Robust 3D Face Identification in the Presence of Occlusions. In Face Recognition in Adverse Conditions; IGI Global: Hershey, PA, USA, 2014; pp. 124–146. [Google Scholar]
  21. Balaban, S. Deep learning and face recognition: The state of the art. In Proceedings of the Biometric and Surveillance Technology for Human and Activity Identification XII, Baltimore, MD, USA, 20–24 April 2015. [Google Scholar]
  22. Schroff, F.; Kalenichenko, D.; Philbin, J. Facenet: A unified embedding for face recognition and clustering. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 815–823. [Google Scholar]
  23. Huang, G.B.; Mattar, M.; Berg, T.; Learned-Miller, E. Labeled faces in the wild: A database forstudying face recognition in unconstrained environments. In Proceedings of the Workshop on Faces in ‘Real-Life’ Images: Detection, Alignment, and Recognition, Marseille, France, 12–18 October 2008. [Google Scholar]
  24. Liu, X.; Sun, X.; He, R.; Tan, T. Recent advances on cross-domain face recognition. In Chinese Conference on Biometric Recognition; Springer: Cham, Switzerland, 2016; pp. 147–157. [Google Scholar]
  25. Bagga, M.; Singh, B. Spoofing detection in face recognition: A review. In Proceedings of the 2016 3rd International Conference on Computing for Sustainable Global Development (INDIACom), New Delhi, India, 16–18 March 2016; pp. 2037–2042. [Google Scholar]
  26. Kollreider, K.; Fronthaler, H.; Faraj, M.I.; Bigun, J. Real-time face detection and motion analysis with application in “liveness” assessment. IEEE Trans. Inf. Forensics Secur. 2007, 2, 548–558. [Google Scholar] [CrossRef]
  27. Bao, W.; Li, H.; Li, N.; Jiang, W. A liveness detection method for face recognition based on optical flow field. In Proceedings of the 2009 International Conference on Image Analysis and Signal Processing, Linhai, China, 11–12 April 2009; pp. 233–236. [Google Scholar]
  28. Mahmood, Z.; Muhammad, N.; Bibi, N.; Ali, T. A review on state-of-the-art face recognition approaches. Fractals 2017, 25, 1750025. [Google Scholar] [CrossRef]
  29. Jin, Y.; Wang, Y.; Ruan, Q.; Wang, X. A new scheme for 3D face recognition based on 2D Gabor Wavelet Transform plus LBP. In Proceedings of the 2011 6th International Conference on Computer Science & Education (ICCSE), Singapore, 3–5 August 2011; pp. 860–865. [Google Scholar]
  30. Alyuz, N.; Gokberk, B.; Akarun, L. 3-D face recognition under occlusion using masked projection. IEEE Trans. Inf. Forensics Secur. 2013, 8, 789–802. [Google Scholar] [CrossRef]
  31. Mohammadzade, H.; Hatzinakos, D. Iterative closest normal point for 3D face recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 35, 381–397. [Google Scholar] [CrossRef]
  32. Jahanbin, S.; Choi, H.; Bovik, A.C. Passive multimodal 2-D+ 3-D face recognition using Gabor features and landmark distances. IEEE Trans. Inf. Forensics Secur. 2011, 6, 1287–1304. [Google Scholar] [CrossRef]
  33. Huang, D.; Soltana, W.B.; Ardabilian, M.; Wang, Y.; Chen, L. Textured 3D face recognition using biological vision-based facial representation and optimized weighted sum fusion. In Proceedings of the CVPR Workshops, Colorado Springs, CO, USA, 20–25 June 2011; pp. 1–8. [Google Scholar]
  34. Albakri, G.; Alghowinem, S. The effectiveness of depth data in liveness face authentication using 3D sensor cameras. Sensors 2019, 19, 1928. [Google Scholar] [CrossRef]
  35. Wu, B.; Pan, M.; Zhang, Y. A review of face anti-spoofing and its applications in china. In International Conference on Harmony Search Algorithm; Springer: Cham, Switzerland, 2019; pp. 35–43. [Google Scholar]
  36. Jourabloo, A.; Liu, Y.; Liu, X. Face de-spoofing: Anti-spoofing via noise modeling. In European Conference on Computer Vision (ECCV); Springer: Cham, Switzerland, 2018; pp. 290–306. [Google Scholar]
  37. Liu, A.; Li, X.; Wan, J.; Liang, Y.; Escalera, S.; Escalante, H.J.; Madadi, M.; Jin, Y.; Wu, Z.; Yu, X.; et al. Cross-ethnicity face anti-spoofing recognition challenge: A review. IET Biom. 2021, 10, 24–43. [Google Scholar] [CrossRef]
  38. Labati, R.D.; Genovese, A.; Piuri, V.; Scotti, F. Touchless fingerprint biometrics: A survey on 2D and 3D technologies. J. Internet Technol. 2014, 15, 325–332. [Google Scholar]
  39. Jung, J.; Lee, W.; Kang, W.; Shin, E.; Ryu, J.; Choi, H. Review of piezoelectric micromachined ultrasonic transducers and their applications. J. Micromech. Microeng. 2017, 27, 113001. [Google Scholar] [CrossRef]
  40. Iula, A. Ultrasound systems for biometric recognition. Sensors 2019, 19, 2317. [Google Scholar] [CrossRef] [PubMed]
  41. Yu, Y.; Wang, H.; Sun, H.; Zhang, Y.; Chen, P.; Liang, R. Optical Coherence Tomography in Fingertip Biometrics. Opt. Lasers Eng. 2022, 151, 106868. [Google Scholar] [CrossRef]
  42. Erdogmus, N.; Marcel, S. Spoofing face recognition with 3D masks. IEEE Trans. Inf. Forensics Secur. 2014, 9, 1084–1097. [Google Scholar] [CrossRef]
  43. Lee, W.J.; Wilkinson, C.M.; Hwang, H.S. An accuracy assessment of forensic computerized facial reconstruction employing cone-beam computed tomography from live subjects. J. Forensic Sci. 2012, 57, 318–327. [Google Scholar] [CrossRef] [PubMed]
  44. Sun, Z.L.; Lam, K.M.; Gao, Q.W. Depth estimation of face images using the nonlinear least-squares model. IEEE Trans. Image Process. 2012, 22, 17–30. [Google Scholar] [PubMed]
  45. Raghavendra, R.; Raja, K.B.; Pflug, A.; Yang, B.; Busch, C. 3D face reconstruction and multimodal person identification from video captured using smartphone camera. In Proceedings of the 2013 IEEE International Conference on Technologies for Homeland Security (HST), Waltham, MA, USA, 12–14 November 2013; pp. 552–557. [Google Scholar]
  46. Dou, P.; Zhang, L.; Wu, Y.; Shah, S.K.; Kakadiaris, I.A. Pose-robust face signature for multi-view face recognition. In Proceedings of the 2015 IEEE 7th International Conference on Biometrics Theory, Applications and Systems (BTAS), Arlington, VA, USA, 8–11 September 2015; pp. 1–8. [Google Scholar]
  47. Wang, K.; Wang, X.; Pan, Z.; Liu, K. A two-stage framework for 3D face reconstruction from RGBD images. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 36, 1493–1504. [Google Scholar] [CrossRef]
  48. Hu, G.; Mortazavian, P.; Kittler, J.; Christmas, W. A facial symmetry prior for improved illumination fitting of 3D morphable model. In Proceedings of the 2013 International Conference on Biometrics (ICB), Madrid, Spain, 4–7 June 2013; pp. 1–6. [Google Scholar]
  49. Fooprateepsiri, R.; Kurutach, W. A general framework for face reconstruction using single still image based on 2D-to-3D transformation kernel. Forensic Sci. Int. 2014, 236, 117–126. [Google Scholar] [CrossRef]
  50. Wang, Y.; Chen, S.; Li, W.; Huang, D.; Wang, Y. Face anti-spoofing to 3D masks by combining texture and geometry features. In Chinese Conference on Biometric Recognition; Springer: Cham, Switzerland, 2018; pp. 399–408. [Google Scholar]
  51. Segundo, M.P.; Silva, L.; Bellon, O.R.P. Improving 3d face reconstruction from a single image using half-frontal face poses. In Proceedings of the 2012 19th IEEE International Conference on Image Processing, Orlando, FL, USA, 30 September–3 October 2012; pp. 1797–1800. [Google Scholar]
  52. Dou, P.; Kakadiaris, I.A. Multi-view 3D face reconstruction with deep recurrent neural networks. Image Vis. Comput. 2018, 80, 80–91. [Google Scholar] [CrossRef]
  53. Li, J.; Long, S.; Zeng, D.; Zhao, Q. Example-based 3D face reconstruction from uncalibrated frontal and profile images. In Proceedings of the 2015 International Conference on Biometrics (ICB), Phuket, Thailand, 19–22 May 2015; pp. 193–200. [Google Scholar]
  54. Abate, A.F.; Narducci, F.; Ricciardi, S. Biometrics empowered ambient intelligence environment. Atti Accad. Peloritana Pericolanti-Cl. Sci. Fis. Mat. Nat. 2015, 93, 4. [Google Scholar]
  55. van Dam, C.; Veldhuis, R.; Spreeuwers, L. Landmark-based model-free 3d face shape reconstruction from video sequences. In Proceedings of the 2013 International Conference of the BIOSIG Special Interest Group (BIOSIG), Darmstadt, Germany, 5–6 September 2013; pp. 1–5. [Google Scholar]
  56. Betta, G.; Capriglione, D.; Corvino, M.; Lavatelli, A.; Liguori, C.; Sommella, P.; Zappa, E. Metrological characterization of 3D biometric face recognition systems in actual operating conditions. Acta IMEKO 2017, 6, 33–42. [Google Scholar] [CrossRef]
  57. Yin, J.; Yang, X. 3D facial reconstruction of based on OpenCV and DirectX. In Proceedings of the 2016 International Conference on Audio, Language and Image Processing (ICALIP), Shanghai, China, 11–12 July 2016; pp. 341–344. [Google Scholar]
  58. Abate, A.F.; Nappi, M.; Ricciardi, S. A biometric interface to ambient intelligence environments. In Information Systems: Crossroads for Organization, Management, Accounting and Engineering; Springer: Berlin/Heidelberg, Germany, 2012; pp. 155–163. [Google Scholar]
  59. Crispim, F.; Vieira, T.; Lima, B. Verifying kinship from rgb-d face data. In International Conference on Advanced Concepts for Intelligent Vision Systems; Springer: Cham, Switzerland, 2020; pp. 215–226. [Google Scholar]
  60. Naveen, S.; Rugmini, K.; Moni, R. 3D face reconstruction by pose correction, patch cloning and texture wrapping. In Proceedings of the 2016 International Conference on Communication Systems and Networks (ComNet), Thiruvananthapuram, India, 21–23 July 2016; pp. 112–116. [Google Scholar]
  61. Zhang, Z.; Zhang, M.; Chang, Y.; Esche, S.K.; Chassapis, C. A virtual laboratory combined with biometric authentication and 3D reconstruction. In Proceedings of the ASME International Mechanical Engineering Congress and Exposition, Phoenix, AZ, USA, 11–17 November 2016; Volume 50571, p. V005T06A049. [Google Scholar]
  62. Rychlik, M.; Stankiewicz, W.; Morzynski, M. 3D facial biometric database–search and reconstruction of objects based on PCA modes. In International Conference on Universal Access in Human-Computer Interaction; Springer: Cham, Switzerland, 2014; pp. 125–136. [Google Scholar]
  63. Tahiri, M.A.; Karmouni, H.; Tahiri, A.; Sayyouri, M.; Qjidaa, H. Partial 3D Image Reconstruction by Cuboids Using Stable Computation of Hahn Polynomials. In WITS 2020; Springer: Singapore, 2022; pp. 831–842. [Google Scholar]
  64. Xiong, W.; Yang, H.; Zhou, P.; Fu, K.; Zhu, J. Spatiotemporal Correlation-Based Accurate 3D Face Imaging Using Speckle Projection and Real-Time Improvement. Appl. Sci. 2021, 11, 8588. [Google Scholar] [CrossRef]
  65. Abate, A.F.; De Maio, L.; Distasi, R.; Narducci, F. Remote 3D face reconstruction by means of autonomous unmanned aerial vehicles. Pattern Recognit. Lett. 2021, 147, 48–54. [Google Scholar] [CrossRef]
  66. Li, X.; Wu, S. Multi-attribute regression network for face reconstruction. In Proceedings of the 2020 25th International Conference on Pattern Recognition (ICPR), Milan, Italy, 10–15 January 2021; pp. 7226–7233. [Google Scholar]
  67. Kneis, B.; Zhang, W. 3D Face Recognition using Photometric Stereo and Deep Learning. In Proceedings of the 10th International Conference on Web Intelligence, Mining and Semantics, Biarritz, France, 30 June–3 July 2020; pp. 255–261. [Google Scholar]
  68. Li, J.; Liu, Z.; Zhao, Q. Exploring shape deformation in 2D images for facial expression recognition. In Chinese Conference on Biometric Recognition; Springer: Cham, Switzerland, 2019; pp. 190–197. [Google Scholar]
  69. Sopiak, D.; Oravec, M.; Pavlovičová, J.; Bukovčíková, Z.; Dittingerová, M.; Bil’anská, A.; Novotná, M.; Gontkovič, J. Generating face images based on 3D morphable model. In Proceedings of the 2016 12th International Conference on Natural Computation, Fuzzy Systems and Knowledge Discovery (ICNC-FSKD), Changsha, China, 13–15 August 2016; pp. 58–62. [Google Scholar]
  70. Chuchvara, A.; Georgiev, M.; Gotchev, A. A framework for fast low-power multi-sensor 3D scene capture and reconstruction. In International Workshop on Biometric Authentication; Springer: Cham, Switzerland, 2014; pp. 40–53. [Google Scholar]
  71. Narayana, S.; Rohit; Rajagopal; Rakshith; Antony, J. 3D face reconstruction using frontal and profile views. In Proceedings of the 2013 7th Asia Modelling Symposium, Hong Kong, China, 23–25 July 2013; pp. 132–136. [Google Scholar]
  72. Kumar, A.; Kwong, C. Towards Contactless, Low-Cost and Accurate 3D Fingerprint Identification. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 681–696. [Google Scholar] [CrossRef] [PubMed]
  73. Labati, R.D.; Genovese, A.; Piuri, V.; Scotti, F. Toward unconstrained fingerprint recognition: A fully touchless 3-D system based on two views on the move. IEEE Trans. Syst. Man Cybern. Syst. 2015, 46, 202–219. [Google Scholar] [CrossRef]
  74. Liu, F.; Zhang, D. 3D fingerprint reconstruction system using feature correspondences and prior estimated finger model. Pattern Recognit. 2014, 47, 178–193. [Google Scholar] [CrossRef]
  75. Xu, J.; Hu, J. Direct Feature Point Correspondence Discovery for Multiview Images: An Alternative Solution When SIFT-Based Matching Fails. In International Conference on Testbeds and Research Infrastructures; Springer: Cham, Switzerland, 2016; pp. 137–147. [Google Scholar]
  76. Chatterjee, A.; Bhatia, V.; Prakash, S. Anti-spoof touchless 3D fingerprint recognition system using single shot fringe projection and biospeckle analysis. Opt. Lasers Eng. 2017, 95, 1–7. [Google Scholar] [CrossRef]
  77. Yin, X.; Zhu, Y.; Hu, J. 3D fingerprint recognition based on ridge-valley-guided 3D reconstruction and 3D topology polymer feature extraction. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 43, 1085–1091. [Google Scholar] [CrossRef]
  78. Kang, W.; Liu, H.; Luo, W.; Deng, F. Study of a full-view 3D finger vein verification technique. IEEE Trans. Inf. Forensics Secur. 2019, 15, 1175–1189. [Google Scholar] [CrossRef]
  79. Ma, Z.; Fang, L.; Duan, J.; Xie, S.; Wang, Z. Personal identification based on finger vein and contour point clouds matching. In Proceedings of the 2016 IEEE International Conference on Mechatronics and Automation, Harbin, China, 7–10 August 2016; pp. 1983–1988. [Google Scholar]
  80. Veldhuis, R.; Spreeuwers, L.; Ton, B.; Rozendal, S. A high-quality finger vein dataset collected using a custom-designed capture device. In Handbook of Vascular Biometrics; Springer: Cham, Switzerland, 2020; pp. 63–75. [Google Scholar]
  81. Xu, H.; Yang, W.; Wu, Q.; Kang, W. Endowing rotation invariance for 3d finger shape and vein verification. Front. Comput. Sci. 2022, 16, 1–16. [Google Scholar] [CrossRef]
  82. Cho, S.Y. 3D ear shape reconstruction and recognition for biometric applications. Signal Image Video Process. 2013, 7, 609–618. [Google Scholar] [CrossRef]
  83. Liu, C.; Mu, Z.; Wang, K.; Zeng, H. 3D ear modeling based on SFS. In Proceedings of the 10th World Congress on Intelligent Control and Automation, Beijing, China, 6–8 July 2012; pp. 4837–4841. [Google Scholar]
  84. Bastias, D.; Perez, C.A.; Benalcazar, D.P.; Bowyer, K.W. A method for 3D iris reconstruction from multiple 2D near-infrared images. In Proceedings of the 2017 IEEE International Joint Conference on Biometrics (IJCB), Denver, CO, USA, 1–4 October 2017; pp. 503–509. [Google Scholar]
  85. Benalcazar, D.P.; Zambrano, J.E.; Bastias, D.; Perez, C.A.; Bowyer, K.W. A 3D iris scanner from a single image using convolutional neural networks. IEEE Access 2020, 8, 98584–98599. [Google Scholar] [CrossRef]
  86. Benalcazar, D.P.; Bastias, D.; Perez, C.A.; Bowyer, K.W. A 3D iris scanner from multiple 2D visible light images. IEEE Access 2019, 7, 61461–61472. [Google Scholar] [CrossRef]
  87. López-Fernández, D.; Madrid-Cuevas, F.J.; Carmona-Poyato, A.; Muñoz-Salinas, R.; Medina-Carnicer, R. Entropy volumes for viewpoint-independent gait recognition. Mach. Vis. Appl. 2015, 26, 1079–1094. [Google Scholar] [CrossRef]
  88. López-Fernández, D.; Madrid-Cuevas, F.J.; Carmona-Poyato, A.; Marín-Jiménez, M.J.; Muñoz-Salinas, R.; Medina-Carnicer, R. Independent gait recognition through morphological descriptions of 3D human reconstructions. Image Vis. Comput. 2016, 48, 1–13. [Google Scholar] [CrossRef]
  89. Imoto, D.; Kurosawa, K.; Honma, M.; Yokota, R.; Hirabayashi, M.; Hawai, Y. Model-Based Interpolation for Continuous Human Silhouette Images by Height-Constraint Assumption. In Proceedings of the 2020 4th International Conference on Vision, Image and Signal Processing, Bangkok, Thailand, 9–11 December 2020; pp. 1–11. [Google Scholar]
  90. Lorkiewicz-Muszyńska, D.; Kociemba, W.; Żaba, C.; Łabęcka, M.; Koralewska-Kordel, M.; Abreu-Głowacka, M.; Przystańska, A. The conclusive role of postmortem computed tomography (CT) of the skull and computer-assisted superimposition in identification of an unknown body. Int. J. Leg. Med. 2013, 127, 653–660. [Google Scholar] [CrossRef]
  91. Svoboda, J.; Klubal, O.; Drahanskỳ, M. Biometric recognition of people by 3D hand geometry. In Proceedings of the International Conference on Digital Technologies 2013, Zilina, Slovakia, 29–31 May 2013; pp. 137–141. [Google Scholar]
  92. Chu, B.; Romdhani, S.; Chen, L. 3D-aided face recognition robust to expression and pose variations. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 1899–1906. [Google Scholar]
  93. Parkhi, O.M.; Vedaldi, A.; Zisserman, A. Deep Face Recognition. In Proceedings of the British Machine Vision Conference, Swansea, UK, 7–10 September 2015; pp. 41.1–41.12. [Google Scholar]
  94. Victor, B.; Bowyer, K.; Sarkar, S. An evaluation of face and ear biometrics. In Proceedings of the 2002 International Conference on Pattern Recognition, Quebec City, QC, Canada, 11–15 August 2002; Volume 1, pp. 429–432. [Google Scholar]
  95. Abdi, H.; Williams, L.J. Principal component analysis. Wiley Interdiscip. Rev. Comput. Stat. 2010, 2, 433–459. [Google Scholar] [CrossRef]
  96. Fisher, R.A. The use of multiple measurements in taxonomic problems. Ann. Eugen. 1936, 7, 179–188. [Google Scholar] [CrossRef]
  97. Ghojogh, B.; Karray, F.; Crowley, M. Fisher and kernel Fisher discriminant analysis: Tutorial. arXiv 2019, arXiv:1906.09436. [Google Scholar]
  98. Li, C.; Mu, Z.; Zhang, F.; Wang, S. A novel 3D ear reconstruction method using a single image. In Proceedings of the 10th World Congress on Intelligent Control and Automation, Beijing, China, 6–8 July 2012; pp. 4891–4896. [Google Scholar]
  99. Blanz, V.; Vetter, T. A morphable model for the synthesis of 3D faces. In Proceedings of the 26th Annual Conference on Computer Graphics and Interactive Techniques, Los Angeles, CA, USA, 8–13 August 1999; pp. 187–194. [Google Scholar]
  100. Daugman, J. How iris recognition works. In The Essential Guide to Image Processing; Elsevier: Amsterdam, The Netherlands, 2009; pp. 715–739. [Google Scholar]
  101. Moulon, P.; Bezzi, A. Python Photogrammetry Toolbox: A Free Solution for Three-Dimensional Documentation; ArcheoFoss: Napoli, Italiy, 2011; pp. 1–12. [Google Scholar]
  102. Zheng, C.; Cham, T.J.; Cai, J. T2net: Synthetic-to-realistic translation for solving single-image depth estimation tasks. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 767–783. [Google Scholar]
  103. Alhashim, I.; Wonka, P. High quality monocular depth estimation via transfer learning. arXiv 2018, arXiv:1812.11941. [Google Scholar]
  104. Whitaker, L.A. Anthropometry of the Head and Face in Medicine. Plast. Reconstr. Surg. 1983, 71, 144–145. [Google Scholar] [CrossRef]
  105. Vanezis, P.; Blowes, R.; Linney, A.; Tan, A.; Richards, R.; Neave, R. Application of 3-D computer graphics for facial reconstruction and comparison with sculpting techniques. Forensic Sci. Int. 1989, 42, 69–84. [Google Scholar] [CrossRef]
  106. Vandermeulen, D.; Claes, P.; De Greef, S.; Willems, G.; Clement, J.; Suetens, P. Automated facial reconstruction. In Craniofacial Identification; Cambridge University Press: Cambridge, UK, 2012; Volume 203. [Google Scholar]
  107. Vezzetti, E.; Marcolin, F.; Tornincasa, S.; Moos, S.; Violante, M.G.; Dagnes, N.; Monno, G.; Uva, A.E.; Fiorentino, M. Facial landmarks for forensic skull-based 3D face reconstruction: A literature review. In International Conference on Augmented Reality, Virtual Reality and Computer Graphics; Springer: Cham, Switzerland, 2016; pp. 172–180. [Google Scholar]
  108. Yin, X.; Zhu, Y.; Hu, J. A Survey on 2D and 3D Contactless Fingerprint Biometrics: A Taxonomy, Review, and Future Directions. IEEE Open J. Comput. Soc. 2021, 2, 370–381. [Google Scholar] [CrossRef]
  109. Gorthi, S.S.; Rastogi, P. Fringe projection techniques: Whither we are? Opt. Lasers Eng. 2010, 48, 133–140. [Google Scholar] [CrossRef]
  110. Stoykova, E.; Nazarova, D.; Berberova, N.; Gotchev, A. Performance of intensity-based non-normalized pointwise algorithms in dynamic speckle analysis. Opt. Express 2015, 23, 25128–25142. [Google Scholar] [CrossRef] [PubMed]
  111. Lowe, D.G. Distinctive image features from scale-invariant keypoints. Int. J. Comput. Vis. 2004, 60, 91–110. [Google Scholar] [CrossRef]
  112. Fischler, M.A.; Bolles, R.C. Random sample consensus: A paradigm for model fitting with applications to image analysis and automated cartography. Commun. ACM 1981, 24, 381–395. [Google Scholar] [CrossRef]
  113. Storn, R.; Price, K. Differential evolution–a simple and efficient heuristic for global optimization over continuous spaces. J. Glob. Optim. 1997, 11, 341–359. [Google Scholar] [CrossRef]
  114. Qin, A.K.; Huang, V.L.; Suganthan, P.N. Differential evolution algorithm with strategy adaptation for global numerical optimization. IEEE Trans. Evol. Comput. 2008, 13, 398–417. [Google Scholar] [CrossRef]
  115. Zhu, Y.; Yin, X.; Jia, X.; Hu, J. Latent fingerprint segmentation based on convolutional neural networks. In Proceedings of the 2017 IEEE Workshop on Information Forensics and Security (WIFS), Rennes, France, 4–7 December 2017; pp. 1–6. [Google Scholar]
  116. Chikkerur, S.; Cartwright, A.N.; Govindaraju, V. Fingerprint enhancement using stft analysis. Pattern Recognit. 2007, 40, 198–211. [Google Scholar] [CrossRef]
  117. Sidiropoulos, G.K.; Kiratsa, P.; Chatzipetrou, P.; Papakostas, G.A. Feature Extraction for Finger-Vein-Based Identity Recognition. J. Imaging 2021, 7, 89. [Google Scholar] [CrossRef] [PubMed]
  118. Zhang, Z.; Wang, M. A Simple and Efficient Method for Finger Vein Recognition. Sensors 2022, 22, 2234. [Google Scholar] [CrossRef] [PubMed]
  119. Asaari, M.S.M.; Suandi, S.A.; Rosdi, B.A. Fusion of band limited phase only correlation and width centroid contour distance for finger based biometrics. Expert Syst. Appl. 2014, 41, 3367–3382. [Google Scholar] [CrossRef]
  120. Zuiderveld, K. Contrast limited adaptive histogram equalization. In Graphics Gems; Academic Press Professional, Inc.: San Diego, CA, USA, 1994; pp. 474–485. [Google Scholar]
  121. Cutting, J.E.; Kozlowski, L.T. Recognizing friends by their walk: Gait perception without familiarity cues. Bull. Psychon. Soc. 1977, 9, 353–356. [Google Scholar] [CrossRef] [Green Version]
  122. Horprasert, T.; Harwood, D.; Davis, L.S. A statistical approach for real-time robust background subtraction and shadow detection. In Proceedings of the IEEE ICCV’99 FRAME-RATE Workshop, Corfu, Greece, 20–25 September 1999; pp. 1–19. [Google Scholar]
  123. Koestinger, M.; Wohlhart, P.; Roth, P.M.; Bischof, H. Annotated facial landmarks in the wild: A large-scale, real-world database for facial landmark localization. In Proceedings of the 2011 IEEE International Conference on Computer Vision Workshops (ICCV Workshops), Barcelona, Spain, 6–13 November 2011; pp. 2144–2151. [Google Scholar]
  124. Erdogmus, N.; Marcel, S. Spoofing 2D face recognition systems with 3D masks. In Proceedings of the 2013 International Conference of the BIOSIG Special Interest Group (BIOSIG), Darmstadt, Germany, 5–6 September 2013; pp. 1–8. [Google Scholar]
  125. Savran, A.; Alyüz, N.; Dibeklioğlu, H.; Çeliktutan, O.; Gökberk, B.; Sankur, B.; Akarun, L. Bosphorus database for 3D face analysis. In European Workshop on Biometrics and Identity Management; Springer: Berlin/Heidelberg, Germany, 2008; pp. 47–56. [Google Scholar]
  126. Yin, L.; Wei, X.; Sun, Y.; Wang, J.; Rosato, M.J. A 3D facial expression database for facial behavior research. In Proceedings of the 7th International Conference on Automatic Face and Gesture Recognition (FGR06), Southampton, UK, 10–12 April 2006; pp. 211–216. [Google Scholar]
  127. Zhang, X.; Yin, L.; Cohn, J.F.; Canavan, S.; Reale, M.; Horowitz, A.; Liu, P. A high-resolution spontaneous 3D dynamic facial expression database. In Proceedings of the 2013 10th IEEE International Conference and Workshops on Automatic Face and Gesture Recognition (FG), Shanghai, China, 22–26 April 2013; pp. 1–6. [Google Scholar]
  128. Phillips, P.J.; Moon, H.; Rizvi, S.A.; Rauss, P.J. The FERET evaluation methodology for face-recognition algorithms. IEEE Trans. Pattern Anal. Mach. Intell. 2000, 22, 1090–1104. [Google Scholar] [CrossRef]
  129. Phillips, P.J.; Scruggs, W.T.; O’Toole, A.J.; Flynn, P.J.; Bowyer, K.W.; Schott, C.L.; Sharpe, M. FRVT 2006 and ICE 2006 large-scale experimental results. IEEE Trans. Pattern Anal. Mach. Intell. 2009, 32, 831–846. [Google Scholar] [CrossRef] [PubMed]
  130. Tabula Rasa (Trusted Biometrics under Spoofing Attacks). Available online: http://www.tabularasa-euproject.org/contact-info (accessed on 16 February 2022).
  131. Zafeiriou, S.; Hansen, M.; Atkinson, G.; Argyriou, V.; Petrou, M.; Smith, M.; Smith, L. The photoface database. In Proceedings of the CVPR 2011 WORKSHOPS, Colorado Springs, CO, USA, 20–25 June 2011; pp. 132–139. [Google Scholar]
  132. Labeled Faces in the Wild. Available online: http://vis-www.cs.umass.edu/lfw/ (accessed on 16 February 2022).
  133. Wolf, L.; Hassner, T.; Maoz, I. Face recognition in unconstrained videos with matched background similarity. In Proceedings of the CVPR 2011, Colorado Springs, CO, USA, 20–25 June 2011; pp. 529–534. [Google Scholar]
  134. Sim, T.; Baker, S.; Bsat, M. The CMU pose, illumination, and expression (PIE) database. In Proceedings of the Fifth IEEE International Conference on Automatic Face Gesture Recognition, Washington, DC, USA, 21 May 2002; pp. 53–58. [Google Scholar]
  135. Toderici, G.; Evangelopoulos, G.; Fang, T.; Theoharis, T.; Kakadiaris, I.A. UHDB11 database for 3D-2D face recognition. In Pacific-Rim Symposium on Image and Video Technology; Springer: Berlin/Heidelberg, Germany, 2013; pp. 73–86. [Google Scholar]
  136. Kumar, A.; Wu, C. Automated human identification using ear imaging. Pattern Recognit. 2012, 45, 956–968. [Google Scholar] [CrossRef]
  137. AMI Ear Database. Available online: https://ctim.ulpgc.es/research_works/ami_ear_database/ (accessed on 19 February 2022).
  138. Chen, H.; Bhanu, B. Human ear recognition in 3D. IEEE Trans. Pattern Anal. Mach. Intell. 2007, 29, 718–737. [Google Scholar] [CrossRef]
  139. Yan, P.; Bowyer, K.W. Biometric recognition using 3D ear shape. IEEE Trans. Pattern Anal. Mach. Intell. 2007, 29, 1297–1308. [Google Scholar] [CrossRef]
  140. Messer, K.; Matas, J.; Kittler, J.; Luettin, J.; Maitre, G. XM2VTSDB: The extended M2VTS database. In Proceedings of the Second International Conference on Audio and Video-Based Biometric Person Authentication, Washington, DC, USA, 22–24 March 1999; Volume 964, pp. 965–966. [Google Scholar]
  141. López-Fernández, D.; Madrid-Cuevas, F.J.; Carmona-Poyato, Á.; Marín-Jiménez, M.J.; Muñoz-Salinas, R. The AVA multi-view dataset for gait recognition. In International Workshop on Activity Monitoring by Multiple Distributed Sensing; Springer: Cham, Switzerland, 2014; pp. 26–39. [Google Scholar]
  142. Iwashita, Y.; Ogawara, K.; Kurazume, R. Identification of people walking along curved trajectories. Pattern Recognit. Lett. 2014, 48, 60–69. [Google Scholar] [CrossRef]
  143. Gkalelis, N.; Kim, H.; Hilton, A.; Nikolaidis, N.; Pitas, I. The i3dpost multi-view and 3d human action/interaction database. In Proceedings of the 2009 Conference for Visual Media Production, London, UK, 12–13 November 2009; pp. 159–168. [Google Scholar]
  144. Singh, S.; Velastin, S.A.; Ragheb, H. Muhavi: A multicamera human action video dataset for the evaluation of action recognition methods. In Proceedings of the 2010 7th IEEE International Conference on Advanced Video and Signal Based Surveillance, Boston, MA, USA, 29 August–1 September 2010; pp. 48–55. [Google Scholar]
  145. Weinland, D.; Ronfard, R.; Boyer, E. Free viewpoint action recognition using motion history volumes. Comput. Vis. Image Underst. 2006, 104, 249–257. [Google Scholar] [CrossRef]
  146. IIT Delhi Iris Database. Available online: http://https://www4.comp.polyu.edu.hk/~csajaykr/IITD/Database_Iris.htm (accessed on 20 February 2022).
  147. The Hong Kong Polytechnic University 3D Fingerprint Images Database. Available online: https://tinyurl.com/4nmb2keb (accessed on 19 February 2022).
  148. Tharewal, S.; Malche, T.; Tiwari, P.K.; Jabarulla, M.Y.; Alnuaim, A.A.; Mostafa, A.M.; Ullah, M.A. Score-Level Fusion of 3D Face and 3D Ear for Multimodal Biometric Human Recognition. Comput. Intell. Neurosci. 2022, 2022, 3019194. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Three-Dimensional (3D) Taxonomy.
Figure 1. Three-Dimensional (3D) Taxonomy.
Sensors 22 06364 g001
Figure 2. 3D Biometric Treemap.
Figure 2. 3D Biometric Treemap.
Sensors 22 06364 g002
Figure 3. Three-Dimensional (3D) Biometrics Major Categories.
Figure 3. Three-Dimensional (3D) Biometrics Major Categories.
Sensors 22 06364 g003
Figure 4. Three-Dimensional (3D) Face Categories.
Figure 4. Three-Dimensional (3D) Face Categories.
Sensors 22 06364 g004
Figure 5. Three-Dimensional (3D) Hand Categories.
Figure 5. Three-Dimensional (3D) Hand Categories.
Sensors 22 06364 g005
Figure 6. Three-Dimensional (3D) Biometrics Categories.
Figure 6. Three-Dimensional (3D) Biometrics Categories.
Sensors 22 06364 g006
Figure 7. Three-Dimensional (3D) Biometrics General Flowchart.
Figure 7. Three-Dimensional (3D) Biometrics General Flowchart.
Sensors 22 06364 g007
Figure 8. Three-Dimensional (3D) Acquisition Method.
Figure 8. Three-Dimensional (3D) Acquisition Method.
Sensors 22 06364 g008
Figure 9. Three-Dimensional (3D) Acquisition Method per Biometric.
Figure 9. Three-Dimensional (3D) Acquisition Method per Biometric.
Sensors 22 06364 g009
Figure 10. Three-Dimensional (3D) Reconstruction Methods.
Figure 10. Three-Dimensional (3D) Reconstruction Methods.
Sensors 22 06364 g010
Figure 11. Applications with Stereo Vision and SfS by Biometric Category.
Figure 11. Applications with Stereo Vision and SfS by Biometric Category.
Sensors 22 06364 g011
Figure 12. 3D Reconstruction Methods per Biometric.
Figure 12. 3D Reconstruction Methods per Biometric.
Sensors 22 06364 g012
Figure 13. Facial Biometric Methods.
Figure 13. Facial Biometric Methods.
Sensors 22 06364 g013
Table 1. Literature Table.
Table 1. Literature Table.
BiometricMajor CategoryPercentage (%)References
FacialFace59.26[42,43,44,45,46,47,48,49,50,51,52,53,54,55,56,57,58,59,60,61,62,63,64,65,66,67,68,69,70,71,58,71]
FingerprintHand11.11[72,73,74,75,76,77]
Finger VeinHand7.41[78,79,80,81]
EarFace5.56[45,82,83]
IrisFace5.56[84,85,86]
GaitGait5.56[87,88,89]
SkullFace3.70[43,90]
PalmHand1.85[91]
Table 2. Three-Dimensional (3D) Biometrics Dataset Table.
Table 2. Three-Dimensional (3D) Biometrics Dataset Table.
DatasetBiometricNumber of ImagesClassesYearReference
AFLWFacial21,99725,9932011[123]
3D-MADFacial76,500172013[124]
BosphorusFacial46661052008[125]
BU-3DFEFacial25001002006[126]
BU-4DFEFacial60,6001012013[127]
FeretFacial14,12611992000[128]
FRGCFacial50,00012,5002004[129]
MorphoFacial200202013[130]
The Photoface DatabaseFacial73562612011[131]
LFWFacial13,23357492019[132]
Youtube FacesFacial3245 (videos)15952011[133]
PieFacial75,0003372002[134]
UHDB11Facial1656232013[135]
IIT-KanpurEar4651252012[136]
AMIEar7001002008[137]
UCREar9021552007[138]
UNDEar16864152007[139]
XM2VTSEar1180 (videos)2952013[140]
AVAMVGGait200 (videos)202014[141]
KY4DGait168 (videos)422014[142]
i3DPostGait768 (videos)82009[143]
MuHAViGait136 (videos)142010[144]
IXMASGait550 (videos)102006[145]
SCUT LFMB-3DPVFVFinger Vein16,8487022022[81]
IIT Iris DatabaseIris11202242007[146]
Hong Kong Polytechnic 3DFingerprint15602602016[147]
Table 3. Three-Dimensional (3D) Biometrics State of the Art.
Table 3. Three-Dimensional (3D) Biometrics State of the Art.
TitleBiometricScoreDatasetYearReference
Verifying kinship from rgb-d face dataFacial95% (accuracy)Kin3D2020[59]
A novel 3D ear
reconstruction method
using a single image
EarmanualUND2012[98]
A 3D iris scanner from a single image
using convolutional neural networks
Iris99.8% (accuracy)98,520 iris2020[85]
An accuracy assessment of forensic computerized facial reconstruction employing cone-beam computed tomography from live subjectsSkull0.31 mm (error)3 humans2012[43]
3D fingerprint recognition
based on ridge–valley
guided 3D reconstruction
and 3D topology polymer
feature extraction
Fingerprint98% (accuracy)DB1, DB22019[77]
Endowing rotation invariancefor 3D
finger shape and vein verification
Finger Vein2.61 (ER%)3DPVFV2022[81]
Biometric recognition of people by 3D
hand geometry
Palm0.0018 (RMSE)3 palms2013[91]
Model-based interpolation
for continuous human
silhouette images by
height-constraint assumption
Gait95%KY 4D2020[89]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Samatas, G.G.; Papakostas, G.A. Biometrics: Going 3D. Sensors 2022, 22, 6364. https://doi.org/10.3390/s22176364

AMA Style

Samatas GG, Papakostas GA. Biometrics: Going 3D. Sensors. 2022; 22(17):6364. https://doi.org/10.3390/s22176364

Chicago/Turabian Style

Samatas, Gerasimos G., and George A. Papakostas. 2022. "Biometrics: Going 3D" Sensors 22, no. 17: 6364. https://doi.org/10.3390/s22176364

APA Style

Samatas, G. G., & Papakostas, G. A. (2022). Biometrics: Going 3D. Sensors, 22(17), 6364. https://doi.org/10.3390/s22176364

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop