Next Article in Journal
A Driver’s Visual Attention Prediction Using Optical Flow
Previous Article in Journal
Robot-Assisted Autism Therapy (RAAT). Criteria and Types of Experiments Using Anthropomorphic and Zoomorphic Robots. Review of the Research
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Efficient and Accurate Iris Recognition Algorithm Based on a Novel Condensed 2-ch Deep Convolutional Neural Network

School of Microelectronics, Shandong University, Jinan 250100, China
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(11), 3721; https://doi.org/10.3390/s21113721
Submission received: 30 March 2021 / Revised: 1 May 2021 / Accepted: 3 May 2021 / Published: 27 May 2021
(This article belongs to the Section Physical Sensors)

Abstract

:
Recently, deep learning approaches, especially convolutional neural networks (CNNs), have attracted extensive attention in iris recognition. Though CNN-based approaches realize automatic feature extraction and achieve outstanding performance, they usually require more training samples and higher computational complexity than the classic methods. This work focuses on training a novel condensed 2-channel (2-ch) CNN with few training samples for efficient and accurate iris identification and verification. A multi-branch CNN with three well-designed online augmentation schemes and radial attention layers is first proposed as a high-performance basic iris classifier. Then, both branch pruning and channel pruning are achieved by analyzing the weight distribution of the model. Finally, fast finetuning is optionally applied, which can significantly improve the performance of the pruned CNN while alleviating the computational burden. In addition, we further investigate the encoding ability of 2-ch CNN and propose an efficient iris recognition scheme suitable for large database application scenarios. Moreover, the gradient-based analysis results indicate that the proposed algorithm is robust to various image contaminations. We comprehensively evaluated our algorithm on three publicly available iris databases for which the results proved satisfactory for real-time iris recognition.

1. Introduction

Iris texture patterns are believed to be randomly determined during fetal development of the eye and are invariant to age [1]. Hence, the iris pattern of each eye can be seen as a universally unique biometric feature even distinct between twins. As one of the most secure and reliable biometric identification techniques, iris recognition has been widely used in banking, border security control, mobile phones, etc. [2,3,4]. Compared with other mainstream biometric approaches, including face recognition [5], palmprint recognition [6], and fingerprint recognition [7], iris recognition is safer and more sanitary because of its characteristics of being non-contact and having less exposure [8]. The merits of iris recognition have prompted increasing efforts for investigating more accurate and efficient iris feature extraction algorithms under various conditions [9,10,11].
Verification and identification are the two main application scenarios for iris recognition. Given an iris image of an eye, the iris recognition system in verification mode will judge whether it is registered or not according to the previously enrolled iris images, which is usually a “one-against-one” comparison scheme. In the identification mode, the system will answer “who is he” to this iris image, which performs a “one-against-all” comparison scheme most of the time.
The deep learning methods, especially convolutional neural network (CNN), have achieved considerable success in many computer vision (CV) tasks [12,13,14,15,16]. Handcrafted feature extraction approaches have been outclassed by CNN, with its capability to automatically learn relevant features from sufficient training data [17,18]. Recent advances in iris recognition have studied the feasibility of applying the CNN to iris image processing, such as iris segmentation [19,20], iris recognition [21,22,23], and fake iris detection [24,25]. Previous studies on iris recognition [21,26] indicated that the CNN-based methods could effectively learn the inherent characteristics of iris images and achieve superior performance than the classic iris matching method represented by IrisCode [27]. The success of early efforts prompts us to further investigate the potential of deep CNN for addressing challenging problems in real-time iris recognition.

2. Related Work

The earliest automatic iris recognition system can be traced back to 1987. Flom and Safir [28] proposed a conceptual iris recognition system without implementation details and were authorized the first patent. From then on, various iris texture feature extraction and classification methods emerged, which can be approximately divided into classic handcrafted feature engineering methods and recently appeared deep learning methods.
One of the most influential iris recognition algorithms was proposed by Daugman [27,29,30]. In his pioneering works, the boundary of the pupil and iris was first detected by the integrodifferential operator and normalized by Daugman’s rubber sheet model. Then, the extracted iris was transformed into a series of encoding (which is usually referred to as IrisCode) by applying Gabor phase-quadrant feature descriptor. Finally, in the identification or verification stage, the hamming distance between IrisCode was calculated and hence attained the recognition result. Daugman’s algorithm and workflow are still widely utilized in current iris recognition systems. Later, numerous iris feature extraction approaches arose, including variations of the Gabor kernel [31,32,33], the SIFT and SURF-based features [34,35,36], feature fusion methods [37,38,39], and human-in-the-loop methods [40,41]. These methods usually yield a notable performance with few training data. Nevertheless, their feature extractors are required to be delicately designed and not robust to image contaminations such as eyelid and eyebrow, which places higher demands on image quality and preprocessing steps.
Recently, deep learning-based iris recognition approaches have been increasingly studied. Deep CNN is usually used as a feature extractor, which encodes the iris image to a set of feature vectors and then measures their distance as the aforementioned classic method does. Gangwar et al. [42] proposed a deep CNN model with less risk of overfitting for extracting the iris feature. Nguyen et al. [43] explored the encoding ability of the pre-trained CNN architecture, with results showing that the network, such as AlexNet and VGG-net trained on other large-scale image databases, can be effectively transferred to the task of iris texture feature extraction. Raja et al. [44] extracted robust multi-patch iris features by CNN with sparse filters. More recently, Wang et al. [26] and Liu et al. [45] collected iris features using dilated residual network and capsule network, respectively. In addition, deep CNN can also be directly utilized as a classifier. In this way, the pairwise training dataset was generated with all possible combinations of training samples. In the testing phase, paired images were fed into CNN, and the result was provided to examine whether the images belonged to intra-class or inter-class. With this approach, few training samples were needed for the deep neural network. This type of method was first discussed in detail in the work of Zagoruyko et al. [46], and different types of networks, including siamese, pseudo-siamese, 2-channel (2-ch) deep networks, were constructed for image patch comparison. The experimental results showed that the 2-ch network outperformed the other networks at the cost of computational complexity. Some efforts have also been focused on iris verification using 2-ch network. Liu et al. [47] proposed a 2-ch CNN architecture named DeepIris for heterogeneous iris verification. In their algorithm, six forward propagations were required to prevent rotation differences, which led to the heavy computational burden. Špetlík et al. [48] modified the 2-ch CNN with a unit-circle layer for iris verification. Proença et al. [49] integrated an iris segmentation deep learning model and a 2-ch iris classification CNN for segment-less iris verification.
Though the existing deep learning-based methods are proven to be effective for automatic end-to-end iris feature extraction and classification, several issues remain to be further addressed. For example, due to high computational complexity, the 2-ch methods have only been successfully applied to the iris verification scenario. In addition, the deep learning model is sensitive to image contamination and training data scale, which poses a challenging problem to real-time iris recognition. Furthermore, the hyperparameters in CNN architecture, such as the number of layers and kernels, have not been fully optimized.
The objective of this paper is to develop a deep learning approach with strong robustness to various iris contaminations for large-scale iris identification and verification. To meet this goal and overcome the limitations mentioned above, we construct a multi-branch 2-ch CNN with a radial attention layer. This model is trained with online augmentation schemes to gain a robust iris classifier. Structural pruning is conducted for accurate and efficient iris matching. Finally, the encoding ability of the model is explored, and the performance of iris identification and verification is evaluated on three large-scale databases. The key novelties of our work can be concluded as in the following:
  • To allow the 2-ch CNN to apply in the large-scale iris identification scenario, we investigate the encoding ability of the 2-ch CNN in different layers and put forward a hybrid framework which takes advantage of the accurateness of 2-ch CNN and the efficiency of the encoding and matching.
  • A radial attention layer that can guide our model to focus on relevant iris regions along the radial direction is proposed, and branch level model pruning is realized by norm computing.
  • Three types of online augmentation schemes are designed to enhance the robustness of the model. The successful modeling of brightness jitter, iris image rotation, and radial extension occurring in real-time iris recognition can prevent the model from overfitting and allow the model to train on the small-scale training dataset.
  • A condensed 2-ch CNN with optimal architecture is obtained by pruning the model at the channel level as well as the branch level.
The remainder of this paper is organized as follows: Section 3 presents the details of the proposed iris recognition method. Section 4 illustrates the experiment settings and experimental results on three different databases. Section 5 extends the experimental results and makes additional comparisons. Section 6 summarizes the research and draws the conclusion.

3. Methods

The simplified iris identification and verification workflows of the proposed method are illustrated in Figure 1. As shown in Figure 1a, a full-sized CNN is first trained, and the subsequent pruning and finetuning procedures enable the CNN to obtain better performance and efficiency. Finally, a pair of preprocessed and normalized irises is fed into CNN to calculate the inference distance. Meanwhile, the iris identification workflow is illustrated in Figure 1b. Contrary to the “one-against-one” comparison strategy implemented in the verification scenario, the identification scenario needs to conduct more comparison operation in one identification epoch (i.e., the process of a sample paired with all databases). Consequently, we add an external encoding matching step, which can effectively alleviate the computational burden of the 2-ch CNN in iris identification scenario. Further, the details of each step are explained in the following subsections.

3.1. Preprocessing

We uniformly perform iris segmentation and normalization using an open-source tool named OSIRIS v4.1 [50]. Figure 2 depicts a sample for each preprocessing step. In the segmentation module of OSIRIS v4.1, the contours of the iris are detected by the Viterbi algorithm [51]. Subsequently, a modified Daugman’s rubber-sheet model is deployed to perform normalization. With this approach, the original iris image in any resolution is unwrapped into a size-invariant band. In this work, all the iris images are normalized to the size of 103 × 360. To eliminate the inference, we cut off the first 3 rows and the bottom 40 rows and then resize the image to a resolution of 30 × 360. After that, the contrast of the image is enhanced using histogram equalization [52]. In addition, a conditional horizontal cropping step is leveraged in the training stage of full-size network, which crops 20–164 columns and 240–306 columns and hence obtains a 30 × 150 normalized image. It contributes to reducing the time consumption and the probability of overfitting in the training stage.

3.2. Online Augmentation Method

We develop three well-designed augmentation schemes, namely brightness jitter, horizontal shift, and longitudinal scaling, to simulate the variant and contamination of the normalized iris. The explanation for each scheme is expressed as follows:
(1) Brightness jitter: As shown in Figure 3b, a random part of the region is set to be darker than the surrounding area by convoluting the image with a Gaussian kernel G 128 × 128 with a variance of 28 along both the x-axis and y-axis. The transformation simulates the non-uniform illumination in an iris acquisition environment. Additionally, previous studies proved the feasibility of achieving a performance improvement by covering a random region of the input images [53,54].
(2) Horizontal shift: To overcome the varying rotation degree in various subjects, we perform the translation on normalized iris using a random offset. Figure 3c depicts a sample of horizon shifts in the right direction. According to the definition of the normalized iris image, the horizon shift in normalized iris corresponds to the rotation in the original iris image.
(3) Longitudinal scaling: To better adapt to valid iris size changes caused by individual difference or pupil scaling, a longitudinal scaling augmentation is conducted on normalized iris. In this way, the normalized iris is scaled by a random factor F [ 0.75 , 1.25 ] along the longitudinal direction. If F 1 , then the first 30 rows are preserved as a valid region. Otherwise, a number of last certain rows is mirrored to compensate for the original image height. All the image rescaling operations are conducted by nearest-neighbor interpolation. Obviously, the longitudinal scaling refers to the radial scaling in an unsegmented iris. An example of this operation is illustrated in Figure 3d.
Furthermore, two input channels are randomly switched in the training phase. It is worth highlighting that these augmentation operations are performed online, which means the images are stochastically adjusted during the CNN optimization stage, while preserved in the testing phase. To that end, these functions are integrated into an augmentation layer at the front of the CNN architecture, which is performed in each mini-batch of the input data. Benefiting from the online augmentation, only a few classes are needed for model training and finetuning.

3.3. Modified 2-ch Deep Convolutional Neural Network

As mentioned above, the 2-ch CNN method presents great potential in iris recognition. However, the optimal CNN architecture, which can achieve a trade-off between performance and computational complexity, has not been fully explored. In contrast to empirically adjusting the hyperparameters such as the depth of the network and the number of kernels, we employ the pruning method to automatically search for a satisfactory CNN architecture. As shown in Figure 4a, a full-size CNN (Structure A) is first established and trained on the CASIA-V3-Interval training set, which extends a total of three branches with different depths between the first and the last convolutional layer. Hence, three different forward and backward propagation paths are established. The motivation of constructing this full-size architecture is to explore the possibility of capturing and integrating various depth’s feature representation. At the end of the network, a global average pooling (GAP) layer is utilized to replace the fully connected (FC) layer. The GAP operation sums up all elements in each feature map channel regardless of its tensor shape. With this approach, the network can handle any resolution samplings of a normalized iris image [55].
The loss function is the learning objective of a network and, hence, it should be carefully discussed and designed. Since the comparison of paired iris images is more inclined to distance measurement rather than classification, we take MSE with L 2 - norm as the loss function. Experiment results indicate that the MSE loss outperforms the cross-entropy (CE) loss and hinge-based loss [40]. The learning objective is the following:
min ω i = 1 M ( Y i T i ) 2 + δ 2 | | ω | | 2
where ω denote weights of the neural network, M = 256 stands for the mini-batch size, Y i is the i-th network output value, and T i { 0 , 2 } the corresponding target label (with 0 and 2 denoting a matching and a non-matching pair, respectively). The regularization coefficient δ is set to 0.001 in this study. The L 2 regularization is motivated to alleviate the overfitting of the model by constraining the summation of squares of all learnable parameters. With an appropriate penalty coefficient and sufficient optimizing iterations, the network can automatically filter out the inoperative weights and retains the informative convolution kernels so as to enhance the generalization performance.
In Figure 4a, it can be noticed that we place a radial attention layer at the head of each branch. Assume the tensor X H × W × C is feature map fed into the radial attention layer and the column vector W a H is the radial attention weight to be learned, where C, H, and W correspond to the channel, height, and width of the feature map, respectively. Then, the output of this layer Z H × W × C can be expressed as:
Z = repmat ( W a , [ 1   W   C ] ) X
where repmat ( W a , [ 1   W   C ] ) represents the duplication operation in the 2nd and 3rd dimensionalities for W and C times of W a . The operator corresponds to the Hadamard product. On the other hand, the gradient of W a in back propagation is computed as:
d L d W a = W C B d L d Z X
where L indicates the loss function of the network and B the mini-batch size in training phase. The radial attention layer weights different regions along the radial direction in the corresponding original iris image. It is proven that this layer can provide better recognition performance and help to prune the model.

3.4. Structural Model Pruning

The pruning operation permanently drops the less important weights from the trained model for computational efficiency. The structural model pruning schemes such as branch level pruning and channel level pruning can be easily implemented without extra optimization costs. In this study, the full-size CNN (Structure A) is first trained by 22,000 epochs on only 33 classes (874 genuine pairs + 1748 imposter pairs) in the CASIA-V3-Interval iris database. The model is optimized by the Adam optimizer [56], with an initial learning rate of 5 × 10 4 decreasing exponentially to 10 6 . It is noteworthy that all weights in full-size CNN are initialized by the Glorot method [57]. After training, we gather the L 1 norm of all radial layers of a full-size network and compare them with a fixed threshold T p r u n e = 10 3 . Then, the radial layer whose norm cannot reach the T p r u n e and its corresponding branches are discarded. Finally, a branch-selected network is obtained. As depicted in Figure 4b, only the deepest branch is reserved.
Additionally, the branch-pruned network can be further condensed by channel pruning. For this purpose, we calculate the accumulated L 1 norm of the output channel. By applying the aforementioned fixed threshold T p r u n e , the unimportant output channels together with their corresponding input channels are cut off permanently. Similarly, the corresponding weights in batch normalization (BN) layers are pruned. The application of L 1 norm can preserve more useful kernels, which can lead to a less performance loss. Figure 4c depicts the architecture of the final pruned network (Structure C). It is clear that the whole network, especially the last two convolution blocks of the network, has far fewer parameters compared to the network without channel pruning (Structure B). Actually, Structure C only contains 33,268 parameters, which is far smaller than all the CNN architectures employed in previous iris recognition studies to the best of our knowledge. Figure 5 depicts an example of channel pruning. Figure 5a shows the 16 × 32 channel map in the 2nd convolutional layer in branch-pruned 2-ch CNN (Structure B). In this layer, we have a total of 512 convolution kernels. As shown in Figure 5c, the channel map is reduced to a size of 11 × 22. The output channel pruning also leads to the input channel pruning in the next layer, which is presented in Figure 5b,d as horizontal black lines. The 3rd channel map is hence pruned from the size of 32 × 64 to 22 × 51.
After structural pruning, the model needs to retrain for maintaining accuracy. We retrain the model on the same training set of CASIA-V3-Interval by 500 epochs, with a minor learning rate of 10 6 . Finally, we finetune the model on different target database to adapt the domain of the target database. It is worth emphasizing that this finetuning procedure is done with the non-cropped normalized iris images in resolution of 30 × 360, which enables the model to capture more informative features and gain better performance. Specifically, the model is finetuned for 1200 epochs, with an initial learning rate of 5 × 10 5 decreasing exponentially to 2 × 10 7 .

3.5. Efficient Encoding for Large Scale Iris Recognition

A primary limitation of the 2-ch CNN method lies in its massive computational complexity. When a sample is sent in, this sample is paired with all the samples in the database. Each pair is fed into the 2-ch CNN to conduct forward propagation, which is more time-consuming than only calculating the encoding distance. Accordingly, we propose a hybrid method in the identification scenario. The max-pooling (MP) layer marked with a bold red rectangular frame in Figure 4c acts as an encoding layer. When performing iris registration for large scale iris identification scenario, which is time-sensitive, a pair of identical normalized iris images is fed into the 2-ch CNN and then the flattened feature map of the encoding layer is extracted and stored as a unique encoding for discarding the low confidence sample pairs. Meanwhile, the original normalized iris images are also stored for a more accurate classification with condensed 2-ch CNN. Unlike other deep learning-based feature encoding methods, the 2-ch CNN is not trained for the encoding purpose. However, we find that some of the middle layers in the condensed 2-ch CNN model have a powerful encoding ability, which means we can simultaneously obtain an iris image encoder with lower precision but higher efficiency and a pair-wise iris image classifier with reverse properties by training only one model.

4. Experiments and Results

4.1. Experimental Iris Databases

Three databases, namely CASIA-V1 [58], CASIA-V3-Interval [59], and CASIA-V4-Thousand [59], are adopted to conduct assessment and analysis in this study. The detailed description of these databases is reported as follows:
(1) CASIA-V1: A total of 756 iris images with a resolution of 320 × 280 were collected from 108 eyes. For each eye, seven images were captured in two sessions, where three samples were collected in the first session and four in the second session. The pupil regions of all iris images in this database were automatically detected and replaced with a circular region of uniform intensity to mask out the specular reflections from the NIR illuminators. Since the iris pupil is edited and the image quality is extremely clear, we conduct the ideal condition experiment using this database.
(2) CASIA-V3-Interval: In this database, a total of 2639 images with a resolution of 320 × 280 were collected from 249 subjects but 395 eyes. The number of gathered images was not fixed for each eye. Therefore, for the convenience of the experiment, only 233 classes (eyes) with an image number of seven or more are selected in this work. If the number of images in a class is greater than seven, we randomly choose seven images. In this study, the first 33 classes are adopted to train the original full-size network.
(3) CASIA-V4-Thousand: As the first publicly released iris database containing a thousand people, a total of 20,000 iris images were included in the CASIA-V4-Thousand database. Thus, ten pictures with a resolution of 640 × 480 were enrolled for each person’s left and right eye. The dominating variations in the database are eyeglasses and strong specular reflections, which pose a more significant challenge to the iris recognition algorithm. We select the left eye of the first 648 subjects to carry out related experiments. In addition, seven pictures are randomly selected from each class.
For convenience, here we use CASIA-V3 and CASIA-V4 database to represent the CASIA-V3-Interval and CASIA-V4-Thousand database. As shown in Figure 6, a typical sample is randomly picked from each database. It is worth noting that only the training data in the CASIA-V3 dataset are employed to train the original CNN from scratch, and the training data in other databases are used for finetuning.

4.2. Experimental Results for Iris Identification and Verification

To fully assess the accurateness and effectiveness of our algorithm, we test the model on three publicly available databases under different configurations. The identification evaluation criterion is designated to the recognition accuracy, which is defined as the ratio of the number of correct recognitions to the total number of recognitions. In an identification epoch, the sample paired with all samples in the database, and if the sample pairs with the lowest score belong to the same class, then it is called a correct recognition. Since each class in all databases uniformly has seven images, in this study, all the reported identification results are the mean of the 7-fold cross-validation. On the other hand, we select the equal error rate (EER) as the evaluation criterion of verification scenario. The evaluation program pairs all possible combinations of the sample in the database and gained the false acceptance rate (FAR) and false rejection rate (FRR) at different thresholds. With a particular threshold, the FAR is equal to FRR and also EER.
Table 1 illustrates ten comparative experimental results evaluated in the identification and verification scenario. In the identification scenario, two typical conditions, including one picture registered and the six pictures registered, are considered. At the same time, the EER, along with the FRR at FAR = 0.1% and FRR at FAR = 0.01%, is regarded as the important performance indicator for verification assessment. As described above, the proposed network is first trained on 33 classes in the CASIA-V3 database from scratch, and the model is then retrained on the same 33-class training set after pruning operation. The retrained model yields an excellent EER of 0.76%, and outstanding recognition accuracy of 98.95% is fulfilled with only one picture registered in each class. Further, the full receiver operating characteristic (ROC) curve is plotted in Figure 7. For the CASIA-V1 database, the result achieved by the transferred model is reasonably good because of its high image quality. Moreover, if the model is finetuned with 20 classes to adapt its domain, the performance cannot gain much improvement. The CASIA-V4-Thousand database is widely recognized as one of the most challenging iris databases. Therefore, we comprehensively evaluate our proposed algorithm on it. As can be seen from experiments 4–10, various classes are employed for finetuning, and more than 600 classes (approximately 10M pairs) serve as the testing set. Experiment 4 indicates that the transferred model can reach an accuracy of 98.21% in the identification scenario and an EER of 3.54% in the verification scenario. However, the transferred model may not be suitable for the few pictures registered condition or high security required application scenarios. Sequentially, we finetuned the model with 5–30 classes from the target CASIA-V4 database to fit the target domain feature distribution. Interestingly, it can be observed that by exploiting only five classes (138 genuine pairs and 68 imposter pairs) for finetuning, we can gain a significant performance improvement. With the increase of the amount of tuning data, the performance of the model is gradually improved. The results indicate that 30 classes (815 genuine pairs and 1630 imposter pairs) can positively meet the finetuning demands. The adequately finetuned model is able to provide an accuracy of 97.92% to 99.77% in the identification scenario. An extremely low EER of 1.19% can be reached while the FRR is 2.16% and 3.31% under the FAR = 0.1% and FAR = 0.01% criterion. The extraordinary performance of the model ensures it can be applied to iris recognition scenes with high accuracy and security requirements.

4.3. Ablation Study

We perform extensive ablation experiments on the CASIA-V3-Interval database to comparatively demonstrate the effectiveness of each technique we employed in this study. From the results illustrated in Table 2, it can be observed that our well-designed model (i.e., the model in experiment 2) defeats the widely used baseline model codenamed as Resnet-18 with an EER improvement of 0.29%. The L 2 regularization term can not only help to prune the network architecture but also contribute to enhancing the model. According to experiment 4, it can be seen that if the online augmentation layer is removed, the performance of the model will suffer greatly. This can demonstrate the effectiveness and necessity of the online augmentation method for small-scale datasets. The comparison between experiment 5 and experiment 2 shows the superiority of the MSE loss over the CE loss. Experiment 6 indicates that the radial attention layer can reduce EER at the cost of adding very few learnable parameters. From experiments 6, 7, and 8, we can conclude that if we only prune the model without the retraining procedure, the accuracy of the model will decrease. When performing the finetuning, the performance of the pruned model will exceed that of the unpruned model. In experiment 9, we specially build a single branch network with the same structure as the pruned network and train from scratch. Compared with the result in experiment 8, the necessity of pruning is obvious. Finally, we add the reshape operation, which means the model is trained on a cropped normalized iris image dataset with a resolution of 30 × 150 but finetuned by an uncropped dataset with a resolution of 30 × 360. In this way, despite some interferences mixed in, more iris textures can be captured, and hence the best performance is gained.
To better examine the effectiveness of the proposed online augmentation scheme, an ablation study is done on 150 testing classes and five finetuning classes in CASIA-V4. As shown in Table 3, all three types of augmentation schemes can effectively contribute to improving the performance of the model. Compared to the non-augmented situation, scheme BJ., HS., and LS. alone can offer a 0.16%, 0.30%, and 0.09% reduction in EER, respectively. Moreover, the combination of these three online augmentation methods can achieve a better result. When all three augmentation schemes are utilized, the best performance is yielded.

4.4. Encoding Ability Research

Now we explore the encoding ability of the 2-ch CNN model. For convenience, we define the discarding accuracy D-Accuracy as the following equation:
D - A c c u r a c y = i N d j N R 1 { S i j < T h r } N d × N R
where Thr is a threshold, which can be specified manually or by setting the discarding ratio. N d and N R are the number of identification epochs and the number of registered iris images in each class, respectively. S i j is introduced as the matching score of pairing the original sample with its intra-class j-th sample in the i-th identification epoch. The descriptor 1 { } stands for the indicator function. In this experiment, the discarding accuracy is defined to evaluate the encoding performance of the model. We traverse the encoding capabilities of all layers on CASIA-V3 database. As depicted in Table 4, the discarding accuracy of each layer and its corresponding feature length and matching time in one identification epoch is analyzed in detail. It can be observed that layer 17 reaches the best discarding accuracy and the smallest variance simultaneously. Nevertheless, it is not the best choice for encoding matching due to its corresponding lengthy feature vectors and high time consumption. There are two feasible choices for generating encodings. First is layer 19: compared to the 17th layer, the 19th layer has more than doubled the time-consuming reduction with a loss of discarding accuracy of approximately 1%. The second choice is layer 22, which can achieve 90.23% discarding accuracy even if the feature length is only 960. In this work, we choose the 19th layer as the encoding layer.
Figure 8a demonstrates the discarding accuracy of the proposed modified 2-ch CNN model in different discarding ratios ranging from 0% to 90% and a different number of registered pictures ranging from 1 to 6. We can conclude that a discarding accuracy over 95% can be achieved with a 90% discarding ratio regardless of the number of registered images per class. Such a satisfying result demonstrates the feasibility of taking the 2-ch CNN as a feature extractor. Besides, Figure 8b plots the curve of identification accuracy varying with the discarding ratio under a different number of registered pictures. It reveals the effectiveness of the discarding process for the identification scenario. Meanwhile, the impact of different number of registered pictures on identification accuracy is also well demonstrated. We can see that when only one picture is registered per class, the identification accuracy is mainly restricted by the discarding accuracy. However, when more than two pictures are registered in each class, the identification accuracy is almost independent of the discarding accuracy. Even in some cases, the screening process can contribute a slight improvement to identification accuracy. This experimental result indicates that only three pictures of each eye are needed to ensure an identification accuracy of more than 99%.

5. Discussion

5.1. Weight Visualization

We further visualized some examples of the 2D convolution kernel at the microscopic level learned from each convolution layer. It can be seen in Figure 9 that the convolution kernels in the first four layers learn somewhat chaotic kernel maps. Using these kernels, the model can adapt to various inputs with stochastic perturbations and thereby gain the feature in abstract and robustness. On the other hand, the kernel map of the last two layers seems more specific and regular. Such a phenomenon may be because the feature maps processed from the previous layer have been regularly reshaped, which is insensitive to input disturbances.
As aforementioned, the radial attention layer acts as a branch selection gate as well as an iris region weight function. Figure 10 illustrates the weights that the radial attention layers learned. The smaller value of the X-axis stands for the radial part closer to the pupil. It can be seen from the weight distribution of both radial attention layers that the closer to the pupil, the larger the weight.

5.2. Time Consumption Experiments

In order to further ascertain the effectiveness of the encoding matching process, we calculate the time consumption of the algorithm under different parameter configurations and different devices, which are shown in Figure 11. Traditional iris recognition algorithms are usually deployed on the central processing unit (CPU) [60], and they are hard to parallelize. By contrast, CNN can easily be parallelized and deployed on the graphics processing unit (GPU) by using the mainstream deep learning framework [61]. To comprehensively evaluate the proposed algorithm’s execution efficiency in different application situations, GPU (NVIDIA GTX-1080) and CPU (i7-8700K, 3.7GHz) are considered in this experiment. This experiment is also conducted on CASIA-V3-Interval database. An interesting phenomenon can be observed from Figure 11a,b that with the increase of the number of registered pictures in each class, the computational time consumption of the encoding matching process grows linearly on GPU and nonlinearly on CPU. This may be caused by different caching mechanisms in the CPU. The elapsed time in identification procedure represented in Figure 11c,d can correspond to the identification accuracy in Figure 8b under the same conditions. For GPU devices, it takes only 24 ms (with 90% discarding ratio and one picture registered per class) to 677 ms (with no discarding and six pictures registered per class) to recognize 1400 images. In contrast, for CPU devices, it takes 95 to 5816ms under the same conditions. It should be specified that since the 2-ch CNN network is affected by the stack order of the input channels, we conducted double forward propagation with different channel orders. About 0.25 ms is needed for our model to process one pair of images, which is nearly three times less time-consuming than the DeepIris model mentioned in the literature [47]. Based on this premise, we can choose the corresponding device and discarding ratio according to the actual demand to balance the efficiency and expenses.

5.3. Interpretability Analysis

Despite the 2-ch CNN method has proven its effectiveness in image patch comparison, no efforts have been made to understand how the model obtained its score. In this study, we employed the gradient of the model regression output with respect to the last convolutional layer in a network to find which parts of the iris dominate the output score. The Grad-Cam approach [62] is deployed for this purpose, and the analysis results are shown in Figure 12. The red part of the image in both two channels can be regarded as the region with the most distinctive iris texture. Figure 12a,d demonstrate two typical cases of ideal recognition. For the iris pair identified as intra-class, only a few areas are marked in red, while for the iris pair between classes, most regions are considered inconsistent. In addition, two situations in which the confidence of the output score is relatively low are depicted in Figure 12b,e. We can see that contaminations, including eyelids and eyelashes, are enrolled in these normalized iris images, which affects the judgment of the model. At last, we provide two examples of false detection in Figure 12c,f. The false-negative in Figure 12c is caused by the uneven width of the segmented iris image, which may be the result of pupil dilation generated with dramatic light changes or other interferences. Correspondingly, a false-positive is described in Figure 12f. We observe that this iris pair is too similar and that it can be easily mistakenly identified even by human visual inspection. On considering all these images, it can be found that the different parts marked by our model are seldomly located in the polluted area, which is adequate proof that the model has strong robustness to the stochastic contaminations in the iris image.

5.4. Comparison of Iris Recognition Results

Table 5 compares our proposed method with other methods assessed on CASIA databases in recent years. For the classic iris recognition algorithm, the comparison is made with IrisCode. Othman et al. [50] constructed the OSIRIS framework and presented the classic iris recognition chain, which reproduced the IrisCode algorithm proposed by Daugman [29]. The IrisCode is a handcrafted feature, but it can be applied to the new database without training data. Our proposed method fulfills an EER of 3.54% in the cross-database case on the CASIA-V4 dataset, which is approximately the same as the IrisCode method. However, when a few samples are utilized for finetuning the model, a significant performance advantage is shown by our method. For the deep learning-based method, the encoding ability of the off-the-shelf CNN features are explored in No. 2, 3, 7 entries of Table 5. Moreover, in examples from the literature [45,63,64,65], new CNN models were established, which were trained from scratch, and then their classification ability investigated. It can be seen that all the mentioned methods need a great number of samples for training, which is impossible for practical application scenarios. Moreover, the CNN models utilized in previous studies also have far more parameters, while the performance is evidently lower than ours. The 2-ch CNN-based method was studied in more recent research [49] and [66]. Proença et al. [49] proposed a segment-less CNN model based on VGG-19, and the model was trained by 45,000 genuine pairs and 1 million imposter pairs on CASIA-V4. As a result, an EER of 3% was acquired on testing phase, which was much lower than ours. Chen et al. [66] proposed a novel loss named Tight Center and assessed it with three types of classic architectures on the CASIA-V4 database by adopting a cross-database scheme. As a result, the best result with an EER of 2.36% and an accuracy of 99.58% was reported, which was slightly better than our cross-database results. However, their method was trained on 50,632 images on the ND-IRIS-0405 iris database, and then the model was evaluated on 38,573 genuine pairs and 107,589 imposter pairs on the CASIA-V4 database. In contrast, our method is trained on only 231 images in the CASIA-V3 iris database and test on more than 10M pairs in the CASIA-V4 database. Moreover, we compute the number of parameters and computational cost of our model and other compared models. As illustrated in Table 5, our model has a total number of 33K parameters and floating-point operations (FLOPs) of 49.1M, which are lower than most of the previous models. The model employed in [64] has the least computational cost, but its identification accuracy is far lower than ours. Overall, our proposed condensed 2-ch CNN method achieves state-of-the-art performance on three publicly available databases with few sample tuning and much fewer model parameters.

6. Conclusions

This work presents a new framework for large-scale iris verification and identification using 2-ch CNN. Four key innovations, including the hybrid framework for large-scale iris identification and verification, radial attention layer for weighting different iris regions, online augmentation schemes for enhancing the robustness, and structural pruning for alleviating computational burden, are introduced in 2-ch CNN to improve the performance. The proposed method is evaluated on three publicly available databases. The experimental results indicate that our method has outstanding efficiency and performance over previous studies utilizing deep learning-based methods and handcrafted feature-based methods. Moreover, the satisfying results achieved on the CASIA-V4-Thousand database indicate that the proposed method can be applied in challenging iris recognition situations. This work also investigates the encoding ability of the 2-ch CNN and finds that some middle layers have excellent encoding ability. This enables the 2-ch CNN to be applied to large-scale iris identification. Since all three types of online augmentation schemes carefully designed in this study are proven to be beneficial for model performance, we will continue to develop these schemes and consider more contaminations in iris images, such as eyelids and eyelashes. Additionally, the multi-modal identification approach, which combines iris recognition with other biometrics approaches such as face recognition and palmprint recognition, is suggested for future work.

Author Contributions

Conceptualization, G.L. and W.Z.; Data curation, H.X.; Formal analysis, W.L.; Funding acquisition, W.Z. and L.T.; Investigation, G.L. and Y.L.; Methodology, G.L.; Project administration, W.Z.; Resources, L.T. and W.L.; Software, G.L.; Supervision, W.Z.; Validation, G.L., Y.L. and H.X.; Visualization, G.L., W.L. and Y.L.; Writing—original draft, G.L.; Writing—review and editing, W.Z. and L.T. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported in part by the Key Program of Natural Science Foundation of Shandong Province (No. ZR2020LZH009), the Research Funds of Science and Technology Innovation Committee of Shenzhen Municipality (No. JCYJ20180305164357463), the National Natural Science Foundation of China (No. 11474185), and the Key R&D Project of Shandong Province (No. 2016GGX101028).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available in http://biometrics.idealtest.org/ (accessed on 21 October 2019).

Acknowledgments

We gratefully acknowledge the support from the above funds.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Bowyer, K.W.; Hollingsworth, K.; Flynn, P.J. Image understanding for iris biometrics: A survey. Comput. Vis. Image Underst. 2008, 110, 281–307. [Google Scholar] [CrossRef]
  2. Nguyen, K.; Fookes, C.; Jillela, R.; Sridharan, S.; Ross, A. Long range iris recognition: A survey. Pattern Recognit. 2017, 72, 123–143. [Google Scholar] [CrossRef]
  3. Sheela, S.; Vijaya, P. Iris recognition methods-survey. Int. J. Comput. Appl. Technol. 2010, 3, 19–25. [Google Scholar] [CrossRef]
  4. Winston, J.J.; Hemanth, D.J. A comprehensive review on iris image-based biometric system. Soft Comput. 2019, 23, 9361–9384. [Google Scholar] [CrossRef]
  5. Bonnen, K.; Klare, B.F.; Jain, A.K. Component-based representation in automated face recognition. IEEE Trans. Inf. Forensics Secur. 2012, 8, 239–253. [Google Scholar] [CrossRef] [Green Version]
  6. Meraoumia, A.; Chitroub, S.; Bouridane, A. Palmprint and Finger-Knuckle-Print for efficient person recognition based on Log-Gabor filter response. Analog Integr. Circuits Signal Process. 2011, 69, 17–27. [Google Scholar] [CrossRef]
  7. Jain, A.K.; Arora, S.S.; Cao, K.; Best-Rowden, L.; Bhatnagar, A. Fingerprint recognition of young children. IEEE Trans. Inf. Forensics Secur. 2016, 12, 1501–1514. [Google Scholar] [CrossRef]
  8. Alqahtani, A. Evaluation of the reliability of iris recognition biometric authentication systems. In Proceedings of the 2016 International Conference on Computational Science and Computational Intelligence (CSCI’16), Las Vegas, NV, USA, 15–17 December 2016; pp. 781–785. [Google Scholar]
  9. Benalcazar, D.P.; Zambrano, J.E.; Bastias, D.; Perez, C.A.; Bowyer, K.W. A 3D Iris Scanner from a Single Image Using Convolutional Neural Networks. IEEE Access 2020, 8, 98584–98599. [Google Scholar] [CrossRef]
  10. Boyd, A.; Yadav, S.; Swearingen, T.; Kuehlkamp, A.; Trokielewicz, M.; Benjamin, E.; Maciejewicz, P.; Chute, D.; Ross, A.; Flynn, P. Post-Mortem Iris Recognition—A Survey and Assessment of the State of the Art. IEEE Access 2020, 8, 136570–136593. [Google Scholar] [CrossRef]
  11. Vyas, R.; Kanumuri, T.; Sheoran, G.; Dubey, P. Smartphone based iris recognition through optimized textural representation. Multimed. Tools Appl. 2020, 79, 14127–14146. [Google Scholar] [CrossRef]
  12. LeCun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef] [Green Version]
  13. Zhao, Z.-Q.; Zheng, P.; Xu, S.-T.; Wu, X. Object detection with deep learning: A review. IEEE Trans. Neural Netw. Learn Syst. 2019, 30, 3212–3232. [Google Scholar] [CrossRef] [Green Version]
  14. Guo, G.; Zhang, N. A survey on deep learning based face recognition. Comput. Vis. Image Underst. 2019, 189, 102805. [Google Scholar] [CrossRef]
  15. Voulodimos, A.; Doulamis, N.; Doulamis, A.; Protopapadakis, E. Deep learning for computer vision: A brief review. Comput. Intell. Neurosci. 2018, 2018, 13. [Google Scholar] [CrossRef] [PubMed]
  16. Guo, Y.; Liu, Y.; Oerlemans, A.; Lao, S.; Wu, S.; Lew, M.S. Deep learning for visual understanding: A review. Neurocomputing 2016, 187, 27–48. [Google Scholar] [CrossRef]
  17. Goodfellow, I.; Bengio, Y.; Courville, A. Deep Learning; MIT Press: Cambridge, MA, USA, 2016. [Google Scholar]
  18. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. In Proceedings of the Twenty-sixth Annual Conference on Neural Information Processing Systems (NIPS), Lake Tahoe, NV, USA, 3–8 December 2012. [Google Scholar]
  19. Bazrafkan, S.; Thavalengal, S.; Corcoran, P. An end to end deep neural network for iris segmentation in unconstrained scenarios. Neural Netw. 2018, 106, 79–95. [Google Scholar] [CrossRef] [Green Version]
  20. Arsalan, M.; Naqvi, R.A.; Kim, D.S.; Nguyen, P.H.; Owais, M.; Park, K.R. IrisDenseNet: Robust iris segmentation using densely connected fully convolutional networks in the images by visible light and near-infrared light camera sensors. Sensors 2018, 18, 1501. [Google Scholar] [CrossRef] [Green Version]
  21. Jayanthi, J.; Lydia, E.L.; Krishnaraj, N.; Jayasankar, T.; Babu, R.L.; Suji, R.A. An effective deep learning features based integrated framework for iris detection and recognition. J. Ambient Intell. Humaniz. Comput. 2020, 12, 3271–3281. [Google Scholar] [CrossRef]
  22. Hamd, M.H.; Ahmed, S.K. Biometric system design for iris recognition using intelligent algorithms. Inter. J. Educ. Mod. Comp. Sci. 2018, 10, 9. [Google Scholar] [CrossRef]
  23. Park, K.; Song, M.; Kim, S.Y. The design of a single-bit CMOS image sensor for iris recognition applications. Sensors 2018, 18, 669. [Google Scholar] [CrossRef] [Green Version]
  24. Agarwal, R.; Jalal, A.S. Presentation attack detection system for fake Iris: A review. Multimed. Tools. Appl. 2021, 80, 15193–15214. [Google Scholar] [CrossRef]
  25. Nguyen, D.T.; Pham, T.D.; Lee, Y.W.; Park, K.R. Deep learning-based enhanced presentation attack detection for iris recognition by combining features from local and global regions based on NIR camera sensor. Sensors 2018, 18, 2601. [Google Scholar] [CrossRef] [Green Version]
  26. Wang, K.; Kumar, A. Toward more accurate iris recognition using dilated residual features. IEEE Trans. Inf. Forensics Secur. 2019, 14, 3233–3245. [Google Scholar] [CrossRef]
  27. Daugman, J.G. High confidence visual recognition of persons by a test of statistical independence. IEEE Trans. Pattern Anal. Mach. Intell. 1993, 15, 1148–1161. [Google Scholar] [CrossRef] [Green Version]
  28. Flom, L.; Safir, A. Iris Recognition System. Google Patents US4641349A, 3 February 1987. [Google Scholar]
  29. Daugman, J. New methods in iris recognition. IEEE Trans. Syst. Man Cybern. Syst. Cybern Part B (Cybern) 2007, 37, 1167–1175. [Google Scholar] [CrossRef] [Green Version]
  30. Daugman, J. How iris recognition works. In The Essential Guide to Image Processing; Elsevier: Amsterdam, The Netherlands, 2009; pp. 715–739. [Google Scholar]
  31. Barpanda, S.S.; Majhi, B.; Sa, P.K.; Sangaiah, A.K.; Bakshi, S. Iris feature extraction through wavelet mel-frequency cepstrum coefficients. Opt. Laser Technol. 2019, 110, 13–23. [Google Scholar] [CrossRef]
  32. Nalla, P.R.; Kumar, A. Toward more accurate iris recognition using cross-spectral matching. IEEE Trans. Image Process. 2016, 26, 208–221. [Google Scholar] [CrossRef] [PubMed]
  33. Yao, P.; Li, J.; Ye, X.Y.; Zhuang, Z.Q.; Li, B. Iris recognition algorithm using modified Log-Gabor filters. In Proceedings of the 18th International Conference on Pattern Recognition (ICPR’06), Hong Kong, China, 20–24 August 2006; pp. 461–464. [Google Scholar]
  34. Alonso-Fernandez, F.; Tome-Gonzalez, P.; Ruiz-Albacete, V.; Ortega-Garcia, J. Iris recognition based on SIFT features. In Proceedings of the First IEEE International Conference on Biometrics, Identity and Security (BIdS 2009), Tampa, FL, USA, 22–23 September 2009; p. 8. [Google Scholar]
  35. Bakshi, S.; Das, S.; Mehrotra, H.; Sa, P.K. Score level fusion of SIFT and SURF for iris. In Proceedings of the 2012 International Conference on Devices, Circuits and Systems (ICDCS), Coimbatore, India, 15–16 March 2012; pp. 527–531. [Google Scholar]
  36. Zhu, R.; Yang, J.; Wu, R. Iris recognition based on local feature point matching. In Proceedings of the 2006 International Symposium on Communications and Information Technologies, Bangkok, Thailand, 18–20 October 2006; pp. 451–454. [Google Scholar]
  37. Juneja, K.; Rana, C. Compression-Robust and Fuzzy-Based Feature-Fusion Model for Optimizing the Iris Recognition. Wirel. Pers. Commun. 2021, 116, 267–300. [Google Scholar] [CrossRef]
  38. Santos, G.; Hoyle, E. A fusion approach to unconstrained iris recognition. Pattern Recognit. Lett. 2012, 33, 984–990. [Google Scholar] [CrossRef]
  39. Tajbakhsh, N.; Araabi, B.N.; Soltanianzadeh, H. Feature fusion as a practical solution toward noncooperative iris recognition. In Proceedings of the 11th International Conference on Information Fusion, Cologne, Germany, 30 June–3 July 2008; pp. 1–7. [Google Scholar]
  40. Chen, J.; Shen, F.; Chen, D.Z.; Flynn, P.J. Iris recognition based on human-interpretable features. IEEE Trans. Inf. Forensics Secur. 2016, 11, 1476–1485. [Google Scholar] [CrossRef]
  41. Shen, F. A Visually Interpretable Iris Recognition System with Crypt Features. Ph.D. Thesis, University of Notre Dame, Notre Dame, IN, USA, 2014. [Google Scholar]
  42. Gangwar, A.; Joshi, A. DeepIrisNet: Deep iris representation with applications in iris recognition and cross-sensor iris recognition. In Proceedings of the 2016 IEEE International Conference on Image Processing (ICIP), Phoenix, AZ, USA, 25–28 September 2016; pp. 2301–2305. [Google Scholar]
  43. Nguyen, K.; Fookes, C.; Ross, A.; Sridharan, S. Iris recognition with off-the-shelf CNN features: A deep learning perspective. IEEE Access 2017, 6, 18848–18855. [Google Scholar] [CrossRef]
  44. Raja, K.B.; Raghavendra, R.; Venkatesh, S.; Busch, C. Multi-patch deep sparse histograms for iris recognition in visible spectrum using collaborative subspace for robust verification. Pattern Recognit. Lett. 2017, 91, 27–36. [Google Scholar] [CrossRef]
  45. Liu, M.; Zhou, Z.; Shang, P.; Xu, D. Fuzzified image enhancement for deep learning in iris recognition. IEEE Trans. Fuzzy Syst. 2019, 28, 92–99. [Google Scholar] [CrossRef]
  46. Zagoruyko, S.; Komodakis, N. Learning to compare image patches via convolutional neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 4353–4361. [Google Scholar]
  47. Liu, N.; Zhang, M.; Li, H.; Sun, Z.; Tan, T. DeepIris: Learning pairwise filter bank for heterogeneous iris verification. Pattern Recognit. Lett. 2016, 82, 154–161. [Google Scholar] [CrossRef]
  48. Špetlík, R.; Razumenić, I. Iris verification with convolutional neural network and unit-circle layer. In Proceedings of the 41th German Conference on Pattern Recognition, Dortmund, Germany, 10–13 September 2019. [Google Scholar]
  49. Proença, H.; Neves, J.C. Segmentation-less and non-holistic deep-learning frameworks for iris recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Long Beach, CA, USA, 16–17 June 2019. [Google Scholar]
  50. Othman, N.; Dorizzi, B.; Garcia-Salicetti, S. OSIRIS: An open source iris recognition software. Pattern Recognit. Lett. 2016, 82, 124–131. [Google Scholar] [CrossRef]
  51. Sutra, G.; Garcia-Salicetti, S.; Dorizzi, B. The Viterbi algorithm at different resolutions for enhanced iris segmentation. In Proceedings of the 2012 5th IAPR International Conference on Biometrics (ICB), New Delhi, India, 29 March–1April 2012. [Google Scholar]
  52. Gonzalez, R.C.; Woods, R.E.; Eddins, S.L. Digital Image Processing Using MATLAB; Pearson Education India: Bengaluru, India, 2004. [Google Scholar]
  53. Zhong, Z.; Zheng, L.; Kang, G.; Li, S.; Yang, Y. Random erasing data augmentation. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020. [Google Scholar]
  54. DeVries, T.; Taylor, G.W. Improved regularization of convolutional neural networks with cutout. arXiv 2017, arXiv:1708.04552. [Google Scholar]
  55. Lin, M.; Chen, Q.; Yan, S. Network in network. arXiv 2013, arXiv:1312.4400. [Google Scholar]
  56. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  57. Glorot, X.; Bengio, Y. Understanding the difficulty of training deep feedforward neural networks. In Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics, Chia Laguna Resort, Sardinia, Italy, 13–15 May 2010. [Google Scholar]
  58. CASIA-IrisV1. Available online: http://biometrics.idealtest.org/ (accessed on 21 October 2019).
  59. Database, C.I.I. Available online: http://biometrics.idealtest.org/ (accessed on 21 October 2019).
  60. Rakvic, R.N.; Ulis, B.J.; Broussard, R.P.; Ives, R.W.; Steiner, N. Parallelizing iris recognition. IEEE Trans. Inf. Forensics Secur. 2009, 4, 812–823. [Google Scholar] [CrossRef]
  61. Subramanian, V. Deep Learning with PyTorch: A Practical Approach to Building Neural Network Models Using PyTorch; Packt Publishing Ltd.: Birmingham, UK, 2018. [Google Scholar]
  62. Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-cam: Visual explanations from deep networks via gradient-based localization. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017. [Google Scholar]
  63. Wang, Z.; Li, C.; Shao, H.; Sun, J. Eye recognition with mixed convolutional and residual network (MiCoRe-Net). IEEE Access 2018, 6, 17905–17912. [Google Scholar] [CrossRef]
  64. Tobji, R.; Di, W.; Ayoub, N. FMnet: Iris segmentation and recognition by using fully and multi-scale CNN for biometric security. Appl. Sci. 2019, 9, 2042. [Google Scholar] [CrossRef] [Green Version]
  65. Lee, Y.W.; Kim, K.W.; Hoang, T.M.; Arsalan, M.; Park, K.R. Deep residual CNN-based ocular recognition based on rough pupil detection in the images by NIR camera sensor. Sensors 2019, 19, 842. [Google Scholar] [CrossRef] [Green Version]
  66. Chen, Y.; Wu, C.; Wang, Y. T-Center: A Novel Feature Extraction Approach towards Large-Scale Iris Recognition. IEEE Access 2020, 8, 32365–32375. [Google Scholar] [CrossRef]
  67. Alaslani, M.G.; Elrefaei, L.A. Convolutional neural network based feature extraction for iris recognition. J. Comput. Sci. Inf. Technol. 2018, 10, 65–78. [Google Scholar] [CrossRef] [Green Version]
  68. Boyd, A.; Czajka, A.; Bowyer, K. Deep learning-based feature extraction in iris recognition: Use existing models, fine-tune or train from scratch? In Proceedings of the 2019 IEEE 10th International Conference on Biometrics Theory, Applications and Systems (BTAS), Tampa, FL, USA, 23–26 September 2019. [Google Scholar]
Figure 1. The whole architecture of our iris recognition algorithm. (a,b) demonstrate the verification and identification workflow, respectively.
Figure 1. The whole architecture of our iris recognition algorithm. (a,b) demonstrate the verification and identification workflow, respectively.
Sensors 21 03721 g001
Figure 2. The preprocessing stage including location and segmentation (a), normalization (b), longitudinal cropping and image enhancement (c), and optional horizontal cropping step (d). The region between inner and outer green boundary in (a) is the segmented iris.
Figure 2. The preprocessing stage including location and segmentation (a), normalization (b), longitudinal cropping and image enhancement (c), and optional horizontal cropping step (d). The region between inner and outer green boundary in (a) is the segmented iris.
Sensors 21 03721 g002
Figure 3. The examples of the output for each different online augmentation layer. (a) is a normalized and enhanced iris image randomly picked from CASIA-V3-Interval database. (bd) are the example of brightness jitter, horizontal shift, and longitudinal scaling operation, respectively. The red rectangular window shown in (d) marks the mirrored part of the iris.
Figure 3. The examples of the output for each different online augmentation layer. (a) is a normalized and enhanced iris image randomly picked from CASIA-V3-Interval database. (bd) are the example of brightness jitter, horizontal shift, and longitudinal scaling operation, respectively. The red rectangular window shown in (d) marks the mirrored part of the iris.
Sensors 21 03721 g003
Figure 4. The architecture of the proposed convolutional neural network. (a) presents the full-size 2-ch CNN (Structure A). (b,c) illustrate the branch-pruned and channel-pruned CNN (Structure B and C), respectively. For convenience, a convolutional layer, batch normalization layer, and a ReLu activation layer are integrated into a convolution block (conv) in order. For a specific convolution block, the kernel size and the number of output channels are marked in the upper left and right corners of the box, respectively. All the convolution operations are with the stride of 1 in each direction. Moreover, for the max-pooling layer (maxpool), the pooling region and pooling stride are also marked in the upper left and right corners of the box, respectively.
Figure 4. The architecture of the proposed convolutional neural network. (a) presents the full-size 2-ch CNN (Structure A). (b,c) illustrate the branch-pruned and channel-pruned CNN (Structure B and C), respectively. For convenience, a convolutional layer, batch normalization layer, and a ReLu activation layer are integrated into a convolution block (conv) in order. For a specific convolution block, the kernel size and the number of output channels are marked in the upper left and right corners of the box, respectively. All the convolution operations are with the stride of 1 in each direction. Moreover, for the max-pooling layer (maxpool), the pooling region and pooling stride are also marked in the upper left and right corners of the box, respectively.
Sensors 21 03721 g004
Figure 5. The demonstration of channel level sparsity. Each entry in the matrix represents the L 1 norm of the kernel. (a,b) illustrate the 2nd and 3rd convolutional layer channel map. The brighter element represents the more important kernel. (c,d) are the corresponding pruned channel map. The white regions are reserved while the black regions are discarded.
Figure 5. The demonstration of channel level sparsity. Each entry in the matrix represents the L 1 norm of the kernel. (a,b) illustrate the 2nd and 3rd convolutional layer channel map. The brighter element represents the more important kernel. (c,d) are the corresponding pruned channel map. The white regions are reserved while the black regions are discarded.
Sensors 21 03721 g005
Figure 6. The sample iris images randomly picked from three databases.
Figure 6. The sample iris images randomly picked from three databases.
Sensors 21 03721 g006
Figure 7. The ROC curve of the proposed algorithm on database CASIA-V3-Interval.
Figure 7. The ROC curve of the proposed algorithm on database CASIA-V3-Interval.
Sensors 21 03721 g007
Figure 8. (a) The discarding accuracies under different discarding ratio and different number of registered pictures. (b) The identification accuracies under different discarding ratio and different number of registered pictures.
Figure 8. (a) The discarding accuracies under different discarding ratio and different number of registered pictures. (b) The identification accuracies under different discarding ratio and different number of registered pictures.
Sensors 21 03721 g008
Figure 9. The visualization of the convolution kernels randomly picked from each convolution layer. Each kernel in size of 3 × 3 or 5 × 5 is resized to a higher resolution by cubic interpolation.
Figure 9. The visualization of the convolution kernels randomly picked from each convolution layer. Each kernel in size of 3 × 3 or 5 × 5 is resized to a higher resolution by cubic interpolation.
Sensors 21 03721 g009
Figure 10. The visualization of the radial attention layer in proposed CNN architecture. (a,b) correspond to the attention weight for radial attention layer 1 and 2, respectively.
Figure 10. The visualization of the radial attention layer in proposed CNN architecture. (a,b) correspond to the attention weight for radial attention layer 1 and 2, respectively.
Sensors 21 03721 g010
Figure 11. The summary of the time consumption in different conditions. (a,b) correspond to the time consumption of screening procedure achieved by GPU and CPU, respectively. (c,d) the time consumption of identification procedure achieved by GPU and CPU, respectively.
Figure 11. The summary of the time consumption in different conditions. (a,b) correspond to the time consumption of screening procedure achieved by GPU and CPU, respectively. (c,d) the time consumption of identification procedure achieved by GPU and CPU, respectively.
Sensors 21 03721 g011
Figure 12. (af) are the heat maps of ROI visualized by Grad-CAM algorithm. The most discriminative iris texture area is marked by red and yellow. (g) is the colorbar of the heat map, where the yellow and red area represents a higher score while the green and blue area correspond to the medium score, and the bottom 20% score is set to zero.
Figure 12. (af) are the heat maps of ROI visualized by Grad-CAM algorithm. The most discriminative iris texture area is marked by red and yellow. (g) is the colorbar of the heat map, where the yellow and red area represents a higher score while the green and blue area correspond to the medium score, and the bottom 20% score is set to zero.
Sensors 21 03721 g012
Table 1. The identification and verification results for different database using different model configuration.
Table 1. The identification and verification results for different database using different model configuration.
No.Database (Method)Training ClassesTesting ClassesIdentificationVerification
Register 1 PictureRegister 6 PicturesEERFRR@ FAR = 0.1%FRR@ FAR = 0.01%
1CASIA-V33320098.95%100%0.76%1.24%1.45%
2CASIA-V1010899.51%100%0.35%0.57%1.32%
3CASIA-V1208899.76%100%0.33%0.43%1.46%
4CASIA-V4064889.53%98.21%3.54%16.92%31.13%
5CASIA-V4561594.89%99.47%2.20%5.84%12.40%
6CASIA-V41061596.10%99.58%1.80%4.28%8.87%
7CASIA-V41561597.12%99.72%1.30%2.89%5.40%
8CASIA-V42061597.65%99.79%1.23%2.39%3.96%
9CASIA-V42561597.86%99.81%1.18%2.25%3.50%
10CASIA-V43061597.92%99.77%1.19%2.16%3.31%
Table 2. The ablation study under different model configuration.
Table 2. The ablation study under different model configuration.
No.StructureLossReg.Aug.Att.Pru.Fin.EER
1ResnetMSE×××1.55%
2AMSE×××1.26%
3AMSE××××1.86%
4AMSE××××3.40%
5ACE×××1.55%
6AMSE××1.09%
7CMSE×1.22%
8CMSE1.03%
9C MSE1.26%
10C oMSE0.76%
Reg., Aug., Att., Pru., and Fin. are the abbreviation of regularization term, augmentation layer, attention layer, pruning, and finetuning. † The model is trained from scratch. o The uncropped iris images with resolution of 30 × 360 serve as input.
Table 3. The ablation study of the online augmentation method.
Table 3. The ablation study of the online augmentation method.
No.BJ.HS.LS.EERFRR@ FAR = 0.1%FRR@ FAR = 0.01%
1×××2.80%9.46%17.05%
2××2.64%8.00%14.86%
3××2.50%8.89%16.83%
4××2.71%9.30%17.43%
5×2.36%7.84%16.16%
6×2.50%6.48%14.41%
7×2.36%7.84%14.48%
82.27%6.06%13.97%
BJ., HS., and LS. stand for the brightness jitter, horizontal shift, and longitudinal scaling.
Table 4. The comparison of the encoding ability in different depth of the proposed CNN.
Table 4. The comparison of the encoding ability in different depth of the proposed CNN.
LayerTypeDiscarding Accuracy (%)Feature LengthTime (ms)
1Input85.61 ± 2.4321,60021.00
2Attention85.35 ± 2.6121,60020.76
3Convolution77.65 ± 2.86118,800104.56
4BN72.37 ± 3.24118,800104.09
5ReLu72.19 ± 3.18118,800104.09
6MP75.32 ± 2.9229,70027.57
7Attention73.36 ± 3.2129,70028.48
8Convolution71.37 ± 3.1759,40052.91
9BN70.82 ± 3.2359,40053.02
10ReLu76.00 ± 3.2659,40053.74
11MP83.58 ± 2.6815,84015.67
12Convolution90.79 ± 1.3036,72033.68
13BN91.70 ± 0.9536,72033.37
14ReLu89.48 ± 1.8336,72033.28
15MP95.88 ± 0.8618,36018.31
16Convolution96.52 ± 0.7010,08010.99
17BN96.55 ± 0.6810,08010.82
18ReLu96.27 ± 1.0910,08010.90
19MP95.42 ± 1.3333605.00
20Convolution80.92 ± 2.669602.25
21BN84.39 ± 3.079602.25
22ReLu90.23 ± 1.929602.41
23Convolution77.75 ± 2.5781609.08
24BN77.48 ± 2.2281609.10
25ReLu72.51 ± 2.5081609.04
26GAP26.79 ± 1.75681.33
27FC 16.83 ± 1.2311.05
Attention refers to the radial attention layer.
Table 5. Comparison on the performance of different methods proposed in recent years.
Table 5. Comparison on the performance of different methods proposed in recent years.
NoStudiesYearMethodParameters/FLOPsEvaluation ProtocolAugmentationResult
1Othman et al. [50]2016IrisCode (2D-Gabor filter + Hamming Distance)-/-CASIA-V4: 602 classes (for testing)NoneCASIA-V4: 3.5% (Verification EER)
2Nguyen et al. [43]2017Pre-trained CNN (Dense-Net) + SVM-/-CASIA-V4: 1000 classes (Train: 70%, Test: 30%) *NoneCASIA-V4: 98.8% (Identification Accuracy)
3Alaslani et al. [67]2018Pre-trained CNN (Alex-Net) + SVM41 M/2.2 BCASIA-V1: 60 classes
CASIA-V3: 60 classes
CASIA-V4: 60 classes
(Train: 70%, Test: 30%) *
NoneCASIA-V1: 98.3%
CASIA-V3: 89%
CASIA-V4: 98%
(Identification Accuracy)
4Wang et al. [63]2018MiCoRe-Net>1.4 M/>50 MCASIA-V3: 218 classes (Train: 1346 images, Test: 218 images) *
CASIA-V4: 1000 classes (Train: 9000 images, Test: 1000 images) *
Rotation and CroppingCASIA-V3: 99.08%
CASIA-V4: 88.7%
(Identification Accuracy)
5Tobji et al. [64]2019FMnet15 K/10 MCASIA-V4: 1000 classes (Train: 70%, Test: 30%) *NoneCASIA-V4: 95.63% (Identification Accuracy)
7Boyd et al. [68]2019Pre-trained/Finetuned CNN (ResNet-50) + SVM25 M/5.1 BCASIA-V4: 1000 classes (Train: 70%, Test: 30%) *NoneCASIA-V4: 99.03% (Identification Accuracy)
6Liu et al. [45]2019Fuzzified image + Capsule network>4 M/-CASIA-V4: 1000 classes (Train: 80%, Test: 20%)NoneCASIA-V4: 83.1% (Identification Accuracy)
8Lee et al. [65]2019Deep ResNet-152 +Matching distance>53 M/>10 BCASIA-V4: 1000 classes (Train: 50%, Test: 50%)Translation and CroppingCASIA-V4: 1.33% (Verification EER)
9Proença et al. [49]2019VGG-19 based CNN138 M/-CASIA-V4: 2000 classes (Train: 1000 classes, Test: 1000 classes)Scale transform and Intensity transformCASIA-V4: 3.0% (Verification EER)
10Chen et al. [66]2020Tiny-VGG based CNN>10 M/>1.3 BCASIA-V4: 140 K pairs
(Train: 50,632 images on another database)
Contrast, Brightness, and DistortionCASIA-V4: 99.58% (Identification Accuracy)
CASIA-V4: 2.36% (Verification EER)
11Proposed Method2021Condensed 2-ch CNN33 K/49.1 MCASIA-V1: 108 classes (Finetune: 20 classes, Test: 88 classes)
CASIA-V3: 233 classes (Train: 33 classes, Test: 200 classes)
CASIA-V4: 648 classes (Finetune: 30 classes, Test: 615 classes)
Brightness jitter, Horizontal shift, and Longitudinal scaling
(Online)
CASIA-V1: 100%
CASIA-V3: 100%
CASIA-V4: 99.77%
(Identification Accuracy)
CASIA-V1: 0.33%
CASIA-V3: 0.76%
CASIA-V4: 1.19%
(Verification EER)
* Training set and testing set share same classes. K-Kilo, M-Million, B-Billion.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Liu, G.; Zhou, W.; Tian, L.; Liu, W.; Liu, Y.; Xu, H. An Efficient and Accurate Iris Recognition Algorithm Based on a Novel Condensed 2-ch Deep Convolutional Neural Network. Sensors 2021, 21, 3721. https://doi.org/10.3390/s21113721

AMA Style

Liu G, Zhou W, Tian L, Liu W, Liu Y, Xu H. An Efficient and Accurate Iris Recognition Algorithm Based on a Novel Condensed 2-ch Deep Convolutional Neural Network. Sensors. 2021; 21(11):3721. https://doi.org/10.3390/s21113721

Chicago/Turabian Style

Liu, Guoyang, Weidong Zhou, Lan Tian, Wei Liu, Yingjian Liu, and Hanwen Xu. 2021. "An Efficient and Accurate Iris Recognition Algorithm Based on a Novel Condensed 2-ch Deep Convolutional Neural Network" Sensors 21, no. 11: 3721. https://doi.org/10.3390/s21113721

APA Style

Liu, G., Zhou, W., Tian, L., Liu, W., Liu, Y., & Xu, H. (2021). An Efficient and Accurate Iris Recognition Algorithm Based on a Novel Condensed 2-ch Deep Convolutional Neural Network. Sensors, 21(11), 3721. https://doi.org/10.3390/s21113721

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop