=Paper=
{{Paper
|id=Vol-2744/paper31
|storemode=property
|title=Hybrid Iris Segmentation Method Based on CNN and Principal Curvatures
|pdfUrl=https://ceur-ws.org/Vol-2744/paper31.pdf
|volume=Vol-2744
|authors=Varvara Tikhonova,Elena Pavelyeva
}}
==Hybrid Iris Segmentation Method Based on CNN and Principal Curvatures==
Hybrid Iris Segmentation Method Based on CNN and Principal Curvatures Varvara Tikhonova[0000−0002−7386−0250] and Elena Pavelyeva[0000−0002−3249−2156] Faculty of Computational Mathematics and Cybernetics, Lomonosov Moscow State University, Moscow, Russia varvara.a.tikhonova@gmail.com, paveljeva@yandex.ru Abstract. In this article the new hybrid iris image segmentation method based on convolutional neural networks and mathematical methods is proposed. Iris boundaries are found using modified Daugman’s method. Two UNet-based con- volutional neural networks are used for iris mask detection. The first one is used to predict the preliminary iris mask including the areas of the pupil, eyelids and some eyelashes. The second neural network is applied to the enlarged image to specify thin ends of eyelashes. Then the principal curvatures method is used to combine the predicted by neural networks masks and to detect eyelashes cor- rectly. The proposed segmentation algorithm is tested using images from CASIA IrisV4 Interval database. The results of the proposed method are evaluated by the Intersection over Union, Recall and Precision metrics. The average metrics values are 0.922, 0.957 and 0.962, respectively. The proposed hybrid iris image segmen- tation approach demonstrates an improvement in comparison with the methods that use only neural networks. Keywords: Biometrics, Iris Identification, Iris Segmentation, Principal Curva- tures, Convolutional Neural Network. 1 Introduction Iris recognition is one of the most accurate methods of biometric identification. In order to determine the features which are necessary for person identification by the iris image, the image segmentation must be performed. It includes determination of the inner and outer boundaries of the iris (iris localization) and the areas where eyelids and eyelashes overlap an eye. The result of segmentation is a mask, which is a binary image of the visible iris region. The mask helps to identify areas suitable for further parameterization. Iris segmentation is an important stage of the iris image preprocessing. Unlike the texture of the iris, the position of eyelashes and eyelids are not constant. Nevertheless, only regular features could be used for the identification, so it is necessary to exclude all the variable parameters in the eye image. There are some mathematical methods for iris localization: integro-differential Daug- man’s method [1], Hough’s transform-based method [2], fast method named ”Pulling and Pushing” [3]. Eyelids and eyelashes are often found as parabolas and noises [4, 5]. Copyright c 2020 for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0). 2 V. Tikhonova, E. Pavelyeva The evolution of the neural networks increases the popularity of their use. Neu- ral network-based methods give usually better results than mathematical methods. The most popular architectures for iris segmentation are: convolutional neural network (CNN) such as U-Net [6, 7], SegNet [8]; hierarchical convolutional neural networks (HCNN) and multi-task fully convolutional networks (MTFCN) [9]; deep neural networks (DNN) [10]; densely connected fully convolutional networks (IrisDenseNet) [11]; dense deep convolutional neural networks (DDNet) [12] and other architectures. However, it is nec- essary to control the results of neural networks work. Therefore, the use of a hybrid method improves the results of CNN application. The hybrid iris image segmentation is proposed in the article. The Daugman’s integro-differential method is modified and used for iris localization. To obtain iris mask image two UNet-based convolutional neural networks are used. First CNN is used to predict the preliminary mask of the iris, and the second one is used to specify thin ends of eyelashes. The modified principal curvatures method is used to combine pre- dicted masks and to detect eyelashes correctly. Test results using CASIA IrisV4 Interval database [13] show the effectiveness of the proposed iris segmentation method. 2 Iris Localization The iris localization method proposed by John Daugman [1] is based on the formula: I ∂ I(x, y) max(r,x0 ,y0 ) Gσ (r) ∗ , (1) ∂r 2·π·r r,x0 ,y0 where Gσ (r) is a Gaussian function and I(x, y) is an iris image intensity function, in- tegration is carried out along a circular contour with a radius r centered at the point (x0 , y0 ) of the image. Usually different values of smoothing parameter σ are used for inner and outer iris boundaries detection. For outer iris boundary the integration is car- ried out not over the entire circle, but only along the right and left arcs, divided into two parts (Fig. 1). The measure of the arc in each of the areas is calculated by the formula [14]: 90◦ · Ii αi = 4 ; i ∈ {1, 2, 3, 4}, (2) P Ij j=1 where Ii is the average image intensity in the corresponding region. Iris pupil boundary is close to the circle but it is necessary to precise it for the accu- rate iris segmentation. The image is translated into a polar coordinate system centered at the pupil center and the Canny filter [15] is applied to the image (Fig. 2). The compar- ison of the internal boundaries found by the modified and classical methods is shown in Fig. 3. The modified Daugman’s integro-differential method accurately determines the in- ner and outer iris boundaries, but it is not able to determine the areas of eyelids or eyelashes. Hybrid Iris Segmentation Method Based on CNN and Principal Curvatures 3 Fig. 1. Divided regions of the image. Fig. 2. Determination of the inner iris boundary. Fig. 3. Comparison of iris localization using the classical (red line) and modified (blue line) Daug- man’s method. 3 Iris Segmentation 3.1 CNN for iris segmentation In order to determine the areas of eyelids and eyelashes, a U-net–based [7] convolu- tional neural network is used (Fig. 4). Binary cross-entropy is used as a loss function, the Jacquard measure is used as a metric [16], the Adam optimizer [17] for optimiza- tion. Ground truth images are needed to train the neural network. There is the open- source database of labeled images [18], but eyelashes are not marked there. So, 150 eye images from the Casia IrisV4 Interval database [13] were manually labeled in related work: 100 for the training set, 27 for the validation and 23 for the test set. The training is conducted through 40 epochs, with 100 iterations in each. To increase the number 4 V. Tikhonova, E. Pavelyeva of images in the training data, augmentations were used – 16 pixels left and right, 14 pixels up and down shifts and rotations up to 5 degrees. The validation set is used after passing through each epoch in order to adjust the learning rate if the accuracy values and loss functions do not improve over six epochs. Fig. 4. Neural network architecture. However, the neural network trained on the source data is not able to distinguish the thin ends of the eyelashes due to the strong size reduction of the image after the third layer (Fig. 5), while two layers are not enough for the right segmentation. There are several ways to determine the ends of thin eyelashes: to add a dilated convolution layer to increase image size or to use a second neural network with the same architecture, which is trained on the double sized with bicubic interpolation images from the same training dataset. This network is able to determine eyelashes better, some iris features could be recognized as eyelashes. The results of both networks are combined using the curvature method. 3.2 Principal curvatures method The image can be represented as a three-dimensional surface, taking the intensity at each point as the value of the z-coordinate [19, 20]. Let L(x, y) be the Gaussian smoothed image, Lxx (x, y), Lxy (x, y), Lyx (x, y), Lyy (x, y) be its second derivatives. Local char- acteristics of the image L(x, y) at the point (x, y) can be determined using the Hessian matrix: Lxx (x, y) Lxy (x, y) H(x, y) = . (3) Lyx (x, y) Lyy (x, y) Eigenvalues of the Hessian matrix H(x, y) are denoted by λ1 , λ2 , where |λ1 | > |λ2 |, and the eigenvectors corresponding to them are ν1 , ν2 . The direction and value of the maximum curvature at the point (x, y) will correspond to the vector ν1 and the value Hybrid Iris Segmentation Method Based on CNN and Principal Curvatures 5 (a) (b) (c) (d) Fig. 5. Comparison of masks predicted by the first and second neural networks. (a) is the original image, (b) is the mask predicted by the first network, (c) is the mask predicted by the second network, (d) is a comparison of the masks. The matching white areas are shown in green, the areas that only the first network has identified as the iris are shown in red, the areas that only the second network has identified as the iris are shown in yellow. of λ1 , the minimum – ν2 and λ2 [20]. Vector ν2 is directed along the tubular-shaped area, while vector ν1 is directed across it. Thus, each next eyelash point can be found by moving from the previous one in the direction of the vector ν2 to a distance proportional to λ1 with coefficient k. Fig. 6 shows the field of such vectors in the eye image: it is clearly seen that the direction of vector ν2 lines up along the eyelashes. However, there are other vectors in the image. To avoid it, the threshold parameter γ is used. At each starting eyelash point of the image the eigenvalues and eigenvectors of the |λ1 | Hessian matrix are calculated. If the point conforms to the rule |λ 2| > γ, the next step of the algorithm is applied at a point located in the direction of the vector ν2 at a distance proportional to k · λ1 from the previous point, where the parameter k is: |λ1 | 1 k = max(0.15, min( · , 0.7)). (4) |λ2 | 100 |λ1 | The algorithm stops if the next step leads to the same point or if the condition |λ 2| >γ is not met. 3.3 Hybrid iris segmentation The proposed iris segmentation method (Fig. 7) can be described as follows: first, the iris mask is obtained using the first CNN. After that, using the second neural network, the iris mask of enlarged image is found to refine thin eyelashes. Then, the principal curvature method that consists of three stages is applied to the results of both networks (Fig. 7 d, e, f). 6 V. Tikhonova, E. Pavelyeva Fig. 6. The vector field of the eye image corresponding to vector ν2 , and the length of vectors is proportional to λ1 . 1. At the first stage, only the points with the image intensity values less than a thresh- old, that were predicted by the first network are taken as the starting points (Fig. 7 d). γ1 = 1.7. 2. At the second stage, only the points marked as eyelashes in the image after the first stage are taken as the starting points. This stage was performed to thicken and connect interrupted eyelashes, the intensity of which in the original image is less than the threshold. (Fig. 7 e) γ2 = 2.0. 3. At the third stage, only the points recognized as eyelashes by the second network and adjoined to the points from the second stage are taken as the starting points. This step is performed to extend the thin ends of the eyelashes. (Fig. 7 f) γ3 = 1.8. Thus, the image mask with eyelashes is obtained. Then, the modified Daugman’s algorithm is applied to the original image to obtain the inner and outer iris boundaries. After that, the region of obtained mask, between the outer and inner iris boundaries is taken as iris image mask. 4 Results The results of the proposed method for the images from the CASIA IrisV4 database are presented in Fig. 8 and Fig. 9. Fig. 8 represents the change of the mask at each stage of the algorithm and comparison of segmentation results with manually marked ground truth segmentation. Green in the last column (Fig. 8 g) indicates the intersec- tion areas, red indicates the areas marked by the proposed algorithm as the iris but not highlighted in the manually marked image, yellow indicates the opposite situation. The results before and after applying the principal curvature method (Fig. 8 d and Fig. 8 e) are evaluated by three metrics: Intersection over Union (IoU), Precision (Prec) and Recall (Rec). The Recall metric value decreases since while clarifying the eyelashes, the areas between them could also be marked as eyelashes that leads to increase in the number of pixels falsely labeled as eyelashes. However, the areas marked as eyelashes Hybrid Iris Segmentation Method Based on CNN and Principal Curvatures 7 Fig. 7. Description of the proposed method: (a) the source image; (b) the mask predicted by the first neural network; (c) the mask predicted by the second neural network; (d)-(f) stages of curvature method; (g) the result of iris localization; (h) the result of the method. are not used for further biometric identification. For this reason, reducing the number of eyelash points falsely marked as iris points is the priority of the method. The increase of the Precision metric indicates that the principal curvatures method improves the re- sults of the CNN-based method. Fig. 9 shows the overlay of the obtained mask on the original image. (a) (b) (c) (d) (e) (f) (g) Fig. 8. The results of the proposed iris segmentation method. (a) the original image, (b) the result of the modified Daugman’s method, (c) the mask predicted by the first neural network, (d) the intersection of the mask predicted by neural network and the area selected by the Daugman’s method, (e) the result of the proposed method: the intersection of the mask specified using the curvature method and the area selected by the Daugman’s method, (f ) the manually marked iris image segmentation, (g) the comparison of the proposed and manually marked iris segmentation. 8 V. Tikhonova, E. Pavelyeva (a) (b) (c) Fig. 9. The results of the proposed method. (a) the original image, (b) the result of the proposed iris segmentation method, (c) the overlay of the segmented iris region on the original image. In the article [6] U-net with different depth values is used for the iris segmentation. The comparison of average values of Precision, Recall and Intersection over Union metrics, gained in this article with the corresponding metrics obtained in [6] is given in Table 1. Table 1. Comparison to competing method. Method Database Avg. Precision Avg. Recall Avg .IoU U-net [6], depth 3 Casia Iris V1 0.937 0.957 0.899 U-net [6], depth 4 Casia Iris V1 0.948 0.955 0.908 U-net [6], depth 5 Casia Iris V1 0.948 0.960 0.912 Proposed method Casia Iris V4 0.962 0.957 0.922 5 Conclusion In this paper the new hybrid iris segmentation method based on CNNs and principal curvatures method is proposed. Experimental results obtained using Casia IrisV4 Inter- val database demonstrate that the proposed approach gives good segmentation results. The evaluation of this method by the IoU metric is 92%, which appears to be higher than the results of the methods that use only a neural network with a similar depth. Hybrid Iris Segmentation Method Based on CNN and Principal Curvatures 9 References 1. Daugman, J.G.: High confidence visual recognition of persons by a test of statistical indepen- dence. IEEE transactions on pattern analysis and machine intelligence 15(11), 1148–1161 (1993) 2. Wildes, R.P.: Iris recognition: an emerging biometric technology. Proceedings of the IEEE 85(9), 1348–1363 (1997) 3. He, Z., Tan, T., Sun, Z., Qiu, X.: Toward accurate and fast iris segmentation for iris bio- metrics. IEEE transactions on pattern analysis and machine intelligence 31(9), 1670–1684 (2008) 4. Min, T.H., Park, R.H.: Eyelid and eyelash detection method in the normalized iris image using the parabolic hough model and otsu’s thresholding method. Pattern recognition letters 30(12), 1138–1143 (2009) 5. Kong, W., Zhang, D.: Accurate iris segmentation based on novel reflection and eyelash de- tection model. In: Proceedings of 2001 International Symposium on Intelligent Multimedia, Video and Speech Processing. ISIMP 2001 (IEEE Cat. No. 01EX489). pp. 263–266. IEEE (2001) 6. Lozej, J., Meden, B., Struc, V., Peer, P.: End-to-end iris segmentation using u-net. In: 2018 IEEE International Work Conference on Bioinspired Intelligence (IWOBI). pp. 1–6. IEEE (2018) 7. Ronneberger, O., Fischer, P., Brox, T.: U-net: Convolutional networks for biomedical im- age segmentation. In: International Conference on Medical image computing and computer- assisted intervention. pp. 234–241. Springer (2015) 8. Badrinarayanan, V., Kendall, A., Cipolla, R.: Segnet: A deep convolutional encoder-decoder architecture for image segmentation. IEEE transactions on pattern analysis and machine in- telligence 39(12), 2481–2495 (2017) 9. Liu, N., Li, H., Zhang, M., Liu, J., Sun, Z., Tan, T.: Accurate iris segmentation in non- cooperative environments using fully convolutional networks. In: 2016 International Confer- ence on Biometrics (ICB). pp. 1–8. IEEE (2016) 10. Bazrafkan, S., Thavalengal, S., Corcoran, P.: An end to end deep neural network for iris segmentation in unconstrained scenarios. Neural Networks 106, 79–95 (2018) 11. Arsalan, M., Naqvi, R., Kim, D., Nguyen, P., Owais, M., Park, K.: Irisdensenet: Robust iris segmentation using densely connected fully convolutional networks in the images by visible light and near-infrared light camera sensors. Sensors 18(5), 1501 (2018) 12. Chen, Y., Zeng, Z., Hu, F.: End to end robust recognition method for iris using a dense deep convolutional neural network. In: Chinese Conference on Biometric Recognition. pp. 364– 375. Springer (2019) 13. Casia iris image database. idealtest.org/dbDetailForUser.do?id=4 14. Pavelyeva, E.A., Krylov, A.S.: Iris identification algorithm using the most informative iris points. In: International Conference on Computer Graphics and Machine Vision ”Graph- icon”. pp. 228–231 (2009) 15. Canny, J.: A computational approach to edge detection. IEEE Transactions on pattern analy- sis and machine intelligence (6), 679–698 (1986) 16. Chen, Y., Wang, W., Zeng, Z., Wang, Y.: An adaptive cnns technology for robust iris seg- mentation. IEEE Access 7, 64517–64532 (2019) 17. Kingma, D.P., Ba, J.: Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014) 18. Hofbauer, H., Alonso-Fernandez, F., Wild, P., Bigun, J., Uhl, A.: A ground truth for iris segmentation. In: 2014 22nd international conference on pattern recognition. pp. 527–532. IEEE (2014) 10 V. Tikhonova, E. Pavelyeva 19. Safronova, E.I., Pavelyeva, E.A.: Palm vein recognition algorithm using multilobe differen- tial filters. In: International Conference on Computer Graphics and Machine Vision ”Graph- icon”. pp. 117–121 (2019) 20. Choi, J.H., Song, W., Kim, T., Lee, S.R., Kim, H.C.: Finger vein extraction using gradient normalization and principal curvature. In: Image Processing: Machine Vision Applications II. vol. 7251, p. 725111. International Society for Optics and Photonics (2009)