Iris Recognition for Human Identification

DOI : 10.17577/IJERTV4IS120376

Download Full-Text PDF Cite this Publication

Text Only Version

Iris Recognition for Human Identification

Reda Hirmon Gebrahiwot Electrical and Computer Engineering Mekelle University

Mekelle, Ethiopia

Pro. Pei Zhijun ()

Electronics Engineering

Tianjin University of Technology and Education (TUTE) Tianjin, China

Abstract Iris recognition is a part of biometric identification which offers a new solution for personal identification, authentication and security by analyzing the random pattern of the iris. The iris recognition system automatically recognizes the identity of a person from a new eye image by comparing it to the human iris patterns stored in an iris template database. The iris template database is created using three steps the first step is segmentation. Hough transform is used to segment the iris region from the eye image of the CASIA database. The noise due to eyelid occlusions, reflections is eliminated in the segmentation stage. The next step is normalization. A technique based on Daugmans rubber sheet model was employed on the iris for creating a dimensionally consistent representation of the iris region. The last step is feature encoding. The normalized iris was convolved with 1D Log-Gabor filters then the phase data from 1D Log-Gabor filters was extracted and quantized to four levels to encode the unique pattern of the iris in to a bit-wise biometric template. At last Template of the new eye image will be compared with the iris template database using hamming distance.

Keywords Feature Extraction; Feature Matching; Biometric Identification; Iris Recognition.

  1. INTRODUCTION

    The human iris begins to form in the third month of gestation and the structure is complete by the eight month, even though the color and pigmentation continue to build through the first year of birth. After that, the structure of the iris remains stable throughout a persons life. It is composed of several layers which gives its unique appearance. This uniqueness is visually apparent when looking at its rich and small details seen in high resolution camera images under proper focus and illumination. The iris is the ring-shape structure that encircles the pupil, the dark centered portion of the eye, and stretches radially to the sclera, the white portion of the eye see Fig. 1.1

    The possibility that the iris of the eye might be used as a kind of optical fingerprint for personal identification was suggested originally by ophthalmologists [1], [2], [3], who noted from clinical experience that every iris had a highly detailed and unique texture, which remained unchanged in clinical photographs spanning decades (contrary to the occult diagnostic claims of iridology). The difference even exists between identical twins and between the left and right eye of the same eye [4].the only characteristics that is dependent on genetics is the pigmentation of the iris which determines its color.

    The iris recognition system automatically recognizes the identity of a person from a new eye image by comparing it to the human iris patterns stored in an iris template database.

    Fig. 1.1 Front view of human eye

    The iris template database is created using three steps see Fig. 1.2. The first step is segmentation. The iris boundaries are modeled as two circles which are not concentric. The inner circle is the papillary boundary between the pupil and the iris where as the outer circle is the limbic boundary between the iris and the sclera. The noise due to eyelid occlusions, reflections is eliminated in the segmentation stage. The next step is normalization of the segmented iris for creating a dimensionally consistent representation of the iris region. The last step is feature encoding. The normalized iris was convolved with filters then the phase data from filters was extracted and quantized to four levels to encode the unique pattern of the iris in to a bit-wise biometric template. Template of the new eye image will be compared with the iris template database and decision will be made.

    Fig. 1.2 basic process of iris recognition system

  2. SEGMENTATION

    The main objective here is to remove any non-useful information, namely the pupil-iris boundary and the part outside the iris (sclera, eyelids, and skin), eyelid occlusion. R. Wildes used Hough transforms to detect the iris contour. Daugman proposed an integro-differential operator to find both the pupil and the iris contour. The segmentation stage is critical to the success of an iris recognition system, since data that is falsely represented as iris pattern data will corrupt the biometric templates generated, resulting in poor recognition rates.

    The proposed algorithm is based on the fact that there is some obvious difference in the intensity around each boundary, so we need an edge detection algorithm. Locating the iris is not a trivial task since its intensity is close to that of the sclera and is often obscured by eyelashes and eyelids. However the pupil, due to its regular size and uniform dark shade, is relatively easy to locate. The success of segmentation depends on the imaging quality of eye images. Persons with darkly pigmented irises will present very low contrast between the pupil and iris region if imaged under natural light, making segmentation more difficult.

    A. Hough transform

    Any object can be described by mathematical functions that define the boundaries of that object. The Hough transform is a technique used in digital image processing to isolate features of a particular shape within an image [5]. This technique finds objects with a particular shape within a certain class of shapes. It uses parameters that describe any shape, creates a mesh of those parameters and then follows a voting procedure to cast votes in the mesh cells in some accumulator space. In most of the cases Hough Transform requires some kind of edge detection of the objects prior to performing transform. Due to various kinds of noises in the image, the detected edges may not be in a perfect shape. Hough Transform performs shape detection by grouping these edge points using a voting procedure in an accumulator space.

    Circular Hough Transform uses the parameterized equation of circle for this purpose the equation of circle can be written as:

    ( )2 + ( )2 = 2 (2.1)

    Circular Hough Transformer (CHT) is used to detect the inner and outer boundaries of the iris. First the eye image is edge detected in the vertical then Based on CHT algorithm for every edge a circle is drawn at last the maximum in the Hough space will be chosen when the maximum in the Hough space is chosen we find the circle parameters i.e. the center coordinates and the radius of the iris.

    The P position in Fig. 2.1 is getting the majority of votes so it will be the maximum.

    Fig. 2.1 accumulator hough space

    Fig. 2.2 flowchart of localization of outer iris boundary

    To improve the segmentation of pupil the brightness of the image was calculated. If the brightness was less than threshold Histogram equalization was applied to the previously detected iris if not the iris-pupil boundary is bright enough so histogram equalization is not necessary.

    Fig. 2.3 flowchart of localization of inner iris boundary

    The last stage in segmentation is detecting noises like eyelid occlusion and reflection. Linear Hough Transform (LHT) is used to detect the eyelid occlusion.

    Fig. 2.4 Segmentation of various eye images

  3. NORMALIZATION

    The main objective of normalization is to map the segmented iris image in to a fixed reference or coordinate system. It is required to extract fixed number of features from the iris region regardlss its spatial resolution. Two images of the same iris might be very different as a result of the size of the image Due to the distance between the eye and the capturing device and the camera optical magnification factor, size of the pupil due to the illumination and orientation of the iris. As described in [6], the invariance to all of these factors can be achieved through the translation of the captured data into a double dimensionless polar coordinate system. :

    A. Daugmans rubber sheet model

    The Cartesian to polar reference transform suggested by Daugman [7] authorizes equivalent rectangular representation of the zone of interest as shown in Fig. 3.1. In this way it is possible to compensate the stretching of the iris texture as the pupil changes in size, and unfold the frequency information contained in the circular texture in order to facilitate next feature extraction.

    This is given by

    = ± (3.1)

    With = ()2 + (oy)2

    = cos( tan1( oy)) )

    Where displacement of the centre of the pupil relative to the centre of the iris is given by , oy, and is the distance between the edge of the pupil and edge of the iris at an angle, around the region, and is the radius of the iris.

    Fig. 3.2 a) an eye image (S2001L03 from the CASIA- iris-lamp database) b) corresponding 20 pixels by 240 pixels data points of normalization c) corresponding polar representation d) corresponding noise.

  4. ENCODING

    After being able to localize the iris and normalize the iris it is time to extract the most discriminating feature in its pattern so that a comparison between templates can be done. The iris pattern provides two types of information: The amplitude information and the phase information. As shown by Oppenheim and Lim, and because of the dependence of the amplitude information on many extraneous factors, only phase information is used to generate the iris code [8]. Wavelets can be used to decompose the data in the iris region into components that appear at different resolutions, allowing therefore features that occur at the same position and resolution to be matched up.

    .

    Fig. 3.1 Daugmans rubber sheet model

    Since the pupil can be non-concentric to the iris, a remapping formula is needed to rescale points depending on the angle around the circle .

    Fig. 4.1 Encoding phase information

    A. Log-gabor filter

    Gabor filters are a traditional choice for obtaining localized frequency information. They offer the best simultaneous localization of spatial and frequency information. However they have two main limitations. The maximum bandwidth of a Gabor filter is limited to approximately one octave and Gabor filters are not optimal if one is seeking broad spectral information with maximal spatial localization.

    An alternative to the Gabor function is the Log-Gabor function proposed by Field. Log-Gabor filters can be constructed with arbitrary bandwidth and the bandwidth can be optimized to produce a filter with minimal spatial extent. The frequency response of a Log-Gabor filter is given as;

    () = (log(/))2/2(log(/))2 (4.1)

    Where represents the centre frequency, and gives the bandwidth of the filter.

    VI. RESULT

    The 100 hamming distance of the same iris capture in different time and condition from CASIA-Iris-Twins database was calculated.their mean, standard deviation and histogram distribution can be seen in Fig. 6.1

    Mean = 0.2791

    Standard deviation = 0.0839

  5. MATCHING

The hamming distance algorithm employed also incorporates noise masking, so that only significant bits are used in calculating the hamming distance between two iris templates. Now when taking the hamming distance, only those bits in the iris pattern that corresponds to 0 bits in noise masks of both iris patterns will be used in the calculation. The hamming distance will be calculated using only the bits generated from the true iris region, and the hamming distance formula is given as

Fig. 6.1 intra-class Hamming distance distribution of CASIA-Iris- Twins database

100 hamming distance of different iris from CASIA-Iris- Twins database was calculated.their mean, standard deviation

= ( 1

) (5.1)

and histogram distribution can be seen in Fig. 6.2

N

N

k=1

Xnk(OR)Ynk

Mean = 0.4649

j=1

= N

xj (XOR) yj (AND)Xnj(AND) ynj

Standard deviation = 0.0171

(5.2)

HD = a × b (5.3)

Where xj and yj are the two bit-wise templates to compare, Xnj and ynj are the corresponding noise masks for xj andyj,N is the number of bits represented by each template and HD is the hamming distance. In order to account for rotational inconsistencies, when the hamming distance of two templates is calculated, one template is shifted left and right bit-wise and a number of hamming distance values are calculated from successive shifts. The number of bits moved during each shift is given by two times the number of filters used, since each filter will generate two bits of information from one pixel of the normalized region. From the calculated hamming distance values, only the lowest is taken, since this corresponds to the best match between two templates.

Fig. 5.1 An illustration of the matching process

Fig. 6.2 inter-class Hamming distance distribution of CASIA-Iris-Twins database

Distribution of 100 hamming distance of the same iris capture in different time and condition from CASIA-Iris- Twins database and 100 hamming distance of different iris from CASIA-Iris-Twins database can be seen in Fig. 6.3 notice the separation in hamming distance of 0.4 that shows not all but most of the hamming distance from different iris is above 0.4 and the hamming distance from same iris is less than 0.4.

Fig. 6.3 inter-class and intra- class Hamming distance distribution of CASIA-Iris-Twins database

TABLE I. FALSE ACCEPTANCE AND REJECTION RATE FOR VARIOUS HAMMING DISTANCE VALUES

VII. CONCLUSION

The aim of this project was to implement methods used for iris recognition and to test those methods on a database of iris provided by the Chinese Academy of Sciences Institute of Automation (CASIA) using matlab .the system had 4 sub systems these are segmentation, normalization, feature encoding and matching .the segmentation was implemented using hough transformer to segment the iris from the eye image and its performance was tested on 100 eye image from CASIA database and it gives 84% success.

Normalization was done using Daugmans rubber sheet model to create dimensionally consistent representation of the iris region and it was tested on images of CASIA-Iris-Lamp it gives excellent result.

The normalized image was encoded using 1D-log Gabor filter and hamming distance was used for matching .hamming distance had to be carefully decide so as to use the iris recognition system with acceptable FAR and FRR. With hamming distance of 0.4 it gives 0% of FAR and 11.11% of FRR.

ACKNOWLEDGMENT

I would like to express my gratitude to my supervisor professor Pei Zhijun () for his guidance, helpful ideas, comments and support. I am sincerely grateful to the Chinese Academy of Sciences' Institute of Automation (CASIA) for generously providing the iris database. I am grateful to the department of Electronics Engineering of Tianjin University of technology and education (TUTE) for their assistance and privilege during my study.

Many thanks, to all my family and friends for supporting me along the way.

HAMMING DISTANCE

FALSE ACCEPTATION RATE (FAR) (%)

FALSE REJECTION RATE

(FRR) (%)

0.1

0.000

100.000

0.15

0.000

96.296

0.2

0.00

80.247

025

0.000

0.494

0.3

0.000

37.037

0.35

0.000

22.222

0.4

0.000

11.111

0.45

19.753

1.234

0.5

98.765

0.000

REFERENCES

  1. F.H. Adler, Physiology of the Eye: Clinical Application, fourth ed. London: the C.V Mosby Company, 1965.

  2. L. Flom and A. safir , U.S. Patent No. 4641349, U.S. Government Printing Office, Washington, DC, 1987.

  3. J. Rohen, Morphology and Pathology of the Trabecular Meshwork, in the structure of the eye, smelser, Ed. New York: Academic Press, pp. 335-341, 1961.

  4. Wilds, R.P, Iris Recognition: An Emerging Biometric Technology, Proceedings of the IEEE, vol. 85, no. 9, pp. 1348-1363, September 1997.

  5. D. H. Ballard, Generalizing the Hough transform to Detect Arbitrary Shapes, In Pattern Recognition, vol. 13, pp. 111-122, 1981.

  6. J. G. Daugman, How Iris Recognition Works, IEEE Trans. Circuits and system for Video Tech. 14(1), 21-30 2004.

  7. J. G. Daugman, High Confidence Visual Recognition of Person by a Test of statistical independence, IEEE Trans. PAMI 15, 1148-1161 (1993).

  8. A. Oppenheim, J. Lim, The Importance of Phase in Signals, Proceedings of the IEEE, vol. 69, pp. 529-541, 1981.

Leave a Reply