Research Article | Open Access

Yaser Daanial Khan, Sher Afzal Khan, Farooq Ahmad, Saeed Islam, "Iris Recognition Using Image Moments and k-Means Algorithm", *The Scientific World Journal*, vol. 2014, Article ID 723595, 9 pages, 2014. https://doi.org/10.1155/2014/723595

# Iris Recognition Using Image Moments and k-Means Algorithm

**Academic Editor:**Y.-B. Yuan

#### Abstract

This paper presents a biometric technique for identification of a person using the iris image. The iris is first segmented from the acquired image of an eye using an edge detection algorithm. The disk shaped area of the iris is transformed into a rectangular form. Described moments are extracted from the grayscale image which yields a feature vector containing scale, rotation, and translation invariant moments. Images are clustered using the k-means algorithm and centroids for each cluster are computed. An arbitrary image is assumed to belong to the cluster whose centroid is the nearest to the feature vector in terms of Euclidean distance computed. The described model exhibits an accuracy of 98.5%.

#### 1. Introduction

Identification of individuals has been an important need over the ages. Conventionally identification documents like an identity card, passport, or driving license have been utilized for the purpose. Such identification methods have been evaded several times by use of forged documents. In the digital world a login and a password or a PIN code is used for identification. Besides shoulder surfing and sniffing several other techniques have evolved which are used to crack such codes and breach security. Undoubtedly a robust identification technique is essential for a safe and well supervised environment. This situation thrives the need of an identification technique using some biological inimitable features of a person. Numerous biological human features are peculiar and unique such as fingerprints, suture patterns, iris patterns, gait, and ear shapes. The patterns found in these structures are unique for every human; hence they can be used as an identification tool. In the recent past, use of iris image of a person for his identification has gained popularity. The radial and longitudinal muscles in the iris of an eye are responsible for the constrictions and dilation of the pupil. The pupil changes its size depending upon the light intensity the eye is exposed to. The muscles of iris form the texture of the iris while the presence or absence of a pigment forms the color of the iris. The color of the iris is genetically dependent, whereas the texture is not. The texture of iris forms random unique patterns for each human. A close observation of an iris may reveal pustules, rings, stripes, and undulations forming a unique pattern.

In the recent past, researchers have developed several mechanisms for matching the pattern that lies within the iris. In [1] the author employs a bank of Gabor filters to form a fixed length vector from the local and global iris characteristics. Iris matching is established based on the weighted Euclidean distance between the two iris images being compared. In another article by Monro et al., a technique is devised using discrete cosine transform. Iris coding is based on the differences of discrete cosine transform coefficients of overlapped angular patches from normalized iris images [2]. Certain researchers have employed various statistical models for the purpose. A nonparametric statistical model, namely, neural networks (NN), is used for pattern matching and data compression in [3–5]. The image processing technique using specially designed kernels for iris recognition is used to capture local characteristics so as to produce discriminating texture features in [6]. Several sorts of transformations also prove helpful in extracting useful features from an iris image. This feature vector is further used to form a classification approach for identifying a person based on his iris image [7, 8]. In the groundbreaking work by Daugman the iris recognition principle is based on the failure of statistical independence tests on iris phase structure encoded by multiscale quadrature wavelets. The combinatorial complexity of this phase information across different persons generates discriminating entropy enabling the most probable decision about a person’s identity [9].

Most of the techniques based on feature extraction are designed for image of a certain fixed resolution. They fail to provide the desired result for the same images with different resolution. This characteristic implies that the model is not scale invariant. Techniques making use of NN incorporate a time taking training procedure. At times this training process may prove to be tricky rendering the model unable to yield quick results. On the other hand, some techniques that make use of certain filters may produce undesired results if the image is rotated which implies that such models are not rotation invariant. In this underlying paper a scale and rotation invariant technique for the same purpose is described. The proposed technique requires little training after which results are produced instantly. It is based on the use of image moments. Moments are properties that describe the characteristics of a certain distribution of data. Image moments (namely, Hu moments) are a quantitative measure of the shape of distribution formed by data collected as image pixel intensities and their locations [10].

In the proposed work the iris is segmented from an eye image. Image moments are computed from the segmented grayscale image. Classification of an iris is performed by the k-means algorithm. The composition of the paper is as described here. Section 2 gives an overview of iris recognition process. Section 3 explains the method used for iris segmentation. Section 4 gives a method for transforming the radial information of the iris into a rectangular form. Section 5 explains how this method can be further optimized. Image moments and method of computation of moments are described in Section 6. Section 7 describes the adoption of k-means algorithm for clustering and classification using moments information. Some of the results are discussed in Sections 8 and 9 presents some conclusions.

#### 2. Iris Recognition

Initially the image of an eye is acquired by a device called iriscope specifically designed for eye image acquisition at a high resolution. A large database of such images is collected having several classes. The iris within the image is segmented using an accurate and sufficiently fast technique. The iris image is of radial nature, rather than rectangular, which makes it unsuitable to be processed by any mathematical or statistical model of linear nature. There are two approaches to resolve this problem. The first approach is to adapt a model capable of processing data in its inherent radial form. Other approaches require transformation of the radial data into multidimensional linear form such that the information pertaining to iris texture is retained. In this piece of work the latter approach is adopted.

The information within the texture of the rectangular image may be used to form a probability density function. The image moments quantify the characteristics of this distribution. Using these raw moments translation, scale and rotation invariant moments are computed. Accumulated, these moments describe the characteristics of the pattern of the iris. This forms a feature vector which is later used for classification of iris images.

#### 3. Iris Segmentation

Each image in the database contains the iris pattern which is of interest; the rest of the image is of no use and therefore is not processed. The iris is extracted from the image using the segmentation process described in [5]. The iris is modeled as a disk-like structure consisting of two concentric circles (see Figure 2). The noise in the eye image is suppressed using numerous iterations of median filtering [11]. The image with reduced noise is filtered to extract edges using an edge detection algorithm like the Canny [12] or the Sobel [13] filter as shown in Figure 1(a). Now using the resultant image the iris outline is extracted. The image is scanned top to bottom, left to right line by line. Each point on the outer and the inner edge is stored in two separate arrays. These points are further used to determine the center and the radii of the concentric circles forming the iris and the pupil as shown in Figure 1(b). Assuming that the outline of the iris is a circle, a point on the circle with the center at satisfies the equation And the radius of the circle is given as Choosing any three arbitrary points from the array containing points of the circle, a system of simultaneous equations is formed. The solution to , , and in terms of the selected three points are derived from the system and is given as where , , and are the three arbitrary points.

**(a) The figure shows an iris image after edge detection**

**(b) The figure shows the disk shaped characteristics of iris. Note the circles overlapping the inner and outer circular edges**

Putting the values of and the value of is determined from the following equation: Moreover the radius is determined using (2). The center and the radii of both the concentric circles are determined in the described manner. The information within the inner circle is left out as it encompasses the pupil while the information bound in between the inner and outer circle contains the significant and unique iris pattern. Several triplets of circle points are used to compute the center of each circle. The best estimation of center is achieved by discarding extreme center points and then taking the mean of the rest of the points. For each center point of inner circle and for each center point for outer circle the mean is computed as Similarly, averages for the radius of the inner circle and the radius for outer circle are computed as The pattern formed by the radial and longitudinal muscles of the iris is of interest. Further, this pattern is extracted to form a rectangular image. A computationally moderate solution to the problem must provide a faster transformation method. One such method is discussed in [5] which transforms the radial image into rectangular form and further makes use of the midpoint algorithm to optimize it as described in the next section.

#### 4. Radial to Linear Transformation

An arbitrary Cartesian point anywhere on the disk-like structure having parametric coordinates is given as Cartesian points along the line starting at parametric point and ending at are discretized at appropriate intervals and are placed in a column of a two-dimensional array. A number of columns are collected starting from and incrementing it in small steps up to . The collection of these columns forms a rectangular canvas containing the required iris pattern, (see Figure 3).

**(a)**

**(b)**

The computations required to compute each point within the disk shaped structures are reduced by exploiting the symmetric properties of a circle. A circular shape exhibits eight-way symmetry [14]. This means for any computed point seven more points are determined that lie on the same circle using its symmetric properties. These seven points are described as , , , , , and given that the center lies at the origin. In case the center lies at an arbitrary point, then these points are translated accordingly. Use of this symmetric property reduces the computations eightfold. Each point on the line is determined by incrementing the -coordinate in discrete steps and calculating the corresponding value of -coordinate using the line equation.

The -coordinate and the -coordinate of the pixels along a single line making an arbitrary angle can be determined incrementally while the starting pixel coordinates are and the coordinates of the endpoint pixel are . Based upon the value of -coordinate of previous pixel, the value of -coordinate for next pixel is calculated by incrementing the previous value of . This value of -coordinate, say , is put into the following equation: which yields the corresponding -coordinate.

#### 5. Optimizing the Algorithm

The determination of points along a line is further optimized by the use of the midpoint method [14]. The computations required to yield a point along a line are reduced to mere addition of a small incremental value. The gradient is computed for this purpose as given in the following equation:

Let where the line end points are and . In accordance with the midpoint method for straight line with tangent between 0 and 1 the value of and . Initially the control variable . If the value of is positive then East Pixel is chosen and if it is negative then North East Pixel is chosen. At each step is updated by adding or accordingly [5, 14].

#### 6. Pattern Recognition Using Image Moments

A perceptive action performed on intricate structures needs to quantify its attributes. The state of any structure is quantifiable into data. Diversification of this data represents interaction or changes in the state. All such quantification methods generate finite data. Data by itself is insignificant, but the information implanted within the data is useful. Information is either extracted directly from the data itself or from the patterns formed by the arrangement of data. Researchers have devised various models for extracting information from data embedded in an image. Applications based on such models do not add to the contents of data rather they find hidden data patterns in order to extract interesting and useful information. A probability density can be formed for any data set. The parameters of the probability density function inform us about the general manner in which data is distributed. Moments are the characteristics of the probability density function which are based on the kurtosis and skewedness of the probability density function. Image moments describe the properties of a distribution formed using the pixel data of the image along its axes. The moments are typically chosen to depict a certain interesting property of the image. Such moment proves beneficial in extracting and summarizing the properties of the image in order to produce useful results. Properties of an image such as centroid, area, and orientation are quantified by this process. Another dividend of image moments is that they bring together the local and global geometric details of a grayscale image [15].

##### 6.1. Extracting Moments from an Image

An image in the real world is modeled using a Cartesian distribution function in its analog form. This function is used to provide moments of the order of over the image plane and is generalized as where is the basis function and is the image plane. Equation (11) yields a weighted average over the plane . The basis function is designed such that it represents some invariant features of the image. Furthermore the properties of the basis function are passed onto moments. An image is of discrete nature; thus it is divided into pixels each having a discrete intensity level. Equation (11) is adopted for the digital image as where is the intensity of a pixel in the digital image at the th row and th column.

In [10, 16] the authors prove that the two-dimensional continuous th order moments are defined using the integral where lies within some finite region of the plane. In case of digital image the integrals are replaced by summations, which is formulated as where is the basis function, and are the dimensions of the image, and the is the cartesian moment for the two-dimensional image. Note that this basis function is highly correlated, that is, nonorthogonal. The moment represents the image, whereas the first order moments are used to find the center of the mass or the centroid of the image is given as where is the centroid.

##### 6.2. Centralized Moments

Once the centroid is determined, it is used to compute the centralized moments. In [15] the central moments for two-dimensional data are given as where are the central moments. Note that these are similar to Cartesian moments translated to the centroid. This depicts the translation invariant property of the centralized moments which are always akin to the centroid of the segmented object. Further simplification of (16) up to order generates the following moments:

##### 6.3. Scale Invariant Moments

These moments are further made scale invariant as explained in [4, 5] and are given as where are scale normalized central moments and where .

##### 6.4. Image Orientation

The second order central moments contain information about the orientation of the image. Using these moments a covariance matrix is derived. Let and then the covariance matrix is given as

The major and minor axes of the image intensity correlate with the eigenvectors of the given covariance matrix. The orientation of the image is described by the eigenvector of the highest eigenvalue. In [4] it is shown that the angle is computed by the following equation: where Using (20) the eigenvalues of the covariance matrix are easily obtained and are given as Notice that these values are proportional to the square of the length of the eigenvector axes. The difference between the eigenvalues marks yet another important characteristic. It shows how elongated the image is. This property is termed eccentricity and is computed as

##### 6.5. Rotation Invariant Moments

Previously we have discussed translation and scale invariant moments. In [16] rotation invariant moments are derived which are usually termed as a Hu set of invariant moments. These are given as follows: Every one of the rotation invariant moments extracts a characteristic attribute of the image. For example represents the moment of inertia along the centroid while extracts skew invariant properties which are useful in differentiating between images which are mirror reflections of each other [10, 15–17].

#### 7. Clustering for Classification

By now the iris image has been segmented and transformed into a rectangular canvas. All described moments are applied and a feature vector is extracted, namely, . This vector contains translation, scale, and rotation invariant and orientation related moments. This vector corresponds to various features of the image; hence it is used for classification. An unsupervised approach is adopted for classification using k-means clustering algorithm. Using a set of multidimensional observations , the k-means algorithm partitions the observations into sets such that generating the set so as to minimize the following objective function: where is mean of all the observations in . Initially extracted moment vectors for an iris image sample are considered to be the initial mean.

The k-means Algorithm has two major steps, namely, the assignment step and the update step. The mean is used to assign each observation to a cluster in the assignment step. An observation is assigned a cluster whose mean makes the closest match. Formally this step generated the set such that

Also an observation should be associated with exactly one even if two or more differences are found comparable. The next step is based on the identification of a cluster for the observation established in the previous step. The mean for the cluster is recalculated as the centroid of the observations as given in following equation: Both steps are iterated and the centroid is readjusted. This process continues until there is no appreciable change in the means. At this stage the means have converged and no further training is required [18, 19].

#### 8. Results

The CASIA database containing thousands of images belonging to hundreds of different people is used to gather test results. Nearly one-fourth of the iris images from each class are retained as test case while the rest are used for training. The distorted images within the database are rejected. Iris portion of the image is marked out using the segmentation algorithm and is later transformed into a rectangular canvas. Further the grey scale rectangular canvas of iris is used to compute image moment vector. This vector contains information which is translation, scale, and rotation invariant and provides orientation information. Using the k-means algorithm each image is assigned to a cluster. The k-means algorithm is iterated until convergence is achieved and the centroid of each cluster is determined. Once the system is fully trained it is ready to accept an arbitrary input and provide a match. The model responds with the correlation of an arbitrary image moments vector to a cluster, if the image belongs to a known class. In Figure 4 various clusters formed are depicted; it also shows how the class of a sample is distinguished based upon the Euclidean distance of the feature vector of the sample from the centroid of an arbitrary cluster. Moreover Figure 5 shows a confusion matrix depicting the accuracy of the model. The confusion matrix shows that the accuracy of the model for certain arbitrary classes is 99.0% while the overall accuracy of the model for all the images in the database is estimated to be 98.5%. Moreover it also reports the level of confidence of match based on Euclidean distance of the sample from the centroid of the identified cluster. Level 0 is the highest which means that the Euclidean distance of the sample from the centroid of the cluster is low and level 4 is the lowest which indicates that the Euclidean distance of the sample from the centroid of any cluster does not lie within a stipulated threshold to confidently indicate a match. Figure 4 shows the clusters formed using the k-means algorithm.

**(a)**

**(b)**

Furthermore a number of experiments were carried out to determine the accuracy and efficiency of the proposed model in comparison with other competitive models. In [20] the authors present a technique which extracts the features of iris using fine-to-coarse approximation at different resolution levels determined through discrete dyadic wavelet transform zero crossing representation. The resultant one-dimensional feature vector is used to find a match by computing various distances with arbitrary feature vectors. Ma et al. present yet another iris recognition technique using Gabor filters [1, 6]. The authors use a bank of Gabor filters to extract a fixed length feature vector signifying the global and local characteristics of the iris image. A match is established by computing the weighted Euclidean distance between feature vectors of arbitrary iris images. Daugman in [9] relies on the morphogenetic randomness of texture in the trabecular meshwork of iris. A failure of statistical independence test on two coded patterns from same iris indicates a match. This method extracts the visible texture of iris from a real time video image. The image is later encoded into a compact sequence of multiscale quadrature 2D Gabor wavelet coefficients. The most significant 256 bytes form the iris code. An exclusive OR operation is performed to generate a decision. All the above-mentioned techniques including the proposed technique are executed in order to obtain results. The genuine acceptance rate (GAR) and false acceptance rate (FAR) are observed for each technique. A receiver operating characteristics (ROC) distribution is plotted for each technique based on the results as shown in Figure 6. The ROC distribution comparatively highlights the accuracy along with the frequency of occurrence of errors of the proposed and other current state-of-art models. The following section briefly provides some discussion about the proposed system along with an interpretation of the ROC distribution formed.

#### 9. Conclusion

Through analysis of data obtained after moments extraction a number of conclusions are inferred. Images of a certain iris differing in orientation yielded varying eigenvalues and eccentricity. However, a change in orientation of an image barely affects the values of rotation invariant moments while raw and scale invariant moments are affected. Change in orientation of an image affects the Euclidean distance of the moment vectors from the centroid. Despite this there still remains a great probability of the image to be classified correctly because of coherence in scale invariant moments. Although the model exhibits scale and rotation invariant attributes but some impairment is offered by luminosity of the image. Two arbitrary images of the same objects yield comparable moments if the luminosity is the same but they may yield differing moments in case luminosity is altered. In the underlying research work it is assumed that the luminosity level will be the same for all the images as each image is obtained by an iriscope working in similar conditions. The model provides resilience towards variation of scale and rotation as compared to other techniques which requires coherence of phase and size. The model can be further improved by incorporation of a technique that will process each image to provide uniform luminosity. Furthermore, the ROC distribution obtained (shown in Figure 6) from all the test cases shows that the performance of proposed model is comparable with Daugman method, while it yields a better performance than the methods described in [1, 6, 20].

#### Conflict of Interests

The authors declare that there is no conflict of interests regarding the publication of this paper.

#### References

- L. Ma, T. Tan, Y. Wang, and D. Zhang, “Efficient iris recognition by characterizing key local variations,”
*IEEE Transactions on Image Processing*, vol. 13, no. 6, pp. 739–750, 2004. View at: Publisher Site | Google Scholar - D. M. Monro, S. Rakshit, and D. Zhang, “DCT-bsed iris recognition,”
*IEEE Transactions on Pattern Analysis and Machine Intelligence*, vol. 29, no. 4, pp. 586–595, 2007. View at: Publisher Site | Google Scholar - K. Roy, P. Bhattacharya, and R. C. Debnath, “Multi-class SVM based iris recognition,” in
*Proceedings of the 10th International Conference on Computer and Information Technology (ICCIT '07)*, pp. 1–6, Dhaka, Bangladesh, December 2007. View at: Publisher Site | Google Scholar - R. H. Abiyev and K. Altunkaya, “Personal iris recognition using neural network,”
*International Journal of Security and Its Applications*, vol. 2, no. 2, pp. 41–50, 2008. View at: Google Scholar - Y. D. Khan, F. Ahmad, and M. W. Anwar, “A neuro-cognitive approach for iris recognition using back propagation,”
*World Applied Sciences Journal*, vol. 16, no. 5, pp. 678–685, 2012. View at: Google Scholar - L. Ma, Y. Wang, and T. Tan, “Iris recognition based on multichannel Gabor filtering,” in
*Proceedings of the 5th Asian Conference on Computer Vision*, pp. 279–283, Melbourne, Australia, 2002. View at: Google Scholar - W. W. Boles and B. Boashash, “A human identification technique using images of the iris and wavelet transform,”
*IEEE Transactions on Signal Processing*, vol. 46, no. 4, pp. 1185–1188, 1998. View at: Publisher Site | Google Scholar - K. Lee, S. Lim, O. Byeon, and T. Kim, “Efficient iris recognition through improvement of feature vector and classifier,”
*ETRI Journal*, vol. 23, no. 2, pp. 61–70, 2001. View at: Google Scholar - J. Daugman, “How iris recognition works,”
*IEEE Transactions on Circuits and Systems for Video Technology*, vol. 14, no. 1, pp. 21–30, 2004. View at: Publisher Site | Google Scholar - M.-K. Hu, “Visual pattern recognition by moment invariants,”
*IRE Tranactions on Information Theory*, vol. 8, no. 2, pp. 179–187, 1962. View at: Publisher Site | Google Scholar - E. Arias-Castro and D. L. Donoho, “Does median filtering truly preserve edges better than linear filtering?”
*Annals of Statistics*, vol. 37, no. 3, pp. 1172–1206, 2009. View at: Publisher Site | Google Scholar - J. Canny, “A computational approach to edge detection,”
*IEEE Transactions on Pattern Analysis and Machine Intelligence*, vol. 8, no. 6, pp. 679–698, 1986. View at: Publisher Site | Google Scholar - N. Kanopoulos, N. Vasanthavada, and R. L. Baker, “Design of an image edge detection filter using the Sobel operator,”
*IEEE Journal of Solid-State Circuits*, vol. 23, no. 2, pp. 358–367, 1988. View at: Publisher Site | Google Scholar - J. D. Foley,
*Computer Graphics: Principles and Practice*, Addison-Wesley, Reading, Mass, USA, 1996. - S. O. Belkasim, M. Shridhar, and M. Ahmadi, “Pattern recognition with moment invariants: a comparative study and new results,”
*Pattern Recognition*, vol. 24, no. 12, pp. 1117–1138, 1991. View at: Publisher Site | Google Scholar - J. Flusser, T. Suk, and B. Zitová,
*Moments and Moment Invariants in Pattern Recognition*, John Wiley & Sons, Chichester, UK, 2009. - W. T. Freeman, D. B. Anderson, P. A. Beardsley et al., “Computer vision for interactive computer graphics,”
*IEEE Computer Graphics and Applications*, vol. 18, no. 3, pp. 42–52, 1998. View at: Publisher Site | Google Scholar - D. Aloise, A. Deshpande, P. Hansen, and P. Popat, “NP-hardness of Euclidean sum-of-squares clustering,”
*Machine Learning*, vol. 75, no. 2, pp. 245–248, 2009. View at: Publisher Site | Google Scholar - J. A. Hartigan and M. A. Wong, “Algorithm AS 136: a k-means clustering algorithm,”
*Journal of the Royal Statistical Society C: Applied Statistics*, vol. 28, no. 1, pp. 100–108, 1979. View at: Google Scholar - C. Sanchez-Avila and R. Sanchez-Reillo, “Two different approaches for iris recognition using Gabor filters and multiscale zero-crossing representation,”
*Pattern Recognition*, vol. 38, no. 2, pp. 231–240, 2005. View at: Publisher Site | Google Scholar

#### Copyright

Copyright © 2014 Yaser Daanial Khan et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.