- About this Journal
- Abstracting and Indexing
- Aims and Scope
- Annual Issues
- Article Processing Charges
- Articles in Press
- Author Guidelines
- Bibliographic Information
- Citations to this Journal
- Contact Information
- Editorial Board
- Editorial Workflow
- Free eTOC Alerts
- Publication Ethics
- Reviewers Acknowledgment
- Submit a Manuscript
- Subscription Information
- Table of Contents
BioMed Research International
Volume 2013 (2013), Article ID 481431, 14 pages
Optimized Periocular Template Selection for Human Recognition
Department of Computer Science and Engineering, National Institute of Technology Rourkela, Odisha 769008, India
Received 8 April 2013; Revised 30 June 2013; Accepted 7 July 2013
Academic Editor: Tatsuya Akutsu
Copyright © 2013 Sambit Bakshi et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
A novel approach for selecting a rectangular template around periocular region optimally potential for human recognition is proposed. A comparatively larger template of periocular image than the optimal one can be slightly more potent for recognition, but the larger template heavily slows down the biometric system by making feature extraction computationally intensive and increasing the database size. A smaller template, on the contrary, cannot yield desirable recognition though the smaller template performs faster due to low computation for feature extraction. These two contradictory objectives (namely, (a) to minimize the size of periocular template and (b) to maximize the recognition through the template) are aimed to be optimized through the proposed research. This paper proposes four different approaches for dynamic optimal template selection from periocular region. The proposed methods are tested on publicly available unconstrained UBIRISv2 and FERET databases and satisfactory results have been achieved. Thus obtained template can be used for recognition of individuals in an organization and can be generalized to recognize every citizen of a nation.
A biometric system comprises a physical or behavioral trait of a person through which he or she can be recognized uniquely. Computer aided identification of a person through face biometric has grown its importance through the last decade and researchers have attempted to find unique facial nodal points. However, change of facial data with expression and age makes it challenging for recognition through face. A stringent necessity to identify a person on partial facial data has been felt in such scenario. There are forensic applications where antemortem information is a partial face. These motives led researchers to derive auxiliary biometric traits from facial image, namely, iris, ear, lip, and periocular region. Recognizing human through iris captured under near infrared (NIR) illumination and constrained scenario yields satisfactory recognition accuracy while recognition under visual spectrum (VS) and unconstrained scenario is relatively challenging. In particular, VS periocular image has been exploited to examine its uniqueness as there exists many nodal points. Classification and recognition through periocular region show significant accuracy, given the fact that periocular biometric uses only approximately 10% of a complete face data (illustrated in Section 4.1). Figure 1 illustrates the working model of a biometric system that employs region around the eye (periocular region) as a trait for recognition. Face is one of the primitive means of human recognition.
Periocular (peripheral area of ocular) region refers to the immediate vicinity of the eye, including eyebrow and lower eye fold as depicted in Figure 2. Face recognition has been main attention of biometric researchers due to its ease of unconstrained acquisition and the uniqueness. Face is proven to have approximately 18 feature points  which can comprise in formation of a unique template for authentication. The major challenges in face detection faced by the researchers are due to change of human face with age, expression, and so forth. With the advent of low-cost hardware to fuse multiple biometrics in real time, the emphasis began to extract a subset of face which can partially resolve the aforementioned issues listed in Table 1. Hence the investigation towards ear, lip, and periocular has started gaining priority. Furthermore, capturing eye or face image automatically acquires periocular image. This gives the flexibility of recognizing an individual using the periocular data along with iris data without extra storage or acquisition cost. Moreover, periocular features can be used when an iris image does not contain subtle details, which mostly occurs due to poor image quality. Periocular biometric also comes into play as a candidate for fusion with face image for better recognition accuracy.
This paper approaches to fit an optimal boundary to the periocular region which is sufficient and necessary for recognition. Unlike other biometric traits, edge information is not the required criteria to exactly localize periocular region. Rather periocular region can be localized where the periphery of eye contains no further information. Researchers have considered a static rectangular boundary around the eye to recognize human and termed the localized rectangle as periocular region. However, this approach is naive as the same static boundary does not work for every face image (e.g., when the face image is captured through different distances from the camera, or when there is a tilt of face or camera during acquisition). So there is a need of deriving a dynamic boundary to describe periocular region. While deciding the periocular boundary, the objective of achieving the highest recognition accuracy also needs to be maintained. The paper specifies few metrics through which periocular region can be optimally localized in scale and rotation invariant manner.
The rest of the paper is organized as follows: Section 2 describes the landmark works in the direction of recognition and classification through periocular region and analyzes the need for optimizing the periocular region considered for recognition pointed in Section 3. In Section 4, four methods of template optimization are described and subsequently Section 5 records experimental results obtained to establish the proposed methods. Finally Section 6 concludes with describing the decided periocular template which is optimal for human recognition and marks its importance for recognition from a large database.
2. Literature Review
Investigations have been made by researchers in the direction of localizing iris from high quality constrained eye images captured in NIR illumination. Table 2 summarizes the comparative study of accuracy obtained by few benchmark iris localization technique. The results conclude that high localization accuracy has been achieved for NIR iris images. Several global and local matching techniques have been applied for matching NIR iris images and researchers have got high accuracy. However, when it comes to recognize a person only through his iris image captured under visible spectrum, the results have been observed to be unsatisfactory. So researchers have been motivated to take into account not only iris but also its peripheral regions while recognizing visible spectrum images.
The task of recognition is more challenging than classification and hence draws more attention. The most commonly used feature extraction techniques in context of periocular recognition are Scale Invariant Feature Transform, Local Binary Pattern. Tables 3 and 4 outline the methods used and performance obtained towards periocular classification and recognition in visual spectrum images, respectively. However, the portion of eye on which it is applied is not computationally justified in the literature. Any arbitrary rectangular portion centering the eye has been taken into account without questioning the following.(a)Will the accuracy obtained from this arbitrary boundary increase if a larger region is considered?(b)How much of the considered periocular region is actually contributing to recognition?(c)Is there any portion within this arbitrary considered periocular region which can be removed and still comparable accuracy can be achieved?
The derivation of optimal dynamic periocular region gives a simultaneous solution to the aforementioned questions.
3. Why Optimal Template for Periocular Region Is Required
Unlike other biometric traits, periocular region has no boundary defined by any edge information. Hence periocular region cannot be detected through differential change in pixel value in different directions. Rather the location of boundary is the region which is smooth in terms of pixel intensity, that is, a region with no information. The authors of  have localized the periocular region statically by taking a rectangle having dimension centering the iris where defines the radius of the iris. But this localization method fails when the eye is tilted or gaze is not frontal. Moreover, the method presumes the location of iris center to be accurately detectable. However, iris center cannot be detected for some eye images due to low-resolution nature of the image.
The objective of the paper is to attain a dynamic boundary around the eye that defines periocular region. The region hence derived should have the following properties: (a) should be able to recognize humans uniquely, (b) should be achievable for low-quality VS images, (c) should contain main identifiable features of eye region identifiable by a human being, and (d) no subset of the derived periocular region should be equally potent as the derived region for recognition.
The optimally selected periocular template can be a template to hold identity of an individual. If such template can be generated for the whole nation, it can serve as authorized identity (i.e., biometric passport ) of every citizen of the nation.
4. Proposed Periocular Template Selection Methods
To achieve the above stated properties, four different dynamic models are proposed through which periocular region can be segmented out. These models are based on (a) human anthropometry, (b) demand of the accuracy of biometric system, (c) human expert judgement, and (d) subdivision approach.
4.1. Through Human Anthropometry
In a given face image, face can be extracted out by neural training to the system or by fast color-segmentation methods. The color-segmentation methods detect skin region in the image and find the connected components in such a region. Depending on connected components having skin color, the system labels the component largest in size as face. Algorithm 1 proposes a binary component analysis based skin detection. The thresholds are experimentally fitted to obtain highest accuracy in segmenting skin region in face images comprising skin colors with different skin tones. The algorithm takes RGB face image as input. It first converts the face image to color space and normalizes the pixel values. In the next step, the average luminance value is calculated by summing up the component values of each pixel and dividing the total number of pixels in the image. A brightness compensated image is generated depending on the value of average luminance as specified in the algorithm. In the obtained brightness compensated image, compound condition is applied and a thresholding is performed to obtain the skin-map finally. Through connected component analysis of the skin map in color space, open eye region can be obtained as explained in Algorithm 2. The reason of segmenting open eye region is to obtain the nonskin region within detected face, which can be labeled as eye and thus to achieve approximate location of eye center.
Once the eye region is detected, the iris center can be obtained using conventional pupil detection and integrodifferential approach for finding the iris boundary and a static boundary can be fitted. As described earlier, the authors of  bounded periocular region with rectangle centering the iris center. But no justification is produced in the paper regarding the empirically taken height and width of this periocular boundary. This process of finding periocular boundary has prerequisite of knowledge of coordinates of iris center and radius of iris.
Anthropometric analysis  of human face and eye region gives the information regarding the ratio of eye and iris and ratio of width of face and eye. A typical block diagram in Figure 6 depicts the ratios of different parts of human face with respect to height or width of face. From the analysis, it is found that where denotes the distance between center of eyebrow and eye center:
This information can be used to decide the boundary of periocular region. In (1), width and height of eye are expressed as a function of the height and width of human face. Hence to gauge the width and height of periocular template boundary, there is no need to have knowledge of iris radius. However, knowledge of coordinates of iris center is necessary. From these information, a bounding box can be fit composing all visible portions of periocular region, for example, eyebrow, eyelashes, tear duct, eye fold, eye corner, and so forth. This approach is crude and dependent on the human supervision or intelligent detection of these nodal points in human eye.
Further, from (2), it is observable that either information of the height or width of periocular region is sufficient to derive the other parameter, provided that the aspect ratio of face is known. This aspect of the localization of periocular is used in Section 4.2. Equation (3) considers elliptical model to represent face while finding the ratio of periocular region and area of a human face. It justifies the usefulness of using an optimally selected periocular template for human recognition rather than a full face recognition system.
This method achieves periocular localization without knowledge of iris radius. Hence it is suitable for localization of periocular region for unconstrained images where iris radius is not detectable by machines due to low-quality, partial closure of eye, or luminance of the visible spectrum eye image.
However, to make the system work in more unconstrained environment, periocular boundary can be achieved through sclera detection, for the scenario when iris cannot be properly located due to unconstrained acquisition of eye or when the image captured is a low-quality color face image captured from a distance.
4.1.1. Detection of Sclera Region and Noise Removal
(1)The input RGB iris image is converted to grayscale image im_gray.(2)The input RGB iris image is converted to HSI color model where component of each pixel can be determined by where R, G, B denotes the Red, Green, and Blue color component of a particular pixel. Let the image hence formed containing S component of each pixel is .(3)If where is a predefined threshold, then that pixel is marked as sclera region, else as a nonsclera region. Authors in  have experimented with to get a binary map of sclera region through binarization of as follows: . Only a noisy binary map of sclera can be found through this process, in which white pixels denote noisy sclera region and black pixels denote non-sclera region.(4)im_bin is formed as follows: for every nonzero pixel in , for every zero pixel in , (5) is formed as follows: (6)All binary connected components present in are removed except the largest and second largest components.(7)If size of the second largest connected component is less than 25% of that of the large one, it is interpreted that the largest component is the single sclera detected and the second largest connected component is removed hence. Else both components are retained as binary map of sclera.
After processing these above specified steps, the binary image would only contain one or two components describing the sclera region, after removing noises.
4.1.2. Content Retrieval of Sclera Region
After a denoised binary map of sclera region within an eye image is obtained, it is necessary to retrieve the information about sclera, whether two parts of sclera on two sides of iris are separately visible, only one of them is detected, or both parts of sclera are detected as a single component.
There can be three exhaustive cases in the binary image found as sclera: (a) the two sides of the sclera is connected and found as a single connected component, (b) two sclera regions are found as two different connected components, and (c) only one side of the sclera is detected due to the pose of eye in the image. If the number of connected components is found to be two, then it is classified as aforementioned Case b (as shown in Figures 3(a), 3(b), and 3(c)) and two components are treated as two portions of sclera. Else, if a single connected component is obtained, it is checked for the ratio of length and breadth of the best fitted oriented bounding rectangle. If the ratio is greater than 1.25, then it belongs to aforementioned Case a, else belongs to Case c (shown in Figure 3(e)). For the aforementioned Case a, the region is subdivided into two components (through detecting minimal cut that divides the joined sclera into two parts) as shown in Figure 3(d) and further processing is performed.
4.1.3. Nodal Points Extraction from Sclera Region
Each sclera is subjected to following processing through which three nodal points are detected from each sclera region, namely (a) center of sclera, (b) center of concave region of sclera, and (c) eye corner. So in general cases where two parts of the sclera are detected, six nodal points will be detected. The method of nodal point extraction is illustrated below.(1)Finding Center of Sclera. The sclera component is subjected to a distance transform where the value of each white pixel (indicating pixels belonging to sclera) is replaced by its minimum distance from any black pixel. The pixel which is farthest from all black pixels will have highest value after this transformation. That pixel is labeled as center of sclera.(2)Finding Center of Concave Region of Sclera. The midpoints of every straight line joining any two border pixels of the detected sclera component are found out as shown in Figure 5. The midpoints lying on the component itself (shown by red point between and in Figure 5) are neglected. The midpoints lying outside the component (shown by yellow point between and in Figure 5) are taken into account. Due to discrete computation of straight lines, midpoints of many straight lines drawn in aforementioned way overlap on a single pixel. A separate matrix having the same size as the sclera itself is introduced, which is having zero value of each pixel initially. For every valid midpoint, the value of corresponding pixel in this new matrix is incremented. Once this process is over, more than one connected components of nonzero values will be obtained in the matrix signifying concave regions. The largest connected component is retained while others are removed. The pixel having maximum value in the largest component is labeled as the center of concave region.(3)Finding the Eye Corner. The distances of all pixels lying on boundary of sclera region from the sclera center are also calculated to find the center of sclera as described above. The boundary pixel which is farthest from the center of the sclera is labeled as the eye corner.
The result of extracting these nodal points from eye image helps in finding the tilt of eye along with the position of iris in eye. Figure 3 depicts five sample images from UBIRISv2 dataset and the outputs obtained from their processing through the aforementioned nodal point extraction technique. This information can be useful in localization of periocular region.
4.2. Through Demand of Accuracy of Biometric System
Beginning with the center of the eye (pupil center), a bounding rectangular box is taken of which only encloses the iris. Figure 4 shows how the eye images changes when it is cropped with pupil center and the bounding size is gradually increased. The corresponding accuracy of every cropped image is tested. In subsequent steps the coverage of this bounding box is increased with a width of 3% of the diameter of the iris and the change in accuracy is observed. After certain iterations of this procedure, the bounding box will come to a portion of periocular region where there is no more change in intensity; hence the region is low entropic. Hence no more local feature can be extracted from this region even if the bounding box is increased. In such scenario, the saturation accuracy is achieved, and on the basis of saturation accuracy, the corresponding minimum bounding box is considered as the desired periocular region. As the demand of different biometric systems may vary, the bounding box corresponding to certain predefined accuracy can also be segmented as periocular region. Similar results have also been observed for FERET database.
The exact method of obtaining the dynamic boundary is as follows.(1)For to 100, follow the steps 2 to 4.(2)For each image in database, find approximate iris location in eye image.(3)For each image in database, centering at the iris center, crop a bounding box whose width of diameter of iris, height of .(4)Find accuracy of the system with this image size.(5)Observe the change in accuracy with .
Figure 7 illustrates a plot of accuracy against which shows that the accuracy of the biometric system saturates after a particular size of the bounding box. Increasing the box further does not increase the accuracy. To carry out this experiment, Local Binary Pattern (LBP)  along with Scale Invariant Feature Transform (SIFT)  are employed as feature extractor from the eye images. First, LBP is applied and resulting image is subjected for extracting local feature through SIFT. In the process, a maximum accuracy of 85.64% is achieved while testing with randomly chosen 50 eye images of 12 subjects from UBIRISv2 dataset . When the same experiment is executed for randomly chosen 50 eye images of 12 subjects from FERET dataset , a maximum accuracy of 78.29% is achieved. These saturation accuracy values are obtained when a rectangular boundary of width 300% of diameter of iris is considered or a wider rectangular eye area is taken into consideration. To validate the experiment run on the sample strongly, the same experiment was conducted on complete UBIRISv2 and FERET dataset which yielded 85.43% and 78.01% accuracy, respectively. This concludes that a subset of a large database can be employed to find the optimal template size and the result found can be used on whole dataset for cropping of images. So to minimize template size without compromising in accuracy, the smallest wide rectangle with saturation accuracy can be used as localization boundary to periocular region. It is also observed that the region beyond 300% of diameter of iris, though does not participate in recognition, increases the matching time as shown in Figure 11. This is also another reason of removing the redundant eye region to make the recognition process fast.
To validate this experiment, the same experiment has been carried out once again on full database of UBIRISv2 and FERET. The obtained accuracy values as depicted in Figure 8 ensure the experimental objective that there is no significant feature in periocular region beyond 300% of diameter of iris which can contribute to recognition. The score distribution of imposter and genuine scores is shown in Figures 9 and 10.
4.3. Human Expert Judgement on Importance of Portions of Eye
Human expertise has been utilized to decide a sorted order of importance of different sections of periocular region towards recognition . This information can be used to detect only the most important section in human eye that is most important towards recognition. If that section is not found in human eye region, the captured image is marked as Failure to Acquire (FTA) and not used for recognition. Hence a predecision on the quality of live query template can increase the accuracy of the system by reducing false rejections. However, this technique is human-supervised while enrolling an image in the database and while a live query comes. The human expert has to verify whether the most important portion of eye is visible in the image and has to guide the biometric system accordingly.
4.4. Through Subdivision Approach and Automation of Human Expertise
During enrolment phase of a biometric system, a human expert needs to verify manually whether the captured image includes expected region of interest. Through automated labeling different sections of an eye, it can be stated which portion of eye is necessary for identification (from human expert knowledge already discussed) and an automated FTA detection system can be made. Hence there is no need of a human expert for verifying the existence of important portions of human eye in an acquired eye image.
The challenge in incorporating this strategy in localization of periocular region is the automatic detection of portions of human eye like eyelid, eye corner, tear duct, lower-eyefold, and so forth. An attempt to do subdivision detection in eye region can be achieved through color detection and analysis and applying different transformations.
5. Experimental Results
There are four methods explained through which an optimal periocular template can be selected for biometric recognition. The first two methods explained in Sections 4.1 and 4.2 are experimentally evaluated using publicly available FERET and UBIRISv2 databases. A brief description of the two databases used for evaluation are illustrated in Table 5. A total of genuine and imposter matching among images from UBIRISv2 and genuine and imposter matching among images from FERET database are experimented to claim the proposition of optimality.
Anthropometry based approach performs accurately along with proper skin detection and sclera detection in eye region. The sample outputs are shown in Figure 3 which are found to be proper when evaluated against ground truth.
Saturation accuracy based approach performs with an accuracy more than 80% with noisy and low-resolution images of UBIRISv2 and FERET, which marks the efficiency of the proposed approach. To analyse the performance more deeply, Receiver Operating Characteristic (ROC) curve is experimented out when the width of the periocular region is 200%, 250%, and 300% of the diameter of iris region, respectively. ROC curve depicts the dependence of false rejection rate (FRR) with false acceptance rate (FAR) for change in the value of threshold. The curve is plotted using linear, logarithmic, or semilogarithmic scales. As plotted in Figures 12 and 13, it is obvious to conclude that the system performs better with low FAR when than when and 250. Hence the ROC curve reveals that the portions of eye lying between 200% and 300% of diameter of iris are very much responsible for the recognition and feature-dense part of a periocular image. Furthermore to have a 1 : matching analysis, Cumulative Match Characteristic (CMC) curves representing the probability of identification at various ranks are also experimented out when the width of the periocular region is 200%, 250%, and 300% of the iris region, respectively (shown in Figures 14 and 15). The index  measures the separation between the arithmetic means of the genuine and imposter probability distribution in standard deviation units is defined as follows where and are mean and standard deviation of genuine and imposter scores. Table 6 yields the change of index of recognition when the width of periocular region is varied. The value of increases monotonically from 1.23 to 2.85 for UBIRISv2 dataset and from 1.19 to 2.69 for FERET dataset with incremental change in . An incremental nature in the values of for to 300 and an insignificant change in the value of for to 400 also establishes the existence of a boundary between regions contributing and not contributing to recognition.
Human expert judging is experimented by Hollingsworth et al.  and the results are used towards the direction of optimal periocular localization. Human subjects are asked which part of eye they feel to be the most important for recognition. Most of the subjects voted that blood vessels are the most important feature to recognize an individual from VS eye image. This information is used to infer which sub-portions of eye must belong to the optimal periocular region for it to be a candidate for recognition. Removal of those important regions will lead to rejection of the template.
Subdivision approach needs manual supervision in the process of proper labeling of the different portions of human eye. Once the enrolled templates are labeled by the expert, an optimal part of the template can be selected for recognition. The method is tested on FERET database and yielded proper localization of periocular region.
Recent research signifies why recognition through visual spectrum periocular image has gained so much importance and how the present approaches work. While developing recognition system for a large database, it is a crucial factor to optimize the template size. Existence of any redundant region in template will increase the matching time but will not contribute to increase the accuracy of matching. Hence removal of redundant region of the template should be accomplished before the matching procedure. As recognition time of identification is dependent on database size n, hence a decrease of 1 : 1 matching time of t will actually decrease nt matching time for identification in total. As n is large (in the range of 109 practical cases), nt is a significant amount of time, especially when concurrent matching is implemented in distributed biometric systems. The paper prescribes four metrics for the optimization of visual spectrum periocular image and experimentally establishes their relevance in terms of satisfying expected recognition accuracy. These methods can be used to localize the periocular region dynamically so that an optimized region can be selected which is best suitable for recognition in terms of two contradictory objectives: (a) minimal template size, and (b) maximal recognition accuracy.
|LBP:||Local Binary Pattern|
|SIFT:||Scale Invariant Feature Transform|
|ROC:||Receiver Operating Characteristic|
|CMC:||Cumulative Match Characteristic|
|FTA:||Failure to Acquire|
|FRR:||False rejection rate|
|FAR:||False acceptance rate.|
- A. Sohail and P. Bhattacharya, “Detection of facial feature points using anthropometric face model,” Signal Processing for Image Enhancement and Multimedia Processing, vol. 31, pp. 189–200, 2008.
- U. Park, R. R. Jillela, A. Ross, and A. K. Jain, “Periocular biometrics in the visible spectrum,” IEEE Transactions on Information Forensics and Security, vol. 6, no. 1, pp. 96–106, 2011.
- T. Camus and R. Wildes, “Reliable and fast eye finding in close-up images,” in Proceedings of the 16th International Conference on Pattern Recognition, vol. 1, pp. 389–394, 2002.
- H. Sung, J. Lim, J.-H. Park, and Y. Lee, “Iris recognition using collarette boundary localization,” in Proceedings of the 17th International Conference on Pattern Recognition (ICPR '04), vol. 4, pp. 857–860, August 2004.
- B. Bonney, R. Ives, D. Etter, and Y. Du, “IRIS pattern extraction using bit planes and standard deviations,” in Proceedings of the 38th Asilomar Conference on Signals, Systems and Computers, vol. 1, pp. 582–586, November 2004.
- X. Liu, K. W. Bowyer, and P. J. Flynn, “Experiments with an improved iris segmentation algorithm,” in Proceedings of the 4th IEEE Workshop on Automatic Identification Advanced Technologies (AUTO ID '05), pp. 118–123, October 2005.
- H. Proença and L. A. Alexandre, “Iris segmentation methodology for non-cooperative recognition,” IEE Proceedings: Vision, Image and Signal Processing, vol. 153, no. 2, pp. 199–205, 2006.
- S. J. Pundlik, D. L. Woodard, and S. T. Birchfield, “Non-ideal iris segmentation using graph cuts,” in Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops (CVPR '08), pp. 1–6, June 2008.
- Z. He, T. Tan, Z. Sun, and X. Qiu, “Toward accurate and fast iris segmentation for iris biometrics,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 31, no. 9, pp. 1670–1684, 2009.
- J. Liu, X. Fu, and H. Wang, “Iris image segmentation based on K-means cluster,” in Proceedings of the IEEE International Conference on Intelligent Computing and Intelligent Systems (ICIS '10), vol. 3, pp. 194–198, October 2010.
- F. Tan, Z. Li, and X. Zhu, “Iris localization algorithm based on gray distribution features,” in Proceedings of the 1st IEEE International Conference on Progress in Informatics and Computing (PIC '10), vol. 2, pp. 719–722, December 2010.
- S. Bakshi, H. Mehrotra, and B. Majhi, “Real-time iris segmentation based on image morphology,” in Proceedings of the International Conference on Communication, Computing and Security (ICCCS '11), pp. 335–338, February 2011.
- R. Abiantun and M. Savvides, “Tear-duct detector for identifying left versus right iris images,” in Proceedings of the 37th IEEE Applied Imagery Pattern Recognition Workshop (AIPR '08), pp. 1–4, October 2008.
- S. Bhat and M. Savvides, “Evaluating active shape models for eye-shape classification,” in Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP '08), pp. 5228–5231, April 2008.
- J. Merkow, B. Jou, and M. Savvides, “An exploration of gender identification using only the periocular region,” in Proceedings of the 4th IEEE International Conference on Biometrics: Theory, Applications and Systems (BTAS '10), September 2010.
- J. R. Lyle, P. E. Miller, S. J. Pundlik, and D. L. Woodard, “Soft biometric classification using periocular region features,” in Proceedings of the 4th IEEE International Conference on Biometrics: Theory, Applications and Systems (BTAS '10), September 2010.
- K. Hollingsworth, K. W. Bowyer, and P. J. Flynn, “Identifying useful features for recognition in near-infrared periocular images,” in Proceedings of the 4th IEEE International Conference on Biometrics: Theory, Applications and Systems (BTAS '10), September 2010.
- D. L. Woodard, S. Pundlik, P. Miller, R. Jillela, and A. Ross, “On the fusion of periocular and iris biometrics in non-ideal imagery,” in Proceedings of the 20th International Conference on Pattern Recognition (ICPR '10), pp. 201–204, August 2010.
- P. E. Miller, J. R. Lyle, S. J. Pundlik, and D. L. Woodard, “Performance evaluation of local appearance based periocular recognition,” in Proceedings of the 4th IEEE International Conference on Biometrics: Theory, Applications and Systems (BTAS '10), September 2010.
- P. E. Miller, A. W. Rawls, S. J. Pundlik, and D. L. Woodard, “Personal identification using periocular skin texture,” in Proceedings of the 25th Annual ACM Symposium on Applied Computing (SAC '10), pp. 1496–1500, March 2010.
- J. Adams, D. L. Woodard, G. Dozier, P. Miller, K. Bryant, and G. Glenn, “Genetic-based type II feature extraction for periocular biometric recognition: less is more,” in Proceedings of the 20th International Conference on Pattern Recognition (ICPR '10), pp. 205–208, August 2010.
- D. L. Woodard, S. J. Pundlik, P. E. Miller, and J. R. Lyle, “Appearance-based periocular features in the context of face and non-ideal iris recognition,” Signal, Image and Video Processing, vol. 5, no. 4, pp. 443–455, 2011.
- D. Malcik and M. Drahansky, “Anatomy of biometric passports,” Journal of Biomedicine and Biotechnology, vol. 2012, Article ID 490362, 8 pages, 2012.
- V. Ramanathan and H. Wechsler, “Robust human authentication using appearance and holistic anthropometric features,” Pattern Recognition Letters, vol. 31, no. 15, pp. 2425–2435, 2010.
- Y. Chen, M. Adjouadi, C. Han et al., “A highly accurate and computationally efficient approach for unconstrained iris segmentation,” Image and Vision Computing, vol. 28, no. 2, pp. 261–269, 2010.
- T. Ojala, M. Pietikäinen, and D. Harwood, “A comparative study of texture measures with classification based on feature distributions,” Pattern Recognition, vol. 29, no. 1, pp. 51–59, 1996.
- D. G. Lowe, “Distinctive image features from scale-invariant keypoints,” International Journal of Computer Vision, vol. 60, no. 2, pp. 91–110, 2004.
- H. Proença, S. Filipe, R. Santos, J. Oliveira, and L. A. Alexandre, “The UBIRIS.v2: a database of visible wavelength iris images captured on-the-move and at-a-distance,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 32, no. 8, pp. 1529–1535, 2010.
- P. Jonathon Phillips, H. Moon, S. A. Rizvi, and P. J. Rauss, “The FERET evaluation methodology for face-recognition algorithms,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 22, no. 10, pp. 1090–1104, 2000.
- H. Proença and L. A. Alexandre, “UBIRIS: a noisy iris image database,” in Proceedings of the 13th International Conference on Image Analysis and Processing, vol. 3617 of Lecture Notes in Computer Science, pp. 970–977, Springer, Cagliari, Italy, 2005.
- A. K. Jain, P. Flynn, and A. A. Ross, Handbook of Biometrics, Springer, New York, NY, USA, 2008.