Adaptive iris matching using database indexing

Information

  • Patent Grant
  • 8630464
  • Patent Number
    8,630,464
  • Date Filed
    Friday, June 11, 2010
    14 years ago
  • Date Issued
    Tuesday, January 14, 2014
    10 years ago
Abstract
An adaptive iris matching approach for processing images of irises having a quality not sufficient for conventional non-adaptive matching approaches. Visible regions in a radial direction on an iris, without segmenting a circumferential of the iris, may be processed. A boundary of the visible region of the iris may be estimated. An iris map may be constructed with the non-visible portion of the iris masked. The iris map may be at least partially encoded. Partial codes of the iris map may be extracted to index at least portions of a database containing iris information. Irises may be retrieved from the database with an iris code as a query.
Description
BACKGROUND

The present invention pertains to recognition systems and particularly to biometric recognition systems; in particular the invention pertains to iris recognition systems.


SUMMARY

The invention is an adaptive iris matching approach for processing images of irises having a quality not sufficient for conventional non-adaptive matching approaches. Visible regions in a radial direction on an iris, without segmenting a circumferential of the iris, may be processed. A boundary of a visible region of the iris may be estimated. An iris map may be constructed with the non-visible portion of the iris masked. The iris map may be partially encoded. Partial codes of the iris map may be extracted to index at least portions of a database containing iris information. Irises may be retrieved from the database with an iris code as a query.





BRIEF DESCRIPTION OF THE DRAWING


FIG. 1
a is a diagram of an iris showing various delineated boundaries;



FIG. 1
b is a diagram of a blurred iris image;



FIG. 1
c is a diagram of a gazed eye;



FIG. 2
a is a diagram of polar segmentation of an iris in an image;



FIG. 2
b is a diagram of processing blocks for an adaptive iris matching approach;



FIG. 2
c is a diagram of adaptive matching;



FIG. 3
a is a diagram of polar segmentation used to estimate a pupil bound;



FIG. 3
b is a diagram of model fitting using a pupil warp function to obtain a symmetric pupil shape and outside iris boundary;



FIG. 3
c is a diagram of an inverse projection or transform as applied to iris, pupil and eyelid boundaries; and



FIG. 4 is a diagram of an operation of an iris image processing approach.





DESCRIPTION

While much progress has been made toward recognition through iris analysis under small variations in pose and lighting, reliable techniques for recognition under more extreme non-cooperative environments appear to have proven elusive. Iris recognition signatures may perform well when the data acquisition parameters are relatively constrained and the acquired data quality is moderately high. However, when the data acquisition and quality constraints are relaxed, the match performance of iris recognition may suffer a considerable decline.


This may be due in part to the variability of the resulting non-ideal iris relative to the constrained signature, and the likelihood that there is insufficient iris exposure to adequately normalize—a procedure required by virtually all existing iris algorithms. Apart from estimating these irregularities, the segmentation routine should also detect reflections due to ambient light sources and occlusions due to eyelashes and eyelids. The challenges of processing such irises may mount mostly in the segmentation process or a loss of iris information due to gazing or heavy obscurations.


One may introduce an adaptive solution based on partial segmentation and database indexing. A key aspect of the present approach may be that it does not necessarily require the usual upfront normalization procedure in order to capture these local variations and thus there is no need to estimate the entire outer border of the iris prior to matching. It is important to note that such a procedure may be crucial to virtually all existing iris recognition techniques.


One may introduce a technical approach to solve the above challenge by initially parsing through the portion of the iris that is visible, and then adaptively extracting features from the rest of the iris based upon data indexing and model fitting approximation. Present solution may extend the POSE technique (see e.g., U.S. patent application Ser. No. 11/043,366, filed Jan. 26, 2005, and entitled “Iris Recognition System and Method”) having capabilities to adaptively extract features from the iris without the prior need to estimate the actual outer bound of the iris. It is arguable that it may suffice only to capture local sharp variations along the radial directions in the nonobscured areas of the iris. The key aspect of the present approach is that it does not require a normalization procedure in order to capture these local sharp variations. It is important to note that such a procedure may be crucial to virtually all existing iris recognition techniques. One may introduce an approach of an iris recognition technique that does not necessarily require normalization of the iris map.


The approach is based on the fact that local details of the iris are spread along the radial directions and thus virtually any visible iris region may be independently processed on the radial direction without estimating the circumferential of the iris.


The merit of the present approach is that it may allow users to enroll or match poor quality iris images that would be rejected by other states of the art technology. In addition, this approach may generate a list of possible matches instead of only the best match when the iris-print is too obscured to be a reliable matcher. Another merit of the present approach is that it may allow the iris recognition system to be more tolerable of noise and could potentially improve the matching process computational efficiency. The database may be easily indexed using a skin-print or eye-print to fewer manageable subsets of individuals.


One may solve the problem in three processing steps. 1) One may first segment only the visible region. Most of the fitting models may provide a reasonable estimate of the border segment to a first order estimation. 2) One then may extract the associated partial codes of the iris barcode to index a smaller population in a large population. 3) One then may recover the noisy information in the obscured area by adapting the appropriate fitting model that provides the best matching bits within the limited subset retrieved in step 1). The last processing step may be beneficial to denoise to some extent the obscured area as long as the iris information is not lost due to heavy gazing or obscurations.


In step 3, alternatively, one may capture the iris dyadic variations by formulating a dynamic encoder that takes into account the varying edge points. Due to the dyadic nature of wavelet functions, one may modify the encoding scheme to make use of a dynamic wavelet function that varies as a function of the angular variable. Unlike the standard wavelet transform, the wavelength scalar is not necessarily a constant but rather may be a smooth continuous function as a function of the angle variable. The output of the wavelet filter may thus be evaluated at each angle using virtually all possible values for the best match. The concatenated matched bits may determine the overall match score between a probe and query barcode.


The initialization step for the adaptive iris model may be based on the perspective projection of an ideal frontal pupil image (i.e., circular shape) into the actual image plane. Since the pupil edges are visible and easy to estimate, one may compute the warping projection function. One may use the inverse transform of the projection on both the pupil and iris edges to estimate the iris boundaries.


One may introduce a technical approach to solve the above problem by initially parsing through a portion of the iris that is visible, and then adaptively extracting features from the rest of the iris based upon data indexing and model fitting approximation. The solution may extend the POSE technique capabilities to adaptively extract features from the iris without the prior need to estimate the actual outer bound of the iris. Arguably, it may suffice only to capture local sharp variations along the radial directions in the non-obscured areas of the iris. A key aspect of the present approach is that it does not necessarily require a normalization procedure in order to capture the local sharp variations. It is important to note that such a procedure is crucial to virtually all existing iris recognition techniques. To one's best knowledge, one may introduce the first approach of an iris recognition technique that does not require normalization of the iris map.


A merit of the present approach is that it may allow users to enroll or match poor quality iris images that would be rejected by the current state of the art technology. In addition, this approach may generate a list of possible matches instead of only the best match when iris-print is too obscured to be a reliable matcher. Another merit of the present approach is that it may allow the iris recognition system to be more tolerable of noise and could potentially improve the matching process computational efficiency. The database may be easily indexed using skin-print or eye-print to fewer manageable subsets of individuals.


The original foundation of the iris technology and algorithms do not appear to address issues such as side looking eyes, self occlusion, non-frontal face images, and so forth. Recognition of this situation has led to solutions to address the real-time operational requirements of a standoff iris recognition system. The present approach is an efficient and robust one-dimensional (1D) fast segmentation approach built around the present approach for the Polar Segmentation (POSE) system (see e.g., U.S. patent application Ser. No. 11/043,366, filed Jan. 26, 2005, and entitled “Iris Recognition System and Method”).


There may be great promise for iris technology with a false reject rate between 0.01-0.03 at a false acceptance rate of 10-3. The false non-matching rate (FNMR) versus false matching rate (FMR) may be better even on larger datasets. The uniqueness and richness of iris patterns when deployed to match billions of subjects may be a requirement that should be met to deploy the technology for things such as border control and the global war on terror.


A challenge to solve may be presented. Current iris recognition solutions do not necessarily reflect the full potential of the technology. Iris recognition signatures may perform well when the data acquisition parameters are relatively constrained and the acquired data quality is moderately high. However, when the data acquisition and quality constraints are relaxed, the match performance of iris recognition may suffer considerably. Very few developed techniques appear to address the true nature of iris irregularity in iris segmentation. This may be due in part to the variability of the iris-print in unconstrained scenarios and less than optimal acquisition conditions, as illustrated by the examples shown in FIGS. 1a, 1b and 1c. Apart from estimating these irregularities, the segmentation routine may also detect reflections due to ambient light sources and occlusions due to eyelashes and eyelids. The difficulties of processing such irises may reside mostly in the segmentation process, loss of spatial frequency content (e.g., blur side effect), or loss of iris information due to gazing or heavy obscurations.


As shown in FIGS. 1a, 1b and 1c, the pupil edges and the eyelids (highlighted at items 11 and 12) may be estimated. The iris borders, on the other hand, may yield multiple fitting models 1, 2 and 3, as noted with numerals 13, 14 and 15, respectively. The uncertainty in the iris segmentation may result in assigning bit codes to the wrong iris pixels during feature extraction, and therefore drastically affecting the matching outcome. Similarly, the blurry low-quality image 16 in FIG. 1b appears to have no high spatial frequency content of the actual iris pattern. The strong gazed eye 17 in FIG. 1c appears to have led to a compression of the iris information that cannot be recovered. An image acceptable to current commercial iris technology may generally require highly constrained conditions such as little to no gaze, high contrast, and rich spatial frequency content. Thus, these illustrative examples may represent a challenge for iris recognition technologies.



FIGS. 1
a, 1b and 1c, from left to right, show examples of challenging cases of iris images: (1a) obscured eye 18 (1b) blurred eye 16 (boxes 19 and 21 indicate areas with key holistic ocular features), (1c) strong gazing 17 (patch(es) 22 indicate discriminating skin textures). No existing iris approach appears to address these types of problems simultaneously, as many seem to rely solely on visible iris-prints and constrained environments.


The present approach is an iris recognition system implementing image quality metrics as a first step of the process to assess the quality of the acquired eye image for reliable operation, and then apply the appropriate iris processing. Images with high quality may be processed using the POSE technique or virtually any third party reliable iris recognition system.


If the iris image does meet a subset of quality metrics (e.g., not blur but obscured images), then the eye image may be passed to the adaptive iris matching process described below. If the iris image does not meet a second set of quality metrics and cannot be rehabilitated, then the eye image may be rejected and a new eye image capture be sought by the eye-finder.


The present technical approach for iris quality measurement is described in U.S. patent application Ser. No. 11/681,614, filed Mar. 2, 2007, and entitled “Iris Recognition System Having Image Quality Metrics”.


An adaptive iris matching technique (AIM) may be noted. Very few techniques appear to have been developed to address the true nature of iris irregularity in standoff iris segmentation. Apart from estimating these irregularities, the segmentation routine should also detect reflections due to ambient light sources and occlusions due to eyelashes, eyelids and gazing.


The present approach may be based on the fact that local details of the iris are spread along the radial directions, and thus virtually any visible iris region may be independently processed on the radial direction without segmenting the circumferential of the iris. It may suffice only to capture local variations along the radial directions in the non-obscured areas of the iris for inferring an iris biometrics. A key aspect of the present approach is that it does not necessarily require the usual upfront normalization procedure in order to capture these local variations; thus, there appears no need to estimate the entire outer border of the iris prior to matching. Such a procedure may be crucial to virtually all existing iris recognition techniques, and use the segmentation functionalities available in the POSE system. The present approach is illustrated in FIGS. 2a, 2b and 2c. FIG. 2a is a diagram of POSE segmentation. FIG. 2b shows AIM processing blocks partial encoding 25, database indexing 26 and adaptive coding 27. The blocks effectively show the following: (1) segment only visible iris regions; (2) index to a manageable database; and (3) search for a best matching model using adaptive encoding. FIG. 2c is a diagram of adaptive matching.


The present approach may solve the issue in three processing steps.


1) Partial encoding 25 may be noted. One may first segment only the visible region (e.g., the iris segment highlighted at 23 in FIG. 1a, as delineated by lid borders 12). The fitting model 15 in FIG. 1a may provide a reasonable first order estimation of the border segment. Since one may estimate the iris boundaries for the visible iris segment 23, one can construct the iris map using a “rubber sheet” model while masking the rest of the iris. One may then proceed to encode the phasor information of the extracted iris map using Log-Gabor or Wavelet filters so that comparisons between templates can be made at least partially.


2) Database indexing 26 may be noted. One may extract the associated partial codes of the iris barcode to index a smaller population in the large database. Verification of an iris barcode against large biometrics databases may be daunting, especially when one has only partial code for an iris. To retrieve irises from a large database, one may use the extracted iris code as a query and compute two feature vectors for the iris query (phasor based clustering and angular clustering). Organization of feature data may be completed hierarchically to compare one feature at a time, starting with the extracted feature elements that are closest to the inner iris boundary and going outward. Current linear database indexing does not necessarily leverage the data structure of the iris barcode. In one approach, one may use the KD-Tree search approach (e.g., see J. L. Bentley, “Multidimensional Binary Search Trees Used for Associative Searching,” Comm. of the ACM, p. 18(9), 1975). Nearly any other alternative search process may be applied to this process step. A primary benefit of the KD-Tree algorithm is that it may enable the use of spatial structure within the data to limit the cost of the search. In one approach, one may use structures from multiple features, and their indexing to increase the efficiency of the search.


Irises may be retrieved from a large collection of databases using a visible iris barcode as a query and then finding the similarity measure for all the features of the iris codes in the database. A search process may be noted in U.S. patent application Ser. No. 11/681,751, filed Mar. 2, 2007, and entitled “Indexing and Database Search System”.


3) Adaptive Encoding 27 may be noted. Appropriate segmentation of the rest of the iris may appear critical to the success of the iris recognition. Since data that is rescaled (e.g., replacing Model 1 (13) with Model 3 (15) in FIG. 1a) iris patterns may corrupt the iris code and the generated template, resulting in poor recognition rates, especially if the visible region is much smaller then the obscured regions. One may linearly adapt the fitting model that provides the best matching bits within the limited subset retrieved in Step 1 (25). The iris pattern deformation due to gazing could generate lateral pressure on some of the tissue that may cause small parts of the iris tissue to fold and change in patterns and some of which might be obscured by the eyelids. The approach may be to adaptively normalize the iris to adjust linearly the rubber sheet iris map to fit fewer identified models, as shown in FIG. 1a, and to align the map to the varying edge points of the model while searching for the minimum possible matching distance. The search may be executed only on the subset of subjects indexed at Step 2 (26) of the process. It is important to note that the map calibration may be applied not necessarily uniformly but rather linearly on the radial direction to fit the varying edge points of the matching model.


Alternatively, in another approach, in Step 3 (27) one may capture the iris dyadic variations by formulating a wavelet based encoder that takes into account the varying edge points. Due to the dyadic nature of wavelet functions, one may modify the encoding scheme to make use of a dynamic wavelet function that varies as a function of the angular variable,












W
α



(
R
)


=



[

α


(
θ
)


]



-
1

/
2







sig


(
r
)





ψ
*



(


r
-
R


α


(
θ
)



)





r





,




(
1
)








where R is the translation parameter and α is the scale parameter. The function ψ(·) may be the mother wavelet. Unlike the standard wavelet transform, the wavelength scale α(θ) may be a smooth continuous function of the angle. The concatenated matched bits may determine the overall match score between a probe and query barcode.


The initialization step for the adaptive iris model may be based on the perspective projection of an ideal frontal pupil image (i.e., circular shape) into the actual image plane. Since the pupil edges are visible and easy to estimate, one may compute the warping projection function. One may use the inverse transform of the projection on both the pupil and iris edges to estimate the iris boundaries as shown in FIGS. 3a, 3b and 3c. Polar segmentation may be used to estimate a pupil bound 31 in FIG. 3a. FIG. 3b shows a model fitting using pupil warp function to obtain a symmetric pupil shape 32 and outside iris boundary 33. The inverse projection or transform is applied to the iris 34, pupil 35 and eyelid 36 boundaries as shown in FIG. 3c. In the preprocessor approach, the iris signature is directly extracted from the reference warped image. By incorporating the warping projection in the adaptive iris matching algorithm, one may report an improvement of the matching accuracy at fixed equal error rate.


The following numbered items are information of the present system. 1) A standoff iris recognition system may have an iris image source, a quality processing module connected to the image source, where the quality processing module is for providing a quality assessment of an iris image.


2) In the system of item 1, the assessment of an iris image may be provided according to one or more of the following image conditions which include blur, obscuration, standoff angle, iris visibility, and so forth.


3) The system of item 2, may further have an evaluator connected to the quality measure, where the evaluator is for receiving a quality assessment of an iris image from the quality processing module, and for indicating whether the iris image is high quality, acceptable (can be processed using adaptive iris matching) or unacceptable for further processing, and the evaluator is for indicating whether an iris image that is unacceptable should be rejected or be rehabilitated for further processing, and the evaluator is for indicating whether an iris image that is acceptable should be processed using the adaptive iris matching process.


4) In the system of item 3, further processing may include segmentation of the iris image using adaptive process having partial encoding, database indexing, and adaptive matching.


5) The system of item 3, where the system medium that provides instructions of item 4 that, when executed by a machine, may further cause the machine to perform operations including normalizing a number of data points from the iris portion along the visible radial at an angle segment.


6) The recognition system that provides instructions of item 5 that, when executed by a machine, may further cause the machine to perform operations to encode the partial visible iris map signature.


7) The system of item 1 may have a approach of indexing: providing a database of templates; grouping the database into a plurality of sub-databases, where: each template in the database has a first number of bits (based on angular indexing), each sub-database of the plurality of sub-databases represents virtually all templates in the database, and each template in a sub-database has a second number of bits (hierarchical structure item, e.g., KD Tree approach).


8) The system of item 7 may further have providing a subset of possible template barcodes for matching, and selecting a second number of bits of a barcode that corresponds to the second number of bits of identified templates.


9) In the system of item 8, providing instructions of item 6 that, when executed by a machine, may further cause the machine to perform operations having normalizing a number of data points from the obscured iris portion to match one of the identified templates (Adaptive Encoding item: The best matching bits within the limited subset retrieved by the database indexing).


10) The recognition system that provides instructions of item 9 that, when executed by a machine, may further cause the machine to perform operations to encode the partial obscured iris map signature. (I.e., it may adaptively normalize the iris to adjust linearly the rubber sheet iris map to fit fewer identified models.)


11) The system of item 1 that provides instructions of item 10 that, when executed by a machine, may further cause the machine to perform operations having comparing the iris barcodes to a previously generated reference iris signature, indicated by item 8. (It may be an item on aligning the map to the varying edge points of the model while searching for the minimum possible matching distance).


The present standoff iris recognition approach using adaptive iris matching (AIM) may tackle the challenging case of iris recognition from images where irises are obscured but partially visible. One may provide an approach that may adaptively extract features from the iris without requiring the actual segmentation of the outer bound of the iris. It may suffice only to capture local sharp variations along the radial directions in the non-obscured areas of the iris. An aspect of the present approach may be that it does not require a normalization procedure in order to capture these local sharp variations. Such a procedure may be crucial to iris recognition techniques. One may introduce an approach of an iris recognition technique that does not require normalization of the iris map.



FIG. 4 is a diagram of an operation of the present approach. A quality evaluation at symbol 42 may be made of an iris image from an iris image acquisition module at symbol 41. If the quality of the image is deemed to be “high” (by reaching a predefined iris quality measure adequate for an iris to be processed using iris signature), the image may be processed using a non-adaptive iris matching approach at symbol 43. Such an approach may be POSE noted herein, or some commercial approach. If the quality of image is deemed to be “low” (does not reach a predefined iris quality level that allows recognition of an identity using iris signature), a second quality evaluation of the image may be made at symbol 44. A result of the second evaluation may be “acceptable low”, such as there is enough percentage of the iris region visible that can be processed, which indicates that the image may be processed by the present adaptive iris matching approach 58 starting at symbol 45. This image quality would not be acceptable to the non-adaptive iris matching approach at symbol 43. An alternative result of the second evaluation at symbol 44 may be “very low”. Here, the image may be rehabilitated at symbol 46. Because of something like a glitch or other issue with the image which could be cleared up, a result may be a “rehabilitated acceptable low” which means that the image may go to the adaptive iris matching approach at symbol 45. The result of rehabilitation at symbol 46 could be a “rehabilitated high” which means that the image could go to the non-adaptive iris matching approach at symbol 43. Or the image may be deemed as an “unrehabilitatable very low” which means that the iris image would be rejected at symbol 47.


At symbol 45, the approach for adaptive iris matching may begin. The following is a summary of the approach for illustrative purposes but could incorporate additional details. There may be a processing of virtually any visible iris region on a radial direction at symbol 48, without segmenting a circumferential of the iris of the image. The iris boundary may be estimated for the visible iris region at symbol 49. An iris map may be constructed using a rubber sheet model at symbol 51. Rest of the iris map may be masked. Phasor information of the iris map may be encoded for at least partial comparisons between templates, as indicated at symbol 52. Associated partial codes of the iris code may be extracted to index a portion of a database at symbol 53. Irises, i.e., images of them, may be retrieved from the database with the iris code as a query and determine a similarity measure for features of iris codes in the database at symbol 54. At symbol 55, the remainder of the iris may be segmented. The iris may be adaptively normalized to linearly adjust the rubber sheet iris map to fit fewer identified models (i.e., narrow the search) at symbol 56. The search may be executed just on the indexed portion of the database at symbol 57.


Relevant applications may include U.S. patent application Ser. No. 11/675,424, filed Feb. 15, 2007, and entitled “Standoff Iris Recognition System”; U.S. patent application Ser. No. 11/372,854, filed Mar. 10, 2006, and entitled “Invariant Radial Iris Segmentation”; U.S. patent application Ser. No. 11/043,366, filed Jan. 26, 2005, and entitled “Iris Recognition System and Method”; U.S. patent application Ser. No. 11/672,108, filed Feb. 7, 2007, and entitled “Approaches and Apparatus for Eye Detection in a Digital Image”; U.S. patent application Ser. No. 11/681,614, filed Mar. 2, 2007, and entitled “Iris Recognition System Having Image Quality Metrics”; U.S. patent application Ser. No. 11/681,751, filed Mar. 2, 2007, and entitled “Indexing and Database Search System”; U.S. patent application Ser. No. 11/681,662, filed Mar. 2, 2007, and entitled “Expedient Encoding System; U.S. patent application Ser. No. 10/979,129, filed Nov. 3, 2004, entitled “System and Method for Gate Access Control”; U.S. patent application Ser. No. 10/655,124, filed Sep. 5, 2003, and entitled “System and Method for Dynamic Stand-Off Biometric Verification” (issued as U.S. Pat. No. 7,183,895); and U.S. Provisional Patent Application 60/778,770, filed Mar. 3, 2006, and entitled “Stand-Off Iris Detection Tracking and Recognition System”; all of which are hereby incorporated by reference.


In the present specification, some of the matter may be of a hypothetical or prophetic nature although stated in another manner or tense.


Although the present system has been described with respect to at least one illustrative example, many variations and modifications will become apparent to those skilled in the art upon reading the specification. It is therefore the intention that the appended claims be interpreted as broadly as possible in view of the prior art to include all such variations and modifications.

Claims
  • 1. A standoff iris recognition system comprising: an iris image acquisition module;an image quality evaluator connected to the iris image acquisition module; andan adaptive iris matching module connected to the image quality evaluator and the eye image acquisition module; andwherein the image quality evaluator determines if an eye image from the iris image acquisition module has a quality which is too low for a non-adaptive iris matching module and is sufficiently high enough for the adaptive iris matching module; andwherein if the quality is high enough for the adaptive iris matching module, the adaptive iris matching module executes a search and generates a list of possible matches without segmenting a circumferential of an iris of the eye image.
  • 2. The system of claim 1, wherein the non-adaptive iris matching module processes an iris image having sufficient quality so that a best match for the iris image can be generated from a database of iris images.
  • 3. The system of claim 2, wherein the adaptive iris matching module processes an iris image having low but sufficient quality so that a group of most probable matches for the iris image instead of one best match can be generated from a database of iris images.
  • 4. The system of claim 3, wherein the image quality evaluator determines whether an iris image has sufficient quality for successful processing by the non-adaptive iris matching module or not sufficient quality for successful processing by the non-adaptive iris matching module but has sufficient quality for processing by the adaptive iris matching module, or not sufficient quality for processing by the non-adaptive iris matching module or by the adaptive iris matching module.
  • 5. The system of claim 4, wherein: if an iris image has not sufficient quality for processing by the non-adaptive iris matching module or by the adaptive iris matching module, then the image is rehabilitated, if possible, for processing by the non-adaptive iris matching module or by the adaptive iris matching module; andif the iris image cannot be rehabilitated, then the iris image is rejected.
  • 6. The system of claim 1, wherein the image quality evaluator determines a quality of an iris image relative to blur, obscuration, standoff angle, and/or iris visibility of the iris image.
  • 7. The system of claim 1, wherein the adaptive iris matching module is for segmenting a visible portion of an iris image, partial encoding, database indexing, and/or adaptive matching relative to the visible portion of the iris image.
  • 8. The system of claim 7, wherein the adaptive iris matching module is further for normalizing data points of an iris portion along a visible radial at an angle segment of an iris image.
  • 9. The system of claim 7, wherein the adaptive iris matching module is further for encoding a partial visible iris map signature of an iris image.
  • 10. The system of claim 7, wherein the database indexing comprises: providing a database having templates; andgrouping the database into a plurality of sub-databases; andwherein:each template in the database has a first number of bits based on angular indexing;each sub-database of the plurality of sub-databases represents virtually all templates in the database; and/oreach template in a sub-database has a second number of bits based on a hierarchical structure.
  • 11. The system of claim 10, wherein the database indexing further comprises: providing a subset of template barcodes for possible matching; and/orselecting a second number of bits of a barcode that corresponds to the second number of bits of each template.
  • 12. The system of claim 11, wherein the adaptive iris matching module is further for normalizing a number of data points from an obscured iris portion of an iris image to match one of the templates.
  • 13. The system of claim 12, wherein the adaptive iris matching module is further for encoding a partial obscured iris map signature.
  • 14. The system of claim 13, wherein the adaptive iris matching module is further for comparing iris barcodes to a previously generated reference iris signature.
  • 15. A method for standoff iris recognition, comprising: segmenting a visible region of an iris, with a model providing a first order estimate of a border segment;extracting associated partial codes of a barcode of the iris to index a smaller population of a whole population; andrecovering noisy information in an obscured area by adapting a fitting model that provides best matching bits within the visible region of the iris;wherein:iris dyadic variations are captured by formulating a dynamic encoder that takes into account varying edge points; andan encoding scheme is modified to make use of a dynamic wavelet function wherein a wavelength scalar of the wavelet function varies as a function of an angular variable.
  • 16. The method of claim 15, wherein: the dynamic wavelet function is a smooth continuous function of the angular variable; andan output of the dynamic wavelet function is enabled at each angle using virtually all possible values for a best match.
  • 17. The method of claim 16, wherein concatenated matched bits can determine an overall match score between a probe and a query barcode.
  • 18. A method for adaptive iris matching, comprising: acquiring an image of an iris;evaluating quality of the image;processing a visible region of the iris in the image on a radial direction without segmenting a circumferential of the iris in the image, if the quality is too low for non-adaptive iris matching but high enough for adaptive iris matching;estimating a boundary for the visible region of the iris in the image;constructing an iris map using a rubber sheet model;masking a non-visible region of the iris;encoding the iris map for at least partial comparisons between templates;extracting associated partial codes of an iris code to index a portion of a database; andretrieving a list of two or more possible iris matches from the database with the iris code as a query.
Parent Case Info

This application claims the benefit of U.S. Provisional Patent Application No. 61/268,678, filed Jun. 15, 2009, and entitled “Adaptive Iris Matching Using Database Indexing”. U.S. Provisional Patent Application No. 61/268,678, filed Jun. 15, 2009, is hereby incorporated by reference.

US Referenced Citations (402)
Number Name Date Kind
4641349 Flom et al. Feb 1987 A
4836670 Hutchinson Jun 1989 A
5231674 Cleveland et al. Jul 1993 A
5291560 Daugman Mar 1994 A
5293427 Ueno et al. Mar 1994 A
5359382 Uenaka Oct 1994 A
5404013 Tajima Apr 1995 A
5543887 Akashi Aug 1996 A
5551027 Choy et al. Aug 1996 A
5572596 Wildes et al. Nov 1996 A
5608472 Szirth et al. Mar 1997 A
5664239 Nakata Sep 1997 A
5671447 Tokunaga Sep 1997 A
5687031 Ishihara Nov 1997 A
5717512 Chmielewski, Jr. et al. Feb 1998 A
5751836 Wildes et al. May 1998 A
5859686 Aboutalib et al. Jan 1999 A
5860032 Iwane Jan 1999 A
5896174 Nakata Apr 1999 A
5901238 Matsushita May 1999 A
5909269 Isogai et al. Jun 1999 A
5953440 Zhang et al. Sep 1999 A
5956122 Doster Sep 1999 A
5978494 Zhang Nov 1999 A
5991429 Coffin et al. Nov 1999 A
6005704 Chmielewski, Jr. et al. Dec 1999 A
6007202 Apple et al. Dec 1999 A
6012376 Hanke et al. Jan 2000 A
6021210 Camus et al. Feb 2000 A
6028949 McKendall Feb 2000 A
6055322 Salganicoff et al. Apr 2000 A
6064752 Rozmus et al. May 2000 A
6069967 Rozmus et al. May 2000 A
6081607 Mori et al. Jun 2000 A
6088470 Camus et al. Jul 2000 A
6091899 Konishi et al. Jul 2000 A
6101477 Hohle et al. Aug 2000 A
6104431 Inoue et al. Aug 2000 A
6108636 Yap et al. Aug 2000 A
6119096 Mann et al. Sep 2000 A
6120461 Smyth Sep 2000 A
6134339 Luo Oct 2000 A
6144754 Okano et al. Nov 2000 A
6246751 Bergl et al. Jun 2001 B1
6247813 Kim et al. Jun 2001 B1
6252977 Salganicoff et al. Jun 2001 B1
6259478 Hori Jul 2001 B1
6282475 Washington Aug 2001 B1
6285505 Melville et al. Sep 2001 B1
6285780 Yamakita et al. Sep 2001 B1
6289113 McHugh et al. Sep 2001 B1
6299306 Braithwaite et al. Oct 2001 B1
6308015 Matsumoto Oct 2001 B1
6309069 Seal et al. Oct 2001 B1
6320610 Van Sant et al. Nov 2001 B1
6320612 Young Nov 2001 B1
6320973 Suzaki et al. Nov 2001 B2
6323761 Son Nov 2001 B1
6325765 Hay et al. Dec 2001 B1
6330674 Angelo et al. Dec 2001 B1
6332193 Glass et al. Dec 2001 B1
6344683 Kim Feb 2002 B1
6370260 Pavlidis et al. Apr 2002 B1
6377699 Musgrave et al. Apr 2002 B1
6393136 Amir et al. May 2002 B1
6400835 Lemelson et al. Jun 2002 B1
6424727 Musgrave et al. Jul 2002 B1
6424845 Emmoft et al. Jul 2002 B1
6433818 Steinberg et al. Aug 2002 B1
6438752 McClard Aug 2002 B1
6441482 Foster Aug 2002 B1
6446045 Stone et al. Sep 2002 B1
6483930 Musgrave et al. Nov 2002 B1
6484936 Nicoll et al. Nov 2002 B1
6490443 Freeny, Jr. Dec 2002 B1
6493363 Roger et al. Dec 2002 B1
6493669 Curry et al. Dec 2002 B1
6494363 Roger et al. Dec 2002 B1
6503163 Van Sant et al. Jan 2003 B1
6505193 Musgrave et al. Jan 2003 B1
6506078 Mori et al. Jan 2003 B1
6508397 Do Jan 2003 B1
6516078 Yang et al. Feb 2003 B1
6516087 Camus Feb 2003 B1
6516416 Gregg et al. Feb 2003 B2
6522772 Morrison et al. Feb 2003 B1
6523165 Liu et al. Feb 2003 B2
6526160 Ito Feb 2003 B1
6532298 Cambier et al. Mar 2003 B1
6540392 Braithwaite Apr 2003 B1
6542624 Oda Apr 2003 B1
6546121 Oda Apr 2003 B1
6553494 Glass Apr 2003 B1
6580356 Alt et al. Jun 2003 B1
6591001 Oda et al. Jul 2003 B1
6591064 Higashiyama et al. Jul 2003 B2
6594377 Kim et al. Jul 2003 B1
6594399 Camus et al. Jul 2003 B1
6598971 Cleveland Jul 2003 B2
6600878 Pregara Jul 2003 B2
6614919 Suzaki et al. Sep 2003 B1
6652099 Chae et al. Nov 2003 B2
6674367 Sweatte Jan 2004 B2
6687389 McCartney et al. Feb 2004 B2
6690997 Rivalto Feb 2004 B2
6708176 Strunk et al. Mar 2004 B2
6709734 Higashi et al. Mar 2004 B2
6711562 Ross et al. Mar 2004 B1
6714665 Hanna et al. Mar 2004 B1
6718049 Pavlidis et al. Apr 2004 B2
6718050 Yamamoto Apr 2004 B1
6718665 Hess et al. Apr 2004 B2
6732278 Baird, III et al. May 2004 B2
6734783 Anbai May 2004 B1
6745520 Puskaric et al. Jun 2004 B2
6750435 Ford Jun 2004 B2
6751733 Nakamura et al. Jun 2004 B1
6753919 Daugman Jun 2004 B1
6754640 Bozeman Jun 2004 B2
6760467 Min et al. Jul 2004 B1
6765470 Shinzaki Jul 2004 B2
6766041 Golden et al. Jul 2004 B2
6775774 Harper Aug 2004 B1
6785406 Kamada Aug 2004 B1
6792134 Chen et al. Sep 2004 B2
6793134 Clark Sep 2004 B2
6819219 Bolle et al. Nov 2004 B1
6829370 Pavlidis et al. Dec 2004 B1
6832044 Doi et al. Dec 2004 B2
6836554 Bolle et al. Dec 2004 B1
6837436 Swartz et al. Jan 2005 B2
6845879 Park Jan 2005 B2
6853444 Haddad Feb 2005 B2
6867683 Calvesio et al. Mar 2005 B2
6873960 Wood et al. Mar 2005 B1
6896187 Stockhammer May 2005 B2
6905411 Nguyen et al. Jun 2005 B2
6920237 Chen et al. Jul 2005 B2
6930707 Bates et al. Aug 2005 B2
6934849 Kramer et al. Aug 2005 B2
6950139 Fujinawa Sep 2005 B2
6954738 Wang et al. Oct 2005 B2
6957341 Rice et al. Oct 2005 B2
6964666 Jackson Nov 2005 B2
6968457 Tam Nov 2005 B2
6972797 Izumi Dec 2005 B2
6992562 Fuks et al. Jan 2006 B2
6992717 Hatano Jan 2006 B2
7003669 Monk Feb 2006 B2
7017359 Kim et al. Mar 2006 B2
7030351 Wasserman et al. Apr 2006 B2
7031539 Tisse et al. Apr 2006 B2
7043056 Edwards et al. May 2006 B2
7053948 Konishi May 2006 B2
7058209 Chen et al. Jun 2006 B2
7071971 Elberbaum Jul 2006 B2
7076087 Wakiyama Jul 2006 B2
7084904 Liu et al. Aug 2006 B2
7092555 Lee et al. Aug 2006 B2
7095901 Lee et al. Aug 2006 B2
7100818 Swaine Sep 2006 B2
7113170 Lauper et al. Sep 2006 B2
7114080 Rahman et al. Sep 2006 B2
7120607 Bolle et al. Oct 2006 B2
7125335 Rowe Oct 2006 B2
7130452 Bolle et al. Oct 2006 B2
7130453 Kondo et al. Oct 2006 B2
7135980 Moore et al. Nov 2006 B2
7136581 Fujii Nov 2006 B2
7145457 Spitz et al. Dec 2006 B2
7146027 Kim et al. Dec 2006 B2
7152085 Tisse Dec 2006 B2
7155035 Kondo et al. Dec 2006 B2
7169052 Beaulieu Jan 2007 B2
7173348 Voda et al. Feb 2007 B2
7174036 Ohba Feb 2007 B2
7177449 Russon et al. Feb 2007 B2
7181049 Ike Feb 2007 B2
7183895 Bazakos et al. Feb 2007 B2
7184577 Chen et al. Feb 2007 B2
7187786 Kee Mar 2007 B2
7191936 Smith et al. Mar 2007 B2
7197166 Jeng Mar 2007 B2
7197173 Jones et al. Mar 2007 B2
7203343 Manasse et al. Apr 2007 B2
7204425 Mosher, Jr. et al. Apr 2007 B2
7206431 Schuessler Apr 2007 B2
7215797 Park May 2007 B2
7226164 Abourizk et al. Jun 2007 B2
7239726 Li Jul 2007 B2
7269737 Robinson Sep 2007 B2
7271839 Lee et al. Sep 2007 B2
7272380 Lee et al. Sep 2007 B2
7272385 Mirouze et al. Sep 2007 B2
7277561 Shin Oct 2007 B2
7277891 Howard et al. Oct 2007 B2
7280984 Phelan, III et al. Oct 2007 B2
7287021 De Smet Oct 2007 B2
7298873 Miller, Jr. et al. Nov 2007 B2
7298874 Cho Nov 2007 B2
7305089 Morikawa et al. Dec 2007 B2
7309126 Mihashi et al. Dec 2007 B2
7312818 Ooi et al. Dec 2007 B2
7313529 Thompson Dec 2007 B2
7315233 Yuhara Jan 2008 B2
7331667 Grotehusmann et al. Feb 2008 B2
7333637 Walfridsson Feb 2008 B2
7333798 Hodge Feb 2008 B2
7336806 Schonberg et al. Feb 2008 B2
7338167 Zelvin et al. Mar 2008 B2
7346195 Lauper et al. Mar 2008 B2
7346779 Leeper Mar 2008 B2
7353399 Ooi et al. Apr 2008 B2
7362210 Bazakos et al. Apr 2008 B2
7362370 Sakamoto et al. Apr 2008 B2
7362884 Willis et al. Apr 2008 B2
7365771 Kahn et al. Apr 2008 B2
7380938 Chmielewski, Jr. et al. Jun 2008 B2
7391865 Orsini et al. Jun 2008 B2
7404086 Sands et al. Jul 2008 B2
7406184 Wolff et al. Jul 2008 B2
7414648 Imada Aug 2008 B2
7417682 Kuwakino et al. Aug 2008 B2
7418115 Northcott et al. Aug 2008 B2
7421097 Hamza et al. Sep 2008 B2
7436986 Caldwell Oct 2008 B2
7443441 Hiraoka Oct 2008 B2
7447911 Chou et al. Nov 2008 B2
7460693 Loy et al. Dec 2008 B2
7466348 Morikawa et al. Dec 2008 B2
7467809 Breed et al. Dec 2008 B2
7471451 Dent et al. Dec 2008 B2
7472283 Angelo et al. Dec 2008 B2
7486806 Azuma et al. Feb 2009 B2
7506172 Bhakta Mar 2009 B2
7512254 Vollkommer et al. Mar 2009 B2
7518651 Butterworth Apr 2009 B2
7537568 Moehring May 2009 B2
7538326 Johnson et al. May 2009 B2
7542945 Thompson et al. Jun 2009 B2
7552333 Wheeler et al. Jun 2009 B2
7580620 Raskar et al. Aug 2009 B2
7593550 Hamza Sep 2009 B2
7639846 Yoda Dec 2009 B2
7722461 Gatto et al. May 2010 B2
7751598 Matey et al. Jul 2010 B2
7756301 Hamza Jul 2010 B2
7756407 Raskar Jul 2010 B2
7761453 Hamza Jul 2010 B2
7777802 Shinohara et al. Aug 2010 B2
7804982 Howard et al. Sep 2010 B2
20010026632 Tamai Oct 2001 A1
20010027116 Baird Oct 2001 A1
20010047479 Bromba et al. Nov 2001 A1
20010051924 Uberti Dec 2001 A1
20020010857 Karthik Jan 2002 A1
20020039433 Shin Apr 2002 A1
20020040434 Elliston et al. Apr 2002 A1
20020062280 Zachariassen et al. May 2002 A1
20020112177 Voltmer et al. Aug 2002 A1
20020142844 Kerr Oct 2002 A1
20020150281 Cho Oct 2002 A1
20020154794 Cho Oct 2002 A1
20020158750 Almalik Oct 2002 A1
20020175182 Matthews Nov 2002 A1
20020186131 Fettis Dec 2002 A1
20020191075 Doi et al. Dec 2002 A1
20020191076 Wada et al. Dec 2002 A1
20020194128 Maritzen et al. Dec 2002 A1
20020194131 Dick Dec 2002 A1
20020198731 Barnes et al. Dec 2002 A1
20030002714 Wakiyama Jan 2003 A1
20030012413 Kusakari et al. Jan 2003 A1
20030038173 Blackson et al. Feb 2003 A1
20030046228 Berney Mar 2003 A1
20030055689 Block et al. Mar 2003 A1
20030055787 Fujii Mar 2003 A1
20030065626 Allen Apr 2003 A1
20030071743 Seah et al. Apr 2003 A1
20030072475 Tamori Apr 2003 A1
20030073499 Reece Apr 2003 A1
20030074317 Hofi Apr 2003 A1
20030074326 Byers Apr 2003 A1
20030080194 O'Hara et al. May 2003 A1
20030092489 Veradej May 2003 A1
20030098776 Friedli May 2003 A1
20030099379 Monk et al. May 2003 A1
20030107097 McArthur et al. Jun 2003 A1
20030107645 Yoon Jun 2003 A1
20030115148 Takhar Jun 2003 A1
20030116630 Carey et al. Jun 2003 A1
20030118212 Min et al. Jun 2003 A1
20030125054 Garcia Jul 2003 A1
20030125057 Pesola Jul 2003 A1
20030126560 Kurapati et al. Jul 2003 A1
20030131245 Linderman Jul 2003 A1
20030133597 Moore et al. Jul 2003 A1
20030140235 Immega et al. Jul 2003 A1
20030140928 Bui et al. Jul 2003 A1
20030141411 Pandya et al. Jul 2003 A1
20030149881 Patel et al. Aug 2003 A1
20030152251 Ike Aug 2003 A1
20030156741 Lee et al. Aug 2003 A1
20030158762 Wu Aug 2003 A1
20030158821 Maia Aug 2003 A1
20030159051 Hollnagel Aug 2003 A1
20030163739 Armington et al. Aug 2003 A1
20030169334 Braithwaite et al. Sep 2003 A1
20030174049 Beigel et al. Sep 2003 A1
20030177051 Driscoll et al. Sep 2003 A1
20030182151 Taslitz Sep 2003 A1
20030182182 Kocher Sep 2003 A1
20030189480 Hamid Oct 2003 A1
20030189481 Hamid Oct 2003 A1
20030191949 Odagawa Oct 2003 A1
20030194112 Lee Oct 2003 A1
20030210139 Brooks et al. Nov 2003 A1
20030225711 Paping Dec 2003 A1
20030236120 Reece et al. Dec 2003 A1
20040002894 Kocher Jan 2004 A1
20040005078 Tillotson Jan 2004 A1
20040006553 de Vries et al. Jan 2004 A1
20040010462 Moon et al. Jan 2004 A1
20040025030 Corbett-Clark et al. Feb 2004 A1
20040025053 Hayward Feb 2004 A1
20040030930 Nomura Feb 2004 A1
20040037450 Bradski Feb 2004 A1
20040039914 Barr et al. Feb 2004 A1
20040042641 Jakubowski Mar 2004 A1
20040044627 Russell et al. Mar 2004 A1
20040046640 Jourdain et al. Mar 2004 A1
20040050924 Mletzko et al. Mar 2004 A1
20040050930 Rowe Mar 2004 A1
20040052405 Walfridsson Mar 2004 A1
20040052418 DeLean Mar 2004 A1
20040059590 Mercredi et al. Mar 2004 A1
20040059953 Purnell Mar 2004 A1
20040117636 Cheng Jun 2004 A1
20040133804 Smith et al. Jul 2004 A1
20040160518 Park Aug 2004 A1
20040162870 Matsuzaki et al. Aug 2004 A1
20040162984 Freeman et al. Aug 2004 A1
20040172541 Ando et al. Sep 2004 A1
20040193893 Braithwaite et al. Sep 2004 A1
20040233038 Beenau et al. Nov 2004 A1
20040252866 Tisse et al. Dec 2004 A1
20040255168 Murashita et al. Dec 2004 A1
20050008201 Lee et al. Jan 2005 A1
20050012817 Hampapur et al. Jan 2005 A1
20050029353 Isemura et al. Feb 2005 A1
20050052566 Kato Mar 2005 A1
20050055582 Bazakos et al. Mar 2005 A1
20050063567 Saitoh et al. Mar 2005 A1
20050084137 Kim et al. Apr 2005 A1
20050084179 Hanna et al. Apr 2005 A1
20050102502 Sagen May 2005 A1
20050125258 Yellin et al. Jun 2005 A1
20050129286 Hekimian Jun 2005 A1
20050138385 Friedli et al. Jun 2005 A1
20050138387 Lam et al. Jun 2005 A1
20050146640 Shibata Jul 2005 A1
20050151620 Neumann Jul 2005 A1
20050152583 Kondo et al. Jul 2005 A1
20050193212 Yuhara Sep 2005 A1
20050199708 Friedman Sep 2005 A1
20050206501 Farhat Sep 2005 A1
20050206502 Bernitz Sep 2005 A1
20050210267 Sugano et al. Sep 2005 A1
20050210270 Rohatgi et al. Sep 2005 A1
20050238214 Matsuda et al. Oct 2005 A1
20050240778 Saito Oct 2005 A1
20050248725 Ikoma et al. Nov 2005 A1
20050249385 Kondo et al. Nov 2005 A1
20050255840 Markham Nov 2005 A1
20060093190 Cheng et al. May 2006 A1
20060147094 Yoo Jul 2006 A1
20060165266 Hamza Jul 2006 A1
20060274919 LoIacono et al. Dec 2006 A1
20070036397 Hamza Feb 2007 A1
20070140531 Hamza Jun 2007 A1
20070160266 Jones et al. Jul 2007 A1
20070189582 Hamza et al. Aug 2007 A1
20070206840 Jacobson Sep 2007 A1
20070211924 Hamza Sep 2007 A1
20070274570 Hamza Nov 2007 A1
20070274571 Hamza Nov 2007 A1
20070286590 Terashima Dec 2007 A1
20080005578 Shafir Jan 2008 A1
20080075334 Determan et al. Mar 2008 A1
20080075441 Jelinek et al. Mar 2008 A1
20080075445 Whillock et al. Mar 2008 A1
20080104415 Palti-Wasserman et al. May 2008 A1
20080148030 Goffin Jun 2008 A1
20080211347 Wright et al. Sep 2008 A1
20080252412 Larsson et al. Oct 2008 A1
20090046899 Northcott et al. Feb 2009 A1
20090092283 Whillock et al. Apr 2009 A1
20090316993 Brasnett et al. Dec 2009 A1
20100033677 Jelinek Feb 2010 A1
20100034529 Jelinek Feb 2010 A1
20100182440 McCloskey Jul 2010 A1
20100239119 Bazakos et al. Sep 2010 A1
Foreign Referenced Citations (187)
Number Date Country
0484076 May 1992 EP
0593386 Apr 1994 EP
0878780 Nov 1998 EP
0899680 Mar 1999 EP
0910986 Apr 1999 EP
0962894 Dec 1999 EP
1018297 Jul 2000 EP
1024463 Aug 2000 EP
1028398 Aug 2000 EP
1041506 Oct 2000 EP
1041523 Oct 2000 EP
1126403 Aug 2001 EP
1139270 Oct 2001 EP
1237117 Sep 2002 EP
1477925 Nov 2004 EP
1635307 Mar 2006 EP
2369205 May 2002 GB
2371396 Jul 2002 GB
2375913 Nov 2002 GB
2402840 Dec 2004 GB
2411980 Sep 2005 GB
9161135 Jun 1997 JP
9198545 Jul 1997 JP
9201348 Aug 1997 JP
9147233 Sep 1997 JP
9234264 Sep 1997 JP
9305765 Nov 1997 JP
9319927 Dec 1997 JP
10021392 Jan 1998 JP
10040386 Feb 1998 JP
10049728 Feb 1998 JP
10137219 May 1998 JP
10137221 May 1998 JP
10137222 May 1998 JP
10137223 May 1998 JP
10248827 Sep 1998 JP
10269183 Oct 1998 JP
11047117 Feb 1999 JP
11089820 Apr 1999 JP
11200684 Jul 1999 JP
11203478 Jul 1999 JP
11213047 Aug 1999 JP
11339037 Dec 1999 JP
2000005149 Jan 2000 JP
2000005150 Jan 2000 JP
2000011163 Jan 2000 JP
2000023946 Jan 2000 JP
2000083930 Mar 2000 JP
2000102510 Apr 2000 JP
2000102524 Apr 2000 JP
2000105830 Apr 2000 JP
2000107156 Apr 2000 JP
2000139878 May 2000 JP
2000155863 Jun 2000 JP
2000182050 Jun 2000 JP
2000185031 Jul 2000 JP
2000194972 Jul 2000 JP
2000237167 Sep 2000 JP
2000242788 Sep 2000 JP
2000259817 Sep 2000 JP
2000356059 Dec 2000 JP
2000357232 Dec 2000 JP
2001005948 Jan 2001 JP
2001067399 Mar 2001 JP
2001101429 Apr 2001 JP
2001167275 Jun 2001 JP
2001222661 Aug 2001 JP
2001292981 Oct 2001 JP
2001297177 Oct 2001 JP
2001358987 Dec 2001 JP
2002119477 Apr 2002 JP
2002133415 May 2002 JP
2002153444 May 2002 JP
2002153445 May 2002 JP
2002260071 Sep 2002 JP
2002271689 Sep 2002 JP
2002286650 Oct 2002 JP
2002312772 Oct 2002 JP
2002329204 Nov 2002 JP
2003006628 Jan 2003 JP
2003036434 Feb 2003 JP
2003108720 Apr 2003 JP
2003108983 Apr 2003 JP
2003132355 May 2003 JP
2003150942 May 2003 JP
2003153880 May 2003 JP
2003242125 Aug 2003 JP
2003271565 Sep 2003 JP
2003271940 Sep 2003 JP
2003308522 Oct 2003 JP
2003308523 Oct 2003 JP
2003317102 Nov 2003 JP
2003331265 Nov 2003 JP
2004005167 Jan 2004 JP
2004021406 Jan 2004 JP
2004030334 Jan 2004 JP
2004038305 Feb 2004 JP
2004094575 Mar 2004 JP
2004152046 May 2004 JP
2004163356 Jun 2004 JP
2004164483 Jun 2004 JP
2004171350 Jun 2004 JP
2004171602 Jun 2004 JP
2004206444 Jul 2004 JP
2004220376 Aug 2004 JP
2004261515 Sep 2004 JP
2004280221 Oct 2004 JP
2004280547 Oct 2004 JP
2004287621 Oct 2004 JP
2004315127 Nov 2004 JP
2004318248 Nov 2004 JP
2005004524 Jan 2005 JP
2005011207 Jan 2005 JP
2005025577 Jan 2005 JP
2005038257 Feb 2005 JP
2005062990 Mar 2005 JP
2005115961 Apr 2005 JP
2005148883 Jun 2005 JP
2005242677 Sep 2005 JP
WO 9717674 May 1997 WO
WO 9721188 Jun 1997 WO
WO 9802083 Jan 1998 WO
WO 9808439 Mar 1998 WO
WO 9932317 Jul 1999 WO
WO 9952422 Oct 1999 WO
WO 9965175 Dec 1999 WO
WO 0028484 May 2000 WO
WO 0029986 May 2000 WO
WO 0031677 Jun 2000 WO
WO 0036605 Jun 2000 WO
WO 0062239 Oct 2000 WO
WO 0101329 Jan 2001 WO
WO 0103100 Jan 2001 WO
WO 0128476 Apr 2001 WO
WO 0135348 May 2001 WO
WO 0135349 May 2001 WO
WO 0140982 Jun 2001 WO
WO 0163994 Aug 2001 WO
WO 0169490 Sep 2001 WO
WO 0186599 Nov 2001 WO
WO 0201451 Jan 2002 WO
WO 0219030 Mar 2002 WO
WO 0235452 May 2002 WO
WO 0235480 May 2002 WO
WO 02091735 Nov 2002 WO
WO 02095657 Nov 2002 WO
WO 03002387 Jan 2003 WO
WO 03003910 Jan 2003 WO
WO 03054777 Jul 2003 WO
WO 03077077 Sep 2003 WO
WO 2004029863 Apr 2004 WO
WO 2004042646 May 2004 WO
WO 2004055737 Jul 2004 WO
WO 2004089214 Oct 2004 WO
WO 2004097743 Nov 2004 WO
WO 2005008567 Jan 2005 WO
WO 2005013181 Feb 2005 WO
WO 2005024698 Mar 2005 WO
WO 2005024708 Mar 2005 WO
WO 2005024709 Mar 2005 WO
WO 2005029388 Mar 2005 WO
WO 2005062235 Jul 2005 WO
WO 2005069252 Jul 2005 WO
WO 2005093510 Oct 2005 WO
WO 2005093681 Oct 2005 WO
WO 2005096962 Oct 2005 WO
WO 2005098531 Oct 2005 WO
WO 2005104704 Nov 2005 WO
WO 2005109344 Nov 2005 WO
WO 2006012645 Feb 2006 WO
WO 2006023046 Mar 2006 WO
WO 2006051462 May 2006 WO
WO 2006063076 Jun 2006 WO
WO 2006081209 Aug 2006 WO
WO 2006081505 Aug 2006 WO
WO 2007101269 Sep 2007 WO
WO 2007101275 Sep 2007 WO
WO 2007101276 Sep 2007 WO
WO 2007103698 Sep 2007 WO
WO 2007103701 Sep 2007 WO
WO 2007103834 Sep 2007 WO
WO 2008016724 Feb 2008 WO
WO 2008019168 Feb 2008 WO
WO 2008019169 Feb 2008 WO
WO 2008021584 Feb 2008 WO
WO 2008031089 Mar 2008 WO
WO 2008040026 Apr 2008 WO
Non-Patent Literature Citations (93)
Entry
Chun et al, “Iris Recognition for Palm-Top Application,” 2004, ICBA 2004, LNCS 3072, pp. 426-433.
Proenca et al, Iris Recognition: Measuring Feature's Quality for the Feature Selection in Unconstrained Image Capture Environments, 2006, IEEE Proceedings of the 2006 International Conference on Computational Intelligence for Homeland Security and Personal Safety—CIHSPS 2006, vol. 1, pp. 1-5.
U.S. Appl. No. 13/077,821, filed Mar. 30, 2011.
Freeboy, “Adaptive Optics Speeds Up Airport Immigration,” Optics.org/ole, 2 pages, Jan. 2009.
http://www.imagine-eyes.com/content/view/100/115/, “INOVEO—Ultra-High Resolution Retinal Imaging with Adaptive Optics,” 2 pages, printed Feb. 22, 2010.
AOptix Technologies, “Introducing the AOptix InSight 2 Meter Iris Recognition System,” 6 pages, 2010.
Belhumeur et al., “Eigenfaces Vs. Fisherfaces: Recognition Using Class Specific Linear Projection,” 14 pages, prior to Jun. 11, 2010.
Bentley et al., “Multidimensional Binary Search Trees Used for Associative Searching,” Communications of the ACM, vol. 18, No. 9, pp. 509-517, Sep. 1975.
Blackman et al., “Chapter 9, Multiple Sensor Tracking: Issues and Methods,” Design and Analysis of Modern Tracking Systems, Artech House, pp. 595-659, 1999.
Brasnett et al., “A Robust Visual Identifier Using the Trace Transform,” 6 pages, prior to Jun. 11, 2010.
Buades et al., “A Review of Image Denoising Algorithms, with a New One,” Multiscale Modeling & Simulation, vol. 4, No. 2, pp. 490-530, 2005.
Chen et al., “Localized Iris Image Quality Using 2-D Wavelets,” LNCS vol. 3832, pp. 373-381, 2005.
Chow et al., “Towards a System for Automatic Facial Feature Detection,” Pattern Recognition vol. 26, No. 12, pp. 1739-1755, 1993.
U.S. Appl. No. 12/792,498, filed Jun. 2, 2010.
U.S. Appl. No. 12/814,232, filed Jun. 11, 2010.
Cula et al., “Bidirectional Imaging and Modeling of Skin Texture,” Proceedings of Texture 2003, 6 pages, Oct. 17, 2003.
Cula et al., “Bidirectional Imaging and Modeling of Skin Texture,” IEEE Transactions on Biomedical Engineering, vol. 51, No. 12, pp. 2148-2159, 2004.
Cula et al., “Compact Representation of Bidirectional Texture Functions,” Proceedings of IEEE Computer Soviet Conference on Computer Vision and Pattern Recognition 2001, 8 pages, 2001.
Cula et al., “Skin Texture Modeling,” International Journal of Computer Vision 2004, 34 pages, 2004.
Dabov et al., “Image Denoising by Sparse 3-D Transform-Domain Collaborative Filtering,” IEEE Transactions on Image Processing, vol. 16, No. 8, pp. 2080-2095, Aug. 2007.
Dabov et al., “Image Restoration by Spars 3D Transform Collaborative Filtering,” SPIE vol. 6812 681207-1, 12 pages, 2008.
Daugman, “High Confidence Visual Recognition of Persons by a Test of Statistical Independence,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 15, No. 11, pp. 1148-1161, 1993.
Daugman, “Probing the Uniqueness and Randomness of Iris Codes: Results from 200 Billion Iris Pair Comparisons,” Proceedings of the IEEE vol. 94, No. 11, pp. 1928-1935, Nov. 2006.
Fooprateepsiri et al., “A Highly Robust Method for Face Authentication,” IEEE 2009 First Asian Conference on Intelligent Information and Database Systems, pp. 380-385, 2009.
Fooprateepsiri et al., “Face Verification Base-On Hausdorff-Shape Context,” IEEE 2009 Asia Conference on Informatics in Control, Automation and Robotics, pp. 240-244, 2009.
Forstner et al., “A Metric for Covariance Matrices,” 16 pages, prior to Jun. 11, 2010.
Gan et al., “Applications of Wavelet Packets Decomposition in Iris Recognition,” LNCS vol. 3832, pp. 443-449, 2005.
Hampapur et al., “Smart Surveillance: Applications, Technologies and Implications,” IEEE, 6 pages, Dec. 15-18, 2003.
Hamza et al., “Standoff Iris Recognition Usin Non-Iterative Polar Based Segmentation,” Proceedings of SPIE vol. 6944, 8 pages, 2008.
Hanna et al., “A System for Non-Intrusive Human Iris Acquisition and Identification,” IAPR Workshop on Machine Vision Applications, pp. 200-203, Nov. 12-14, 1996.
http://en.wikipedia.org/wiki/Radon—transform, “Radon Transform,” 5 pages, printed May 14, 2010.
Ivins et al., “A Deformable Model of the Human Iris for Measuring Small Three-Dimensional Eye Movements,” Machine Vision and Applications, vol. 11, pp. 42-51, 1998.
Kadyrov et al., “The Trace Transform and Its Applications,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 23, No. 8, pp. 811-828, Aug. 2001.
Kadyrov et al., “The Trace Transform as a Tool to Invariant Feature Construction,” 3 pages, prior to Jun. 11, 2010.
Kang et al., “Improved Dual Action Contour for Iris Recognition,” 10 pages, prior to Jun. 11, 2010.
Kawaguchi et al., “Detection of Eyes from Human Faces by Hough Transform and Separability Filter,” IEEE, 4 pages, 2000.
Kong et al., “Detecting Eyelash and Reflection for Accurate Iris Segmentation,” International Journal of Pattern Recognition and Artificial Intelligence, vol. 17, No. 6, pp. 1025-1034, 2003.
Li et al., “Appearance Modeling Using a Geometric Transform,” IEEE Transactions on Image Processing, 17 pages, 2008.
Li et al., “Appearance Modeling Using a Geometric Transform,” Journal Preparation for IEEE Transactions on Image Processing, 30 pages, Nov. 5, 2006.
Ma et al., “Local Intensity Variation Analysis for Iris Recognition,” Pattern Recognition, vol. 37, pp. 1287-1298, 2004.
Ma et al., “Video Sequence Querying Using Clustering of Objects' Appearance Models,” Advances in Visual Computing Third Annual Symposium, ISVC 2007, 14 pages, 2007.
Monro et al., “DCT-Based Iris Recognition,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 29, No. 4, Apr. 2007.
Noh et al., “A Novel Method to Extract Features for Iris Recognition System,” AVBPA 2003, LNCS 2688, pp. 862-868, 2003.
Ojala et al., “Multiresolution Gray-Scale and Rotation Invariant Texture Classification with Local Binary Patterns,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 24, No. 7, 18 pages, Jul. 2002.
Pamudurthy et al., “Dynamic Approach for Face Recognition Using Digital Image Skin Correlation,” Audio and Video Based Person Authentication 5th International Conference, AVBPA 2005, Hilton Rye Town, NY, USA, 11 pages, Jul. 20-22, 2005.
Petrou et al., “The Trace Transform in a Nutshell,” 9 pages, prior to Jun. 11, 2010.
Phillips et al., “FRVT 2006 and ICE 2006 Large-Scale Results,” 56 pages, Mar. 2007.
Porikli et al., “Covariance Tracking Using Model Update Based on Means on Riemannian Manifolds,” 8 pages, prior to Jun. 11, 2010.
Proenca et al., “Toward Noncooperative Iris Recognition: A Classification Approach Using Multiple Signatures,” IEEE Transactions on Patern Analysis and Machine Intellingence, vol. 29, No. 4, pp. 607-612, Apr. 2007.
Ross et al., “Segmenting Non-Ideal Irises Using Geodesic Active Contours,” IEEE 2006 Biometrics Symposium, 3 pages, 2006.
Shapiro et al., pp. 556-559 in Book Entitled “Computer Vision,” Prentice Hall, prior to Jun. 11, 2010.
Stillman et al., “A System for Tracking and Recognizing Multiple People with Multiple Cameras,” 6 pages, Aug. 1998.
Sun et al., “Iris Recognition Based on Non-local Comparisons,” Sinobiometrics 2004, LNCS 3338, pp. 67-77, 2004.
Suzaki et al., “A Horse Identification System Using Biometrics,” Systems and Computer in Japan, vol. 32, No. 14, pp. 12-23, 2001.
Trucco et al., “Robust Iris Location in Close-up Images of the Eye,” Pattern Anal. Applic. vol. 8, pp. 247-255, 2005.
Turan et al., “Trace Transform Based Invariant Object Recognition System,” 4 pages, prior to Jun. 11, 2010.
Turk et al., “Eigenfaces for Recognition,” Journal of Cognitive Neuroscience, vol. 3, No. 1, 16 pages, 1991.
Wang et al., “Recent Developments in Human Motion Analysis,” Pattern Recognition, vol. 36, pp. 585-601, 2003.
Wei et al., “Robust and Fast Assessment of Iris Image Quality,” LNCS vol. 3832, pp. 464-471, 2005.
Zhao et al., “Dynamic Texture Recognition Using Local Binary Patterns with an Application to Facial Expressions,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 29, No. 6, pp. 915-928, Jun. 2007.
Zhi-Hui et al., “Research Iris Serial Images Quality Assessment Method Based on HVS,” Proceedings of SPIE, vol. 6034, 6 pages, 2006.
Avcibas et al., “Steganalysis Using Image Quality Metrics,” IEEE Transactions on Image Processing, vol. 12, No. 2, pp. 221-229, Feb. 2003.
Boles, “A Security System Based on Human Iris Identification Using Wavelet Transform,” IEEE First International Conference on Knowledge Based Intelligent Electronic Systems, May 21-23, Adelaide, Australia, pp. 533-541, 1997.
Bonney et al., “Iris Pattern Extraction Using Bit Planes and Standard Deviations,” IEEE, pp. 582-586, 2004.
Camus et al., “Reliable and Fast Eye Finding in Close-up Images,” IEEE, pp. 389-394, 2002.
Carson et al., “Blobworld: Image Segmentation Using Expectation-Maximization and Its Application to Image Querying,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 24, No. 8, pp. 1026-1038, Aug. 2002.
Cui et al., “A Fast and Robust Iris Localization Method Based on Texture Segmentation,” 8 pages, 2004.
Cui et al., “An Appearance-Based Method for Iris Detection,” 6 pages, 2004.
Cui et al., “An Iris Detection Method Based on Structure Information,” Advances in Biometric Person Authentication, International Workshop on Biometric Recognition Systems, IWBRS 2005, Beijing China, 10 pages, Oct. 22-23, 2005.
Cui et al., “An Iris Image Synthesis Method Based on PCA and Super-Resolution,” IEEE Computer Society, Proceedings of the 17th International Conference on Pattern Recognition, 6 pages, Aug. 23-26, 2004.
Cui et al., “An Iris Recognition Algorithm Using Local Extreme Points,” Biometric Authentication, First International Conference, ICBA 2004, Hong Kong, China, 10 pages, Jul. 15-17, 2004.
Daugman, “Results From 200 Billion Iris Cross-Comparisons,” University of Cambridge Computer Laboratory, Technical Report, No. 635, 8 pages, Jun. 2005.
Daugman, “How Iris Recognition Works,” IEEE 2002 International Conference on Image Processing, vol. I of III, 6 pages, Sep. 22-25, 2002.
Du et al., “A One-Dimensional Approach for Iris Identification,” 11 pages, prior to Jan. 25, 2006.
Guo et al., “A System for Automatic Iris Capturing,” Mitsubishi Electric Research Laboratories, Inc., 10 pages, 2005.
Guo, “Face, Expression, and Iris Recognition Using Learning-Based Approaches,” 132 pages, 2006.
http://www.newscientisttech.com/article/dn11110-invention-covert-iris-sc, “Invention: Covert Iris Scanner,” 3 pages, printed Feb. 8, 2007.
Huang et al., “Iris Model Based on Local Orientation Description,” 5 pages, prior to Jan. 25, 2006.
Huang et al., “An Efficient Iris Recognition System,” IEEE Proceedings of the First International Conference on Machine Learning and Cybernetics, Beijing, pp. 450-454, Nov. 4-5, 2002.
Jalaja et al., “Texture Element Feature Characterizations for CBIR,” IEEE, pp. 733-736, 2005.
Kalka et al., “Image Quality Assessment for Iris Biometric,” Proc. of SPIE vol. 6202 62020D, 11 pages, 2006.
Ko et al., “Monitoring and Reporting of Fingerprint Image Quality and Match Accuracy for a Large User Application,” IEEE Computer Society, Proceedings of the 33rd Applied Imagery Pattern Recognition Workshops, 6 pages, 2004.
Lau et al., “Finding a Small Number of Regions in an Image Using Low-Level Features,” Pattern Recognition 35, pp. 2323-2339, 2002.
Ma et al., “Personal Identification Based on Iris Texture Analysis,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 25, No. 12, pp. 1519-1533, Dec. 2003.
Masek, “Recognition of Human Iris Patterns for Biometric Identification,” 61 pages, 2003.
Maurer et al., “Tracking and Learning Graphs and Pose on Image Sequences of Faces,” IEEE Computer Society Press, International Conference on Automatic Face and Gesture Recognition, pp. 176-181, Oct. 14-16, 1996.
Oppenheim et al, “The Importance of Phase in Signals,” Proceedings of the IEEE, vol. 69, No. 5, pp. 529-541, 1981.
Ratha et al., “A Real-Time Matching System for Large Fingerprint Databases,” IEEE Transactions on Pattern Analysis, and Machine Intelligence, vol. 18, No. 8, pp. 799-812, Aug. 1996.
Sony, “Network Color Camera, SNC-RZ30N (NTSC) ,” 6 pages, Aug. 2002.
Sun et al., “Robust Encoding of Local Ordinal Measures: A General Framework of Iris Recognition,” 13 pages, prior to Jan. 25, 2006.
Wang et al, “Image Quality Assessment: From Error Visibility to Structural Similarity,” IEEE Transactions on Image Processing vol. 13, No. 4, pp. 600-612, Apr. 2004.
Wang et al., “A Universal Image Quality Index,” IEEE Signal Processing Letters, vol. 9, No. 3, pp. 81-84, Mar. 2002.
Wang et al., “Local Phase Coherence and the Perception of Blur,” Advances in Nueral Information Processing Systems 16, pp. 1435-1442, 2004.
Related Publications (1)
Number Date Country
20100315500 A1 Dec 2010 US
Provisional Applications (1)
Number Date Country
61268678 Jun 2009 US