Method and system for indexing and searching an iris image database

Information

  • Patent Grant
  • 7761453
  • Patent Number
    7,761,453
  • Date Filed
    Friday, March 2, 2007
    17 years ago
  • Date Issued
    Tuesday, July 20, 2010
    14 years ago
Abstract
A method and system for indexing and searching a database of iris images having a system to expedite a process of matching a subject to millions (more or less) of templates within a database is disclosed. The system may progressively match an iris image template to an iris template in a database by progressing from a top layer of the database to a lower layer of the database. Such matching or retrieval may use a subject code as a query or probe and then find a similarity measure for the features of codes or templates in the database. A multi-stage hierarchal clustering process may be used to compress codes and/or templates.
Description
BACKGROUND

The present invention pertains to recognition systems and particularly to biometric recognition systems; in particular the invention pertains to iris recognition systems.


Related applications may include U.S. patent application Ser. No. 10/979,129, filed Nov. 3, 2004, which is a continuation-in-part of U.S. patent application Ser. No. 10/655,124, filed Sep. 5, 2003; and U.S. patent application Ser. No. 11/382,373, filed May 9, 2006, which are hereby incorporated by reference.


U.S. Provisional Application No. 60/778,770, filed Mar. 3, 2006, is hereby incorporated by reference.


U.S. patent application Ser. No. 11/275,703, filed Jan. 25, 2006, is hereby incorporated by reference.


U.S. Provisional Application No. 60/647,270, filed Jan. 26, 2005, is hereby incorporated by reference.


U.S. patent application Ser. No. 11/043,366, filed Jan. 26, 2005, is hereby incorporated by reference.


U.S. patent application Ser. No. 11/372,854, filed Mar. 10, 2006, is hereby incorporated by reference.


U.S. patent application Ser. No. 11/672,108, filed Feb. 7, 2007, is hereby incorporated by reference.


U.S. patent application Ser. No. 11/675,424, filed Feb. 15, 2007 is hereby incorporated by reference.


U.S. patent application Ser. No. 11/681,614, filed Mar. 2, 2007 is hereby incorporated by reference.


U.S. patent application Ser. No. 11/681,662, filed Mar. 2, 2007 is hereby incorporated by reference.


SUMMARY

The present invention is an indexing system to expedite the process of matching a subject to millions (more or less) of templates within a database.





BRIEF DESCRIPTION OF THE DRAWING


FIG. 1
a is a diagram of an overall iris recognition system with database indexing.



FIG. 1
b illustrates the progressive matching process;



FIG. 2 shows an overall four stage grouping approach of barcode bits to compute a more compact feature vector per each class;



FIG. 3 is a diagram showing phase based clustering;



FIG. 4 is a diagram showing angular based clustering;



FIG. 5 is a diagram showing radial clustering; and



FIG. 6 is a diagram showing bit formatting.





DESCRIPTION

The present system may relate to biometrics, iris recognition systems, image quality metrics, authentication, access control, monitoring, database retrieval, database query, fast database indexing, fast progressive identification, and security and surveillance systems.


The overall eye detection system is shown in FIG. 1a. It shows a camera 61 that may provide an image with a face in it to an eye finder 62 as noted herein. The eyefinder 62 may provide an image of one or two eyes that go to a preprocessing image quality assessment block 60. Output of block 60 may go to an iris segmentation block 63. A polar segmentation (POSE) system in block 63 may be used to perform the segmentation. POSE may be based on the assumption that image (e.g., 320×240 pixels) has a visible pupil where iris can be partially visible. There may be pupil segmentation at the inner border between the iris and pupil and segmentation at the outer border between the iris and the sclera and iris and eyelids. An output having a segmented image may go to a block 64 for mapping/normalization and feature extraction. An output from block 64 may go to an encoding block 65 which may provide an output, such as a barcode of the images put in terms of ones and zeros. An output of the encoding block 65 may go to an indexing block 68 through the choice of enrolling or matching of diamond or block 66 to group and cluster the bits for either enrolling or matching process. The choice of enrollment made at block 66, i.e., path 71 as input to indexing block 68 of the barcode may provide a basis for storage in block 70 of the eye information based upon sub-blocks of datasets which may be used for enrolling and indexing, at block 70. The block of bit indexing 68 may be configured based upon the clustering choice of the output of block 67 feeding into block 68. The indexing block of the query barcode once triggered by the matching option, i.e., through path 72 as input from block 66 will output the probe barcode to progressive matching of block 69. The progressive matching of block 69 may match the subclasses of the barcode bits using the configuration of the database grouping. The clustering scheme, once selected at block 67 as input to block 68, should be kept the same for both enrollment and matching for consistency at inputs 74 and 75, respectively.


According to related practices, an information content based retrieval approach while clustering analysis may be conducted on bit codes and not on raw iris images. Further retrieval of an iris template from the database when a query iris is given by processing the raw iris image may be a time consuming task. In this invention, one may present a rapid approach to analyze progressively the bit codes to match against a very large datasets. Furthermore, the iris code may be additionally compressed to fewer bits than previously done before extracting the feature vector for indexing. One may selectively use discriminating information in the iris code as a query to compute the feature vectors for the indexing and querying. This may be regarded as a multi-level of clustering to yield a progressive matching using at least three methods of grouping of bits into subclasses using phase, radial, and angular information and then conducting the matching on only limited subclasses of bits and or using a compact bit format to compute a more compact feature vectors of the barcodes.


The present invention may be an approach and system to classify a given query iris into one of the pre-defined classes. Irises may be retrieved from a large collection of databases using partial information of the iris barcode as a probe and then find the similarity measure for all the features of the iris codes in the database. The information may then be augmented progressively to limited numbers of potential probes of barcode templates in the database. Partial features may be extracted using phase, angular or radial clustering. The resulting bits may be used then for matching or additional cumulative bit analysis on a very compressed version of the resulting iris code bits can be deployed instead. The terms probe and query may be used interchangeably herein.


The bit clustering may involve grouping or classifying of a barcode bits into subsets based upon the phase information, angular location or radial range. This classification of bits into subsets (i.e. subclasses) allows a progressive query and matching starting from fewer bits against all probes in the database (i.e., starting with a single or few classes of bits) and iteratively may include additional classes of bits that would limit the matching process to fewer template queries. When grouping the bits using present clustering methods, it is expected that the number of probes used for matching is inversely dependent on the number of bits being deployed in the matching process. At each step of the progressive matching, only potential matches are kept for next steps of analysis. Hence, the more one progresses in the query by deploying additional bits, the more one eliminates non-potential matchers, leading to fewer templates to match against in next rounds. In other words, the more bits deployed, the less number of templates being used to match against. This may be the main benefit of the present indexing method that is to allow matching the barcode probe against only few of the template queries. In the following, one may present an example to illustrate the progressive matching that reduces the required computational burden of matching large databases.


Assume that one has 10,000 templates of enrolled barcodes as queries within the database. Each template may have 240 bytes of information in bits. If one conducts a one-to-many matching process, this may require 2.4 mega bytes of bit matches. In the present invention, each barcode may be classified based upon the phase, angular, and radial information. One may assume to have a 2 bit representation of each pixel. Thus, one may cluster the bits of the barcode into two subgroups using the phase clustering. One may cluster further the bits into 5 classes based upon the angular segments, and in addition may cluster the bits into 3 groups of radial ranges. This will result into 3×5×2=30 subclasses of the databases. And each barcode may be subdivided into 30 subsets of bits; each of subset has only 8 bytes. Thus, the configuration of the database in FIG. 1b will result into 30 different datasets stacked within the database as shown at the right hand side of the database grouping. Each dataset may have a class of the bits of all the 10,000 barcode probes. Thus, at the first matching process, one may match only an 8 byte subset (block 82 in FIG. 1b) of the query barcode against only the dataset 81 of FIG. 1b. The dataset 81 represents only partial information of 8 bytes of each of the 10,000 probes. The matching process of 81 may result into 1000 good potential matches out of the 10,000 probes. At the second step of the progressive matching, one may use only these 100 good matches for analysis. One may match another subset of the barcode of size 8 bytes (block 84 in FIG. 1b) against only the corresponding database 83 and limit the matching to only those 1000 probes. One may note that each dataset like 83 has representation of all 10,000 probes. This may result, for example, into 100 good matches. In step 3, one may select a third 8 byte class of the barcode to match against the resulting 100 probes of dataset 85. One may obtain, for example, 10 good matches out of the 100 probes. One might choose at a certain stage to quit the progressive match given that the numbers of potential probes has been limited to only few probes. One may thus process the rest of the barcode subsections against these 10 probes. The overall required matching operations for this example are (10,000×8+1000×8+100×8+10×240) bytes which is about 0.091 Mega bytes of matching operations, compared to raw matching of 2.4 Meg bytes of operations (that is only 3.8% of computation required in the progressive matching than the regular matching.)


The choice of how one executes bits' grouping into separate subclasses is very crucial to the success of this indexing mechanism. Three methods of bit clustering (phase, angular, and radial) may be used to group the bits into different subclasses. At each clustering stage, one might group bits based upon the types of information and how it pertains to the iris texture representation. For instance, during the progressive matching, one may have to start with the key discriminant bit subclasses at early stage using classes with bits presenting the most significant iris texture information, and end at the least significant classes for last rounds.


The most significant classes are those bits associated with key features of the iris textures. In particular, at the radial axis, one has affirmed that most information of the iris textures is likely to be at the vicinity of the inner border of the iris. At the angular axis, one may realize that the bits representing the pixels of the sclera/iris segment are more reliable than the bits representing the eyelid/iris segments due to high noise-to-signal ratio at the eyelid angular ranges.


An unsupervised iris classification approach may automatically cluster iris codes into a set of prototype classes where the number of desired categories is merely specified a priori. The present iris classification may be based on the most discriminating information content present in an iris pattern and deal with classifying all bits in an iris barcode into one of several classes or themes. To retrieve irises from a large collection of databases, a selective subsection of iris code may be used as a query and one may compute a compact form of the this selective subsection of iris code represented by two feature vectors for the query iris.


Organization of feature data to group the bits of an iris code may be completed in a hierarchical form to compare one subset of bits at a time starting at the extracted feature elements that are closest to the inner iris boundary (row cumulative analysis). If the cumulative feature vector is conducted on the columns, the features that correspond to the middle sclera may be treated first. A simple tree classifier using hamming distance or any other information measure may be utilized as a measure to find similarity for all the features of the irises in the database. The feature vector may be computed using a four stage clustering process to compress the iris barcode to tens of feature elements (i.e., fewer bits of information) used in the class bits.


The present system and approach may provide a fast search and indexing of irises based on four stage selective clustering procedure as shown in FIG. 2. The four stages may include phase clustering, radial clustering, angular clustering and bit clustering. The approach may extract the most valuable discriminant information in the iris at each stage based upon different criteria and then cluster the dataset based upon the extracted significant features of the compressed iris bits. The matching may be executed on a subset of barcode bits or a compact feature vector of 12 bit elements representing the subset of compressed bits. One may deploy an unsupervised tree classifier using a Euclidian distance or binning approach as a measure to find similarity between classes and iris query within sub datasets to break down the datasets into smaller sets.


The present multiple-stage clustering approach and system may begin with a phase clustering stage 11 where the initial size of a code 21 and mask 22 may be 2×Nr×Nθ. The phase clustering 11 may reduce the size of code 21 and mask 22 down to a size of Nr×Nθ, as shown by code 23 and mask 24. A corrupt bits index may be indicated by nk. The stage, radial clustering 12, may reduce code 23 and mask 24 to a size of code 25 and mask 26, which is nk×Nθ. Code 25 may be sclera based. An angular clustering stage 13 may reduce code 25 and mask 26 down to a size of nk×Nθ×Δθs/π, as shown by code 27 and mask 28. In turn, a bit clustering stage 14 may reduce the code 27 and mask 28 size down to a code 29 and mask 30 having a size equal 12nk (if summation of bits is conducted on the horizontal axis) or 12Nθ×Δθs/π if summation is done on the vertical axis.



FIG. 3 is a diagram of phased based clustering 12. In order to encode iris pattern data, decomposition of an iris feature vector 31 having the size 2Nr×Nθ may be accomplished by using a quadratic pair of filters 34 and 35, with real part specified by an even symmetric filter 34 and the imaginary part specified by an odd symmetric filter 35. Both outputs of the filters may be uncorrelated and be treated separately. Thus, a decomposition may occur into parts 32 and 33 each having the size Nr×Nθ.



FIG. 4 is a diagram of angular based clustering. The encoding process may produce a bitwise template containing a number of bits of information covering both the regions bordering the sclera and other adjacent iris segments to top and bottom eyelids. A corresponding noise mask which corresponds to corrupt areas within the iris pattern may be constructed to exclude these noisy bits. Because of the nature of eyelid and eyelash obscurations, most of the corrupt regions may be at the vicinity of the eyelids apart from the sclera central boundaries. Thus, the bits corresponding to the central sclera boundaries may present a more accurate presentation of the actual iris patterns. This stage shows a decomposition of the iris code based upon the boundary limits of the central sclera as shown in FIG. 4.


The sclera portions 41 and 42 of the iris are indicated by boundaries 43 and 44 having sector angles Δθ{tilde over (s)} and Δθs, respectively. The code 45 may have a dimension 2N×Nθ. It may be noted that,

Δθ{tilde over (s)}=[θ→θ+]+φh,
Δθ{tilde over (s)}=Δθs+π,
θ+=[π/4→3π/8], and
θ+,

where φh is the expected head orientation. The decomposition of the iris code 45 based on the boundary 43 and 44 limits may lead to a sclera based code 46 and a lid based code 47. The lid portions may have masking 48 and 49 for the sclera based code 46. The sclera portions may have masking 51 and 52 for the lid based code 47. The dimension of the sclera based code may be

2Nr×Nθ×Δθs/π.

The dimension of the lid based code may be

2Nr×Nθ×(π−Δθs)/π.



FIG. 5 is a diagram of radial based clustering. Many of the iris patterns may reside at the inner boundaries of the iris as shown in the iris map 55 of the Figure, with the exception of a few samples where the iris pattern is spread consistently over the entire iris region. Even with these few exceptions, it is expected that the outer boundary of the iris will include most of the corrupt iris pixels due to obscuration of eyelids and eyelashes which eventually will be masked out of the analysis. A decomposition of the iris code 53, having a size of Nr×Nθ, may be effected based upon a radial segment that has the least number of corrupt bits. To control the number of corrupt bits, one may replace the original mask (of size Nr×Nθ bits) with a partial of the range, example the first single byte (8 bits) only to indicate the index of the least radius nr which has the minimum number of corrupt bits (sum of x bits in horizontal direction that does not exceed a predefined fraction (e.g., 5 percent) of the angular resolution). The corrupt bits index may be indicated by nk. Mask 56 is revealed on map 55. With the masked bits removed, one may have a code 54 having a size of 2nr×Nθ. The number of rows used in the comparison may be defined based on the minimum value of the index of corrupt bits in both the template and the code to be compared. One may note that nr=min (nk, nTemplate).



FIG. 6 is a diagram of a compact bit format. An iris code 57 may be noted with a size 2Nr×Nθ. The bit-to-bit comparison of bar codes may be replaced by a comparison of two 12 bit vectors that represent the iris map. The feature vector {right arrow over (F)} may be computed by summing the row bit values. The first column left of code 57, with heading Δ{right arrow over (F)}, may be a 12 bit presentation of “x” bits (mask bits)=error in elements. The column, adjacent to the first column, may be a 12 bit presentation of “1” bits=feature vector elements. One may track the number of corrupt bits by the 12 bit vector Δ{right arrow over (F)} to account for the error margin during a comparison of feature vectors. One may chose to sum over the angular values so that the comparison will not be affected by any head orientation. A similar approach may be applied to sum through columns rather than rows or both simultaneously.


One may have {tilde over (F)}={right arrow over (F)}+Δ{right arrow over (F)}. {tilde over (F)}, the resultant 24Nθ bit feature vector 58, is shown in the Figure. The present approach may expedite the process by at least the number of shifts times over the related art techniques given that summation over the angular range does not require shifting for actual matching. Summing over the radial may still dictate the use of shifting of these feature vectors to account for head orientation changes.


In the present specification, some of the matter may be of a hypothetical or prophetic nature although stated in another manner or tense.


Although the invention has been described with respect to at least one illustrative example, many variations and modifications will become apparent to those skilled in the art upon reading the present specification. It is therefore the intention that the appended claims be interpreted as broadly as possible in view of the prior art to include all such variations and modifications.

Claims
  • 1. A method of indexing and searching a database comprising: providing a storage module;providing a comparator module communicatively coupled to the storage module;providing a database of iris templates within the storage module, wherein each iris template in the database comprises a barcode configured to have a certain number of bits;hierarchically partitioning the database into a plurality of layers of sub-databases of the iris templates based on phase information, radial information, and/or angular information of an iris image, wherein each iris template in each of the layers of the sub-databases comprises a segment of the iris template barcode and the sum of the iris templates of each layer of sub-databases represents all the iris templates in the database;providing a barcode of an iris image for matching to an iris template in the database, wherein the barcode of the iris image has the same configuration as the configuration of each iris template in the database;selecting a segment of the barcode of the iris image that corresponds to a segment of an iris template in a layer of the sub-databases; andprogressively matching the selected segment of the barcode of the iris image to an iris template in a layer of the sub-databases by progressing from a top layer of the sub-databases to a lower layer of the sub-databases; andwherein the layers of the sub-databases comprise:phase clustering;radial clustering;angular clustering; andbit clustering or binning.
  • 2. The method of claim 1, wherein the segment of the iris template barcode includes only a combination or transformation of a portion of the bits in a barcode.
  • 3. The method of claim 1, wherein hierarchically partitioning the database into a plurality of sub-databases of templates based on phase information, radial information, and/or angular information includes two or more types of information.
  • 4. The method of claim 3, wherein the two or more types of information is progressive relative to the number of layers of the sub-databases in the plurality of sub-databases and a segment of each iris template in the sub-database and the segment is a number of bits representing an iris template of the sub-database.
  • 5. The method of claim 4, wherein a first layer of the sub-databases of the database has iris templates represented by an iris template including a larger number of bits than a second layer of the sub-databases.
  • 6. The method of claim 5, wherein a number of iris templates in the first layer of the sub-databases is the same as or larger than a number of iris templates in the second layer of the sub-databases.
  • 7. The method of claim 1, wherein a number of iris templates used for matching is reduced when progressing from one layer of the sub-databases to another layer of the sub-databases.
  • 8. The method of claim 6, wherein a matching of an iris image barcode is reduced to a least number of iris templates.
  • 9. The method of claim 8, wherein bit summation or combination or transformation of bits on a barcode is used to compress the barcode.
  • 10. The method of claim 9, wherein bit summation or transformation on a barcode bits is used to increase a speed of matching an iris image barcode to an iris template in the database.
  • 11. The method of claim 8, wherein bit summation on a horizontal or vertical axis on the barcode is used to increase a speed of matching the barcode to an iris template in the database.
  • 12. The method of claim 1, wherein the radial information to start is based on information close to an inner border of an iris closest to the pupil borders.
  • 13. The method of claim 1, wherein the angular information is based on angular segments and sclera-iris angular segments away from the eyelids are considered for matching before other segments.
  • 14. A method for matching an iris image barcode to a database comprising: providing a barcode of an iris image;providing a storage module;providing a comparator module communicatively coupled to the storage module;providing a database comprising a certain number of iris templates within the storage module, the iris templates configured to have a certain number of bits;hierarchically partitioning the database into multiple layers of datasets based on phase information, radial information, and/or angular information, the number of iris templates in each dataset are equal to or smaller than the number of iris templates in the database and the certain number of templates in each dataset configured to have a segment of the iris template as a certain number of bits equal to or smaller than the certain number of bits of the iris templates in the database;matching a segment of the barcode of an iris image to a corresponding segment of an iris template stored in the database by progressively matching the certain number of bits of the iris image barcode to the corresponding certain number of bits of the iris templates in the dataset progressing from one layer of the dataset to another layer of the dataset.
  • 15. The method of claim 14 wherein: the matching of the segment of the barcode of an iris image to an iris template of the dataset results in a number of iris templates which are a possible match for the barcode of an iris image; andthe number of iris templates which are a possible match for the barcode of an iris image is less than the number of iris templates in the database.
  • 16. A method for searching a database comprising: providing a storage module;providing a comparator module communicatively coupled to the storage module;providing a database comprising iris templates within the storage module, the iris templates configured to include information as a certain number of bits;providing an iris image barcode configured to include information as a certain number of bits corresponding to the certain number of bits of the iris templates in the database;partitioning the database into a number of sub-databases based on phase information, the number of sub-databases comprising a number of iris templates, the number of iris templates configured to include information as a segment of the iris template having a certain number of bits;partitioning the sub-databases into a number of datasets based on angular information, the number of datasets comprising a number of iris templates, the number of iris templates configured to include information as a segment of the iris template having a certain number of bits;partitioning the datasets into a number of sub-datasets based on radial information, the number of sub-datasets comprising a number of iris templates, the number of iris templates configured to include information as a segment of the iris template having a certain number of bits; andmatching a segment of the iris barcode to a corresponding segment of an iris template in the database by progressively matching the certain number of bits of the iris image barcode to the corresponding certain number of bits of the iris templates progressing from the sub-databases to the datasets, and to the sub-datasets;
  • 17. The method of claim 16, wherein a progression of templates and datasets continues until a dataset exists for all types of available information.
Parent Case Info

This application claims the benefit of U.S. Provisional Application No. 60/778,770, filed Mar. 3, 2006. This application is a continuation-in-part of U.S. patent application Ser. No. 11/275,703, filed Jan. 25, 2006, which claims the benefit of U.S. Provisional Application No. 60/647,270, filed Jan. 26, 2005. This application is a continuation-in-part of U.S. patent application Ser. No. 11/043,366, filed Jan. 26, 2005. This application is a continuation-in-part of U.S. patent application Ser. No. 11/372,854, filed Mar. 10, 2006; This application is a continuation-in-part of U.S. patent application Ser. No. 11/672,108, filed Feb. 7, 2007. This application is a continuation-in-part of U.S. patent application Ser. No. 11/675,424, filed Feb. 15, 2007. This application is a continuation-in-part of U.S. patent application Ser. No. 11/681,614, filed Mar. 2, 2007. This application is a continuation-in-part of U.S. patent application Ser. No. 11/681,662, filed Mar. 2, 2007.

Government Interests

The government may have rights in the present invention.

US Referenced Citations (311)
Number Name Date Kind
4641349 Flom et al. Feb 1987 A
5291560 Daugman Mar 1994 A
5293427 Ueno et al. Mar 1994 A
5551027 Choy et al. Aug 1996 A
5572596 Wildes et al. Nov 1996 A
5608472 Szirth et al. Mar 1997 A
5717512 Chmielewski, Jr. et al. Feb 1998 A
5751836 Wildes et al. May 1998 A
5859686 Aboutalib et al. Jan 1999 A
5901238 Matsushita May 1999 A
5953440 Zhang et al. Sep 1999 A
5956122 Doster Sep 1999 A
5978494 Zhang Nov 1999 A
6005704 Chmielewski, Jr. et al. Dec 1999 A
6007202 Apple et al. Dec 1999 A
6021210 Camus et al. Feb 2000 A
6028949 McKendall Feb 2000 A
6055322 Salganicoff et al. Apr 2000 A
6064752 Rozmus et al. May 2000 A
6069967 Rozmus et al. May 2000 A
6081607 Mori et al. Jun 2000 A
6088470 Camus et al. Jul 2000 A
6091899 Konishi et al. Jul 2000 A
6101477 Hohle et al. Aug 2000 A
6104431 Inoue et al. Aug 2000 A
6108636 Yap et al. Aug 2000 A
6119096 Mann et al. Sep 2000 A
6120461 Smyth Sep 2000 A
6144754 Okano et al. Nov 2000 A
6246751 Bergl et al. Jun 2001 B1
6247813 Kim et al. Jun 2001 B1
6252977 Salganicoff et al. Jun 2001 B1
6282475 Washington Aug 2001 B1
6285505 Melville et al. Sep 2001 B1
6285780 Yamakita et al. Sep 2001 B1
6289113 McHugh et al. Sep 2001 B1
6299306 Braithwaite et al. Oct 2001 B1
6308015 Matsumoto Oct 2001 B1
6309069 Seal et al. Oct 2001 B1
6320610 Van Sant et al. Nov 2001 B1
6320973 Suzaki et al. Nov 2001 B2
6323761 Son Nov 2001 B1
6325765 Hay et al. Dec 2001 B1
6330674 Angelo et al. Dec 2001 B1
6332193 Glass et al. Dec 2001 B1
6344683 Kim Feb 2002 B1
6370260 Pavlidis et al. Apr 2002 B1
6377699 Musgrave et al. Apr 2002 B1
6400835 Lemelson et al. Jun 2002 B1
6424727 Musgrave et al. Jul 2002 B1
6424845 Emmoft et al. Jul 2002 B1
6433818 Steinberg et al. Aug 2002 B1
6438752 McClard Aug 2002 B1
6441482 Foster Aug 2002 B1
6446045 Stone et al. Sep 2002 B1
6483930 Musgrave et al. Nov 2002 B1
6484936 Nicoll et al. Nov 2002 B1
6490443 Freeny, Jr. Dec 2002 B1
6493363 Roger et al. Dec 2002 B1
6493669 Curry et al. Dec 2002 B1
6494363 Roger et al. Dec 2002 B1
6503163 Van Sant et al. Jan 2003 B1
6505193 Musgrave et al. Jan 2003 B1
6516078 Yang et al. Feb 2003 B1
6516087 Camus Feb 2003 B1
6516416 Gregg et al. Feb 2003 B2
6522772 Morrison et al. Feb 2003 B1
6526160 Ito Feb 2003 B1
6532298 Cambier et al. Mar 2003 B1
6540392 Braithwaite Apr 2003 B1
6542624 Oda Apr 2003 B1
6546121 Oda Apr 2003 B1
6553494 Glass Apr 2003 B1
6580356 Alt et al. Jun 2003 B1
6591001 Oda et al. Jul 2003 B1
6591064 Higashiyama et al. Jul 2003 B2
6594377 Kim et al. Jul 2003 B1
6594399 Camus et al. Jul 2003 B1
6598971 Cleveland Jul 2003 B2
6614919 Suzaki et al. Sep 2003 B1
6652099 Chae et al. Nov 2003 B2
6674367 Sweatte Jan 2004 B2
6690997 Rivalto Feb 2004 B2
6708176 Strunk et al. Mar 2004 B2
6711562 Ross et al. Mar 2004 B1
6714665 Hanna et al. Mar 2004 B1
6718049 Pavlidis et al. Apr 2004 B2
6732278 Baird, III et al. May 2004 B2
6734783 Anbai May 2004 B1
6745520 Puskaric et al. Jun 2004 B2
6751733 Nakamura et al. Jun 2004 B1
6753919 Daugman Jun 2004 B1
6754640 Bozeman Jun 2004 B2
6760467 Min et al. Jul 2004 B1
6765470 Shinzaki Jul 2004 B2
6766041 Golden et al. Jul 2004 B2
6775774 Harper Aug 2004 B1
6785406 Kamada Aug 2004 B1
6793134 Clark Sep 2004 B2
6819219 Bolle et al. Nov 2004 B1
6829370 Pavlidis et al. Dec 2004 B1
6832044 Doi et al. Dec 2004 B2
6836554 Bolle et al. Dec 2004 B1
6837436 Swartz et al. Jan 2005 B2
6845479 Park Jan 2005 B2
6853444 Haddad Feb 2005 B2
6867683 Calvesio et al. Mar 2005 B2
6873960 Wood et al. Mar 2005 B1
6896187 Stockhammer May 2005 B2
6905411 Nguyen et al. Jun 2005 B2
6920237 Chen et al. Jul 2005 B2
6930707 Bates et al. Aug 2005 B2
6934849 Kramer et al. Aug 2005 B2
6954738 Wang et al. Oct 2005 B2
6957341 Rice et al. Oct 2005 B2
6972797 Izumi Dec 2005 B2
7053948 Konishi May 2006 B2
7184577 Chen et al. Feb 2007 B2
20010026632 Tamai Oct 2001 A1
20010027116 Baird Oct 2001 A1
20010047479 Bromba et al. Nov 2001 A1
20010051924 Uberti Dec 2001 A1
20010054154 Tam Dec 2001 A1
20020010857 Karthik Jan 2002 A1
20020033896 Hatano Mar 2002 A1
20020039433 Shin Apr 2002 A1
20020040434 Elliston et al. Apr 2002 A1
20020062280 Zachariassen et al. May 2002 A1
20020077841 Thompson Jun 2002 A1
20020089157 Breed et al. Jul 2002 A1
20020106113 Park Aug 2002 A1
20020112177 Voltmer et al. Aug 2002 A1
20020114495 Chen et al. Aug 2002 A1
20020130961 Lee et al. Sep 2002 A1
20020131622 Lee et al. Sep 2002 A1
20020139842 Swaine Oct 2002 A1
20020140715 Smet Oct 2002 A1
20020142844 Kerr Oct 2002 A1
20020144128 Rahman et al. Oct 2002 A1
20020150281 Cho Oct 2002 A1
20020154794 Cho Oct 2002 A1
20020158750 Almalik Oct 2002 A1
20020164054 McCartney et al. Nov 2002 A1
20020175182 Matthews Nov 2002 A1
20020186131 Fettis Dec 2002 A1
20020191075 Doi et al. Dec 2002 A1
20020191076 Wada et al. Dec 2002 A1
20020194128 Maritzen et al. Dec 2002 A1
20020194131 Dick Dec 2002 A1
20020198731 Barnes et al. Dec 2002 A1
20030002714 Wakiyama Jan 2003 A1
20030012413 Kusakari et al. Jan 2003 A1
20030014372 Wheeler et al. Jan 2003 A1
20030020828 Ooi et al. Jan 2003 A1
20030038173 Blackson et al. Feb 2003 A1
20030046228 Berney Mar 2003 A1
20030053663 Chen et al. Mar 2003 A1
20030055689 Block et al. Mar 2003 A1
20030055787 Fujii Mar 2003 A1
20030058492 Wakiyama Mar 2003 A1
20030061172 Robinson Mar 2003 A1
20030061233 Manasse et al. Mar 2003 A1
20030065626 Allen Apr 2003 A1
20030071743 Seah et al. Apr 2003 A1
20030072475 Tamori Apr 2003 A1
20030073499 Reece Apr 2003 A1
20030074317 Hofi Apr 2003 A1
20030074326 Byers Apr 2003 A1
20030076161 Tisse Apr 2003 A1
20030076300 Lauper et al. Apr 2003 A1
20030076984 Tisse et al. Apr 2003 A1
20030080194 O'Hara et al. May 2003 A1
20030091215 Lauper et al. May 2003 A1
20030092489 Veradej May 2003 A1
20030095689 Volkommer et al. May 2003 A1
20030098776 Friedli May 2003 A1
20030099379 Monk et al. May 2003 A1
20030099381 Ohba May 2003 A1
20030103652 Lee et al. Jun 2003 A1
20030107097 McArthur et al. Jun 2003 A1
20030107645 Yoon Jun 2003 A1
20030108224 Ike Jun 2003 A1
20030108225 Li Jun 2003 A1
20030115148 Takhar Jun 2003 A1
20030115459 Monk Jun 2003 A1
20030116630 Carey et al. Jun 2003 A1
20030118212 Min et al. Jun 2003 A1
20030118217 Kondo et al. Jun 2003 A1
20030123711 Kim et al. Jul 2003 A1
20030125054 Garcia Jul 2003 A1
20030125057 Pesola Jul 2003 A1
20030126560 Kurapati et al. Jul 2003 A1
20030131245 Linderman Jul 2003 A1
20030131265 Bhakta Jul 2003 A1
20030133597 Moore et al. Jul 2003 A1
20030140235 Immega et al. Jul 2003 A1
20030140928 Bui et al. Jul 2003 A1
20030141411 Pandya et al. Jul 2003 A1
20030149881 Patel et al. Aug 2003 A1
20030152251 Ike Aug 2003 A1
20030152252 Kondo et al. Aug 2003 A1
20030156741 Lee et al. Aug 2003 A1
20030158762 Wu Aug 2003 A1
20030158821 Maia Aug 2003 A1
20030159051 Hollnagel Aug 2003 A1
20030163739 Armington et al. Aug 2003 A1
20030169334 Braithwaite et al. Sep 2003 A1
20030169901 Pavlidis et al. Sep 2003 A1
20030169907 Edwards et al. Sep 2003 A1
20030173408 Mosher, Jr. et al. Sep 2003 A1
20030174049 Beigel et al. Sep 2003 A1
20030177051 Driscoll et al. Sep 2003 A1
20030182151 Taslitz Sep 2003 A1
20030182182 Kocher Sep 2003 A1
20030191949 Odagawa Oct 2003 A1
20030194112 Lee Oct 2003 A1
20030195935 Leeper Oct 2003 A1
20030198368 Kee Oct 2003 A1
20030200180 Phelan, III et al. Oct 2003 A1
20030210139 Brooks et al. Nov 2003 A1
20030210802 Schuessier Nov 2003 A1
20030218719 Abourizk et al. Nov 2003 A1
20030225711 Paping Dec 2003 A1
20030228898 Rowe Dec 2003 A1
20030233556 Angelo et al. Dec 2003 A1
20030235326 Morikawa et al. Dec 2003 A1
20030235411 Morikawa et al. Dec 2003 A1
20030236120 Reece et al. Dec 2003 A1
20040001614 Russon et al. Jan 2004 A1
20040002894 Kocher Jan 2004 A1
20040005078 Tillotson Jan 2004 A1
20040006553 de Vries et al. Jan 2004 A1
20040010462 Moon et al. Jan 2004 A1
20040012760 Mihashi et al. Jan 2004 A1
20040019570 Bolle et al. Jan 2004 A1
20040023664 Mirouze et al. Feb 2004 A1
20040023709 Beaulieu et al. Feb 2004 A1
20040025030 Corbett-Clark et al. Feb 2004 A1
20040025031 Ooi et al. Feb 2004 A1
20040025053 Hayward Feb 2004 A1
20040029564 Hodge Feb 2004 A1
20040030930 Nomura Feb 2004 A1
20040035123 Kim et al. Feb 2004 A1
20040037450 Bradski Feb 2004 A1
20040039914 Barr et al. Feb 2004 A1
20040042641 Jakubowski Mar 2004 A1
20040044627 Russell et al. Mar 2004 A1
20040046640 Jourdain et al. Mar 2004 A1
20040049687 Orsini et al. Mar 2004 A1
20040050924 Mletzko et al. Mar 2004 A1
20040050930 Rowe Mar 2004 A1
20040052405 Walfridsson Mar 2004 A1
20040052418 DeLean Mar 2004 A1
20040059590 Mercredi et al. Mar 2004 A1
20040059953 Purnell Mar 2004 A1
20040104266 Bolle et al. Jun 2004 A1
20040117636 Cheng Jun 2004 A1
20040133804 Smith et al. Jul 2004 A1
20040146187 Jeng Jul 2004 A1
20040148526 Sands et al. Jul 2004 A1
20040160518 Park Aug 2004 A1
20040162870 Matsuzaki et al. Aug 2004 A1
20040162984 Freeman et al. Aug 2004 A1
20040169817 Grotehusmann et al. Sep 2004 A1
20040172541 Ando et al. Sep 2004 A1
20040174070 Voda et al. Sep 2004 A1
20040179720 Chen et al. Sep 2004 A1
20040190759 Caldwell Sep 2004 A1
20040193893 Braithwaite et al. Sep 2004 A1
20040219902 Lee et al. Nov 2004 A1
20040233038 Beenau et al. Nov 2004 A1
20040240711 Hamza et al. Dec 2004 A1
20040252866 Tisse et al. Dec 2004 A1
20040255168 Murashita et al. Dec 2004 A1
20050008200 Azuma et al. Jan 2005 A1
20050008201 Lee et al. Jan 2005 A1
20050029353 Isemura et al. Feb 2005 A1
20050052566 Kato Mar 2005 A1
20050055582 Bazakos et al. Mar 2005 A1
20050063567 Saitoh et al. Mar 2005 A1
20050084137 Kim et al. Apr 2005 A1
20050084179 Hanna et al. Apr 2005 A1
20050099288 Spitz et al. May 2005 A1
20050102502 Sagen May 2005 A1
20050110610 Bazakos et al. May 2005 A1
20050125258 Yellin et al. Jun 2005 A1
20050127161 Smith et al. Jun 2005 A1
20050129286 Hekimian Jun 2005 A1
20050134796 Zelvin et al. Jun 2005 A1
20050138385 Friedli et al. Jun 2005 A1
20050138387 Lam et al. Jun 2005 A1
20050146640 Shibata Jul 2005 A1
20050151620 Neumann Jul 2005 A1
20050152583 Kondo et al. Jul 2005 A1
20050193212 Yuhara Sep 2005 A1
20050199708 Friedman Sep 2005 A1
20050206501 Farhat Sep 2005 A1
20050206502 Bernitz Sep 2005 A1
20050207614 Schonberg et al. Sep 2005 A1
20050210267 Sugano et al. Sep 2005 A1
20050210270 Rohatgi et al. Sep 2005 A1
20050210271 Chou et al. Sep 2005 A1
20050238214 Matsuda et al. Oct 2005 A1
20050240778 Saito Oct 2005 A1
20050248725 Ikoma et al. Nov 2005 A1
20050249385 Kondo et al. Nov 2005 A1
20050255840 Markham Nov 2005 A1
20060165266 Hamza Jul 2006 A1
20060274919 LoIacono et al. Dec 2006 A1
20070036397 Hamza Feb 2007 A1
20070047772 Matey et al. Mar 2007 A1
Foreign Referenced Citations (160)
Number Date Country
0878780 Nov 1998 EP
0910986 Apr 1999 EP
0962894 Dec 1999 EP
1018297 Jul 2000 EP
1024463 Aug 2000 EP
1028398 Aug 2000 EP
1041506 Oct 2000 EP
1041523 Oct 2000 EP
1126403 Aug 2001 EP
1477925 Nov 2004 EP
2369205 May 2002 GB
2371396 Jul 2002 GB
2375913 Nov 2002 GB
2402840 Dec 2004 GB
2411980 Sep 2005 GB
9161135 Jun 1997 JP
9198545 Jul 1997 JP
9201348 Aug 1997 JP
9147233 Sep 1997 JP
9234264 Sep 1997 JP
9305765 Nov 1997 JP
9319927 Dec 1997 JP
10021392 Jan 1998 JP
10040386 Feb 1998 JP
10049728 Feb 1998 JP
10137219 May 1998 JP
10137221 May 1998 JP
10137222 May 1998 JP
10137223 May 1998 JP
10248827 Sep 1998 JP
10269183 Oct 1998 JP
11047117 Feb 1999 JP
11089820 Apr 1999 JP
11200684 Jul 1999 JP
11203478 Jul 1999 JP
11213047 Aug 1999 JP
11339037 Dec 1999 JP
2000005149 Jan 2000 JP
2000005150 Jan 2000 JP
2000011163 Jan 2000 JP
2000023946 Jan 2000 JP
2000083930 Mar 2000 JP
2000102510 Apr 2000 JP
2000102524 Apr 2000 JP
2000105830 Apr 2000 JP
2000107156 Apr 2000 JP
2000139878 May 2000 JP
2000155863 Jun 2000 JP
2000182050 Jun 2000 JP
2000185031 Jul 2000 JP
2000194972 Jul 2000 JP
2000237167 Sep 2000 JP
2000242788 Sep 2000 JP
2000259817 Sep 2000 JP
2000356059 Dec 2000 JP
2000357232 Dec 2000 JP
2001005948 Jan 2001 JP
2001067399 Mar 2001 JP
2001101429 Apr 2001 JP
2001167275 Jun 2001 JP
2001222661 Aug 2001 JP
2001292981 Oct 2001 JP
2001297177 Oct 2001 JP
2001358987 Dec 2001 JP
2002119477 Apr 2002 JP
2002133415 May 2002 JP
2002153444 May 2002 JP
2002153445 May 2002 JP
2002260071 Sep 2002 JP
2002271689 Sep 2002 JP
2002286650 Oct 2002 JP
2002312772 Oct 2002 JP
2002329204 Nov 2002 JP
2003006628 Jan 2003 JP
2003036434 Feb 2003 JP
2003108720 Apr 2003 JP
2003108983 Apr 2003 JP
2003132355 May 2003 JP
2003150942 May 2003 JP
2003153880 May 2003 JP
2003242125 Aug 2003 JP
2003271565 Sep 2003 JP
2003271940 Sep 2003 JP
2003308522 Oct 2003 JP
2003308523 Oct 2003 JP
2003317102 Nov 2003 JP
2003331265 Nov 2003 JP
2004005167 Jan 2004 JP
2004021406 Jan 2004 JP
2004030334 Jan 2004 JP
2004038305 Feb 2004 JP
2004094575 Mar 2004 JP
2004152046 May 2004 JP
2004163356 Jun 2004 JP
2004164483 Jun 2004 JP
2004171350 Jun 2004 JP
2004171602 Jun 2004 JP
2004206444 Jul 2004 JP
2004220376 Aug 2004 JP
2004261515 Sep 2004 JP
2004280221 Oct 2004 JP
2004280547 Oct 2004 JP
2004287621 Oct 2004 JP
2004315127 Nov 2004 JP
2004318248 Nov 2004 JP
2005004524 Jan 2005 JP
2005011207 Jan 2005 JP
2005025577 Jan 2005 JP
2005038257 Feb 2005 JP
2005062990 Mar 2005 JP
2005115961 Apr 2005 JP
2005148883 Jun 2005 JP
WO 9717674 May 1997 WO
WO 9802083 Jan 1998 WO
WO 9808439 Mar 1998 WO
WO 9932317 Jul 1999 WO
WO 9952422 Oct 1999 WO
WO 9965175 Dec 1999 WO
WO 0028484 May 2000 WO
WO 0029986 May 2000 WO
WO 0031677 Jun 2000 WO
WO 0036605 Jun 2000 WO
WO 0101329 Jan 2001 WO
WO 0103100 Jan 2001 WO
WO 0128476 Apr 2001 WO
WO 0135348 May 2001 WO
WO 0135349 May 2001 WO
0140982 Jun 2001 WO
WO 0163994 Aug 2001 WO
WO 0169490 Sep 2001 WO
WO 0186599 Nov 2001 WO
WO 0201451 Jan 2002 WO
WO 0219030 Mar 2002 WO
WO 0235452 May 2002 WO
WO 0235480 May 2002 WO
WO 02091735 Nov 2002 WO
WO 02095657 Nov 2002 WO
WO 03002387 Jan 2003 WO
WO 03054777 Jul 2003 WO
WO 03077077 Sep 2003 WO
WO 2004029863 Apr 2004 WO
WO 2004042646 May 2004 WO
WO 2004055737 Jul 2004 WO
WO 2004089214 Oct 2004 WO
WO 2004097743 Nov 2004 WO
WO 2005008567 Jan 2005 WO
WO 2005013181 Feb 2005 WO
WO 2005024698 Mar 2005 WO
WO 2005024708 Mar 2005 WO
WO 2005024709 Mar 2005 WO
WO 2005029388 Mar 2005 WO
WO 2005062235 Jul 2005 WO
WO 2005069252 Jul 2005 WO
WO 2005093681 Oct 2005 WO
WO 2005096962 Oct 2005 WO
WO 2005098531 Oct 2005 WO
WO 2005104704 Nov 2005 WO
WO 2005109344 Nov 2005 WO
WO 2006023046 Mar 2006 WO
WO 2006063076 Jun 2006 WO
Related Publications (1)
Number Date Country
20070276853 A1 Nov 2007 US
Provisional Applications (2)
Number Date Country
60778770 Mar 2006 US
60647270 Jan 2005 US
Continuation in Parts (7)
Number Date Country
Parent 11275703 Jan 2006 US
Child 11681751 US
Parent 11043366 Jan 2005 US
Child 11275703 US
Parent 11372854 Mar 2006 US
Child 11043366 US
Parent 11672108 Feb 2007 US
Child 11372854 US
Parent 11675424 Feb 2007 US
Child 11672108 US
Parent 11681614 Mar 2007 US
Child 11675424 US
Parent 11681662 Mar 2007 US
Child 11681614 US