Partial fingerprint scanners are becoming popular for a wide variety of security applications. In contrast to “all at once” fingerprint scanners, which capture an image of an entire fingerprint at the same time, partial fingerprint sensing devices use a sensing area that is smaller than the fingerprint area to be imaged. By imaging only a portion of a fingerprint at any given time, the size and cost of a partial fingerprint sensor can be made considerably smaller and cheaper than that of a full fingerprint sensor. However to capture a full fingerprint image, the user must move his finger and “swipe” it across the sensing zone of the partial finger print sensor.
Various types of partial fingerprint readers exist. Some work by optical means, some by pressure sensor means, and others by capacitance sensing means or radiofrequency sensing means.
For example, one common configuration used for a fingerprint sensor is a one or two dimensional array of CCD (charge coupled devices) or C-MOS circuit sensor elements (pixels). These components are embedded in a sensing surface to form a matrix of pressure sensing elements that generate signals in response to pressure applied to the surface by a finger. These signals are read by a processor and used to reconstruct the fingerprint of a user and to verify identification.
Other devices include one or two dimensional arrays of optical sensors that read light reflected off of a person's finger and onto an array of optical detectors. The reflected light is converted to a signal that defines the fingerprint of the finger analyzed and is used to reconstruct the fingerprint and to verify identification.
Many types of partial fingerprint scanners are comprised of linear (1 dimensional) arrays of sensing elements (pixels). These one dimensional sensors create a two dimensional image of a fingerprint through the relative motion of the finger pad relative to the sensor array.
One class of partial fingerprint sensors that are particularly useful for small device applications are deep finger penetrating radio frequency (RF) based sensors. These are described in U.S. Pat. Nos. 7,099,496; 7,146,024; and patent application Ser. Nos. 11/107,682; 11/112,338; 11,243,100; 11/184,464, and the contents of these patents and patent applications are incorporated herein by reference. These types of sensors are commercially produced by Validity Sensors, Inc, San Jose Calif. This class of sensor mounts the sensing elements (usually arranged in a one dimensional array) on a thin, flexible, and environmentally robust support, and the IC used to drive the sensor in a protected location some distance away from the sensing zone. Such sensors are particularly advantageous in applications where small sensor size and sensor robustness are critical.
The Validity fingerprint sensors measure the intensity of electric fields conducted by finger ridges and valleys, such as deep finger penetrating radio frequency (RF) based sensing technology, and use this information to sense and create the fingerprint image. These devices create sensing elements by creating a linear array composed of many miniature excitation electrodes, spaced at a high density, such as a density of approximately 500 electrodes per inch. The tips of these electrodes are separated from a single sensing electrode by a small sensor gap. The electrodes are electrically excited in a progressive scan pattern and the ridges and valleys of a finger pad alter the electrical properties (usually the capacitive properties) of the excitation electrode—sensing electrode interaction, and this in turn creates a detectable electrical signal. The electrodes and sensors are mounted on thin flexible printed circuit support, and these electrodes and sensors are usually excited and the sensor read by an integrated circuit chip (scanner chip, driver chip, scan IC) designed for this purpose. The end result is to create a one dimensional “image” of the portion of the finger pad immediately over the electrode array and sensor junction.
As the finger surface is moved across the sensor, portions of the fingerprint are sensed and captured by the device's one dimensional scanner, creating an array of one dimensional images indexed by order of data acquisition, and/or alternatively annotated with additional time and/or finger pad location information. Circuitry, such as a computer processor or microprocessor, then creates a full two-dimensional fingerprint image by creating a mosaic of these one dimensional partial fingerprint images.
Often the processor will then compare this recreated two dimensional full fingerprint, usually stored in working memory, with an authorized fingerprint stored in a fingerprint recognition memory, and determine if there is a match or not. Software to fingerprint matching is disclosed in U.S. Pat. Nos. 7,020,591 and 7,194,392 by Wei et. al., and is commercially available from sources such as Cogent systems, Inc., South Pasadena, Calif.
If the scanned fingerprint matches the record of an authorized user, the processor then usually unlocks a secure area or computer system and allows the user access. This enables various types of sensitive areas and information (financial data, security codes, etc.), to be protected from unauthorized users, yet still be easily accessible to authorized users.
The main drawback of partial fingerprint sensors is that in order to obtain a valid fingerprint scan, the user must swipe his or her finger across the sensor surface in a relatively uniform manner. Unfortunately, due to various human factors issues, this usually isn't possible. In the real world, users will not swipe their fingers with a constant speed. Some will swipe more quickly than others, some may swipe at non-uniform speeds, and some may stop partially through a scan, and then resume. In order to account for this type of variation, modern partial fingerprint sensors often incorporate finger position sensors to determine, relative to the fingerprint sensor, how the overall finger position and speed varies during a finger swipe.
One type of finger position indicator, represented by U.S. Pat. No. 7,146,024, and application Ser. Nos. 11/112,338 and 11/107,682 (the contents of which are incorporated herein by reference) detects relative finger position using a long array of electrical drive plate sensors. These plates sense the bulk of a finger (rather than the fine details of the fingerprint ridges), and thus sense the relative position of the finger relative to the linear array used for fingerprint sensing. A second type of fingerprint position indicator, represented by U.S. patent Ser. No. 11/184,464 (the contents of which are incorporated herein by reference), uses two linear partial fingerprint sensors, located about 400 microns apart. The two linear sensors use the slight timing differences that occur when a fingerprint swipe first hits one sensor and then the other sensor to detect when a fingerprint edge passes over the sensors. This technique can also detect relative speed of passage over the two partial sensors. This type of information can be used to deduce overall finger location during the course of a fingerprint swipe.
In either case, once finger position is known, each of the one-dimensional partial fingerprint images can then be annotated with additional (and optional) time data (time stamp) or finger (finger tip, finger pad, fingerprint location) location data (location stamp). This optional annotation information, which supplements the “order of data acquisition” that would normally be used to keep track of the multiple stored partial fingerprint images in memory, can be used to help to correct distortions (artifacts) when the various one dimensional partial images are assembled into a full two dimensional fingerprint image.
For example, if the user momentarily stops moving the finger during the finger swipe, the system will generate a series of nearly identical partial (one dimensional) fingerprint images. These images will have different orders of acquisition, and differing time stamps, which could confuse a processor when it attempts to create a correct two dimensional full fingerprint image. However if the fingerprint scanner also has a finger position sensor, the finger location data stamp associated with these nearly identical one dimensional partial fingerprint images will provide evidence that the finger stopped because the finger location data linked to these various one-dimensional partial fingerprint images will be almost the same. The computer processor that reassembles the partial fingerprint images into the complete fingerprint image can be instructed or programmed to also analyze the finger position (location) data, and perform appropriate image corrections when the location data shows that the finger paused during a scan.
Fingers and fingertips are not totally rigid, however. Finger prints rest on a surface layer of skin, which in turn rests over a bed of non-rigid tissue, which in turn rests on bone. Finger tips are thus flexible and deformable, and also slightly sticky due to the innate property of skin, sweat glands, and the high surface area imparted by fingerprint ridges. Thus when a fingertip slides over a surface, there is a tendency for portions of the fingertip pad to momentarily adhere to the surface, while the top of the fingertip continues its motion. This is called “stiction”. Eventually (usually within a fraction of a second), further motion of the finger tip and bone creates enough force to disrupt the momentary adhesion (stiction) between the finger pad skin and the surface, and the motion of the finger pad skin continues. However if, during this brief period of arrest (stiction), the fingertip skin is arrested over a partial fingerprint scanner, a short series of nearly identical repeated partial fingerprint images will be collected. These nearly identical images will have different order of acquisition indexes, and different time stamps. Because the tip of the finger may have continued in motion while the skin was arrested over the sensor, the nearly identical images may also have different finger position (location) stamps. However, these different finger location stamps are misleading because during the stiction event, the finger pad skin itself may not have moved.
Thus when a processor connected to the partial fingerprint sensor attempts to reassemble the series of one dimensional images into a complete two dimensional fingerprint, a distorted image will result. This distorted data is undesirable because it can confound security authorization schemes that require high quality, undistorted, fingerprint images. Efficient methods to correct for such fingerprint “stiction” artifacts are thus desirable.
The present invention is a method for operating biometric sensors, such as partial fingerprint scanners, to remove common artifacts, such as stiction, that can create sensor artifacts. The method involves assessing the scanner data, determining if the data exhibits one or more known artifacts (such as stiction), and if so removing the artifacts from the data prior to subsequent data analysis.
In one embodiment, partial fingerprint scanner data is assessed to determine if successive partial (one dimensional) fingerprint images are overly similar to each other. If the degree of successive partial image similarity exceeds a preset threshold, then at least some of the overly similar partial images will be removed (redacted) from the overall image dataset. The complete overall (two dimensional) image is generated from the redacted data set.
This method is particularly useful for creating “intelligent” low-cost, low power, partial fingerprint scanners and scanner driver IC chips. These scanners and IC chips can pre-process the partial fingerprint data that is generated during the course of a finger swipe, and remove stiction artifacts on a real-time or near-real time basis. The method is compatible with simple and low-power electrical circuits, and thus is well suited for embedding into scanner driver IC circuits. After image artifacts are removed, the corrected fingerprint data can then be further analyzed as needed.
In general, any biometric imaging system that operates on a partial image basis, and that may encounter distortions and artifacts due to improper motion of the biological target relative to the sensor, may be corrected by these methods. Examples of such biometric imaging systems include retinal scans, iris scans, hand prints, finger prints, and other types of scans. Although, throughout this disclosure, fingerprint stiction correction is used as an example of an image artifact that can be corrected by the present invention, it should be understood that the present methods are not limited to either stiction artifacts, or fingerprint scans.
The present invention is particularly useful for removing “stiction” artifacts from fingerprint scans recorded by one-dimensional partial fingerprint sensors. As previously discussed, such one dimensional partial fingerprint sensors generally consist of a linear array of individual sensor units, typically packed closely together at high density, such as a density of about 500 sensing elements (dots) per inch. These individual sensing elements can detect the hills and valleys of the fingerprint ridges. Usually these sensors then digitize this information into an array of integers that characterize the relative elevation of the fingerprint ridges along the one dimensional array of the sensor. These integers are often values between 0-255 to give 8 bits of precision, but other values, such as 1-bit, 2-bit, 4-bit, 12-bit, 16-bit, etc. precision are also quite possible.
As previously discussed, in addition to the actual fingerprint sensing elements, commercially available partial fingerprint sensors, such as the Validity VFS131 and VFS201 sensors, also contain additional sensors that can be used to determine relative finger location or motion as well.
The VFS131, by contrast, uses a series of electrical sensing plates to determine the relative position of the bulk of the finger. In contrast to the electrodes used to determine the fine structure of the fingerprint ridges and valleys, these electrical sensing plates sense the location of the bulk of the finger, and are differentially excited as the finger moves up and down over the course of a finger swipe.
The VFS201 finger location sensor works by a different method. Instead of finger location plates, the VFS201 contains two linear fingerprint sensors, each parallel to each other, and each separated by a distance of about 400 microns from the other. These two fingerprint sensors can be used to deduce fingertip position and velocity because a finger tip or ridge will first pass over one sensor slightly before the other. The relative timing difference between a fingerprint pattern arriving at the first sensor, relative to the same pattern, can be used to derive finger speed. By keeping track of this finger speed history, the relative position of the sensor relative to the finger pad can be computed.
By contrast the bottom part of
In this example, the numeric order of the partial image acquisition (each row is a different partial fingerprint image from a linear fingerprint sensor) is the same as the time stamp. (To keep the diagram simple, the numeric order in which the partial images arrive is not shown.) The fingerprint ridges (pixel value of 1) are shown in black to allow these ridges to be more easily visualized. Thus in this example, each linear (one dimensional) partial image is 11 pixels by 1 pixel in size, and in this example, 21 of these linear partial images have been assembled to give a “complete” two dimensional fingerprint scan. In this example, the location on the finger pad where each partial image was obtained (in arbitrary distance units) is shown in the “location” column. This information is normally obtained from a finger position sensor. As can be seen, the full two dimensional fingerprint scan, produced when all 21 partial fingerprint images are assembled, shows a portion of a “loop” fingerprint.
In
The time and finger position (location) stamps that are associated with these partial fingerprint images are shown on the “time” and “location” columns. This time and position stamp data can be associated with the partial (one dimensional) fingerprint images by a variety of standard computer memory linkage methods. For example, each partial fingerprint image may contain a pointer to where the additional information associated with that particular partial image may be found in memory. Alternatively, the image annotation data (such as the time stamp and the finger location stamp) can be simply appended onto each partial fingerprint image, and stored in memory along with each partial fingerprint region. Generally, any linking method that reliably links annotation information with the partial fingerprint image data is adequate for these purposes.
Although the resulting fingerprint image looks distorted, there is enough information available to correct for these distortions. This is because each partial image has a time stamp, and by examining time stamps, a fingerprint reconstruction engine (processor), could, for example, correct for this distortion. For example, since the duration was 2× too long, a reconstruction engine could correct by averaging the results from the two nearest neighbor time stamps. Similarly, because (in this example), the finger position where each particular linear partial fingerprint image was taken was also recorded on a position stamp, a processor could also correct for this distortion by a similar process. Thus although the image is distorted by a slower scan, prior art image correction techniques are sufficient to correct for this distortion.
By contrast, on the left side (420), the finger has suffered from a “stiction artifact” during the scan. As a result of this stiction, a portion of the fingerprint pad has remained over the partial fingerprint sensor (424). However, due to the elastic properties of skin, the tip of the finger has continued to move forward during this time. As a result, neither the time stamp nor the finger position stamp data (428) attached to the partial fingerprint images give information that can be used to correct for this stiction artifact, and any correction processor will be helpless to correct the problem. To solve this problem, a new artifact correction technique is required.
Stiction Removal Method:
The present invention teaches a novel “pattern recognition” way to correct for stiction distortions and artifacts in fingerprint images, and related problems in biometric scans. The invention makes use of the fact that stiction creates a series of repeated or nearly repeated partial fingerprint images, which are usually closely associated together in both time and space. The invention sequentially examines partial fingerprint images, determines if the nearby partial images are “too similar” to each other, and if so deletes or redacts the “too similar” partial images.
To allow the post-redaction partial fingerprint images to be reassembled into a non-distorted, artifact-free complete fingerprint image, the present invention may also adjust the time and location partial image annotation data to correct for the effects of the redaction process. That is, if a gap in the partial image data is caused by the redaction, the present invention may also correct the time or location data associated with the partial fingerprint images to correct for the discontinuities caused by partial image redaction.
The invention can be implemented by a variety of different methods and algorithms, and additionally can be implemented on a wide variety of data manipulation processors, devices and circuitry. A flow chart showing an example of one such method and algorithm is shown in
In this example, stiction redaction is done by a repetitive series of comparison loops (500) and (530). The process starts by making the first partial image a reference image (502), (503), and getting the next partial image (504). These partial images are compared with a comparison function or algorithm (506). If the results are not “too similar” (508), then it is likely that there is not a stiction problem. In this case, the new partial image is not redacted, but instead is saved for later use (510). This new partial image then becomes the new reference image (512), (502), (503). The loop (500) then repeats with the still newer partial image (504) being fetched, and now being compared with the previous new reference partial image (503).
However if the results are “too similar” (508), then stiction is likely to be a problem, because stiction produces a group of identical or near-identical images. In this case, the algorithm would instead enter a stiction redaction loop (530). In this stiction redaction loop, new partial images are retrieved (504), (532) and compared versus the same initial reference image (503) that was used right before the stiction problem was observed. If these newer partial images are still too similar to the reference image (534), they are also redacted. Here the reference partial image (503) is kept constant until the stiction redaction loop advances past the series of partial images that are “too similar” (536). These loops usually then continue until all partial fingerprint images that make up the fingerprint have been analyzed.
The redaction can be done by simply not saving the partial fingerprint images, or discarding the partial fingerprint images. Alternatively, the multiple redacted partial fingerprint images may be averaged or analyzed to produce a single overall partial image that is an idealized version of the redacted region, and this single idealized partial image used as one of the partial image frames in the final complete fingerprint image.
Stiction Detection Algorithms:
A large number of different stiction detection algorithms can be used. Here are a few examples are discussed.
As one example of a stiction detection algorithm in operation, consider the results that would happen in the
In this example, assume that the algorithm from
In one embodiment, a comparison algorithm (506) can be as simple as computing the sum of the absolute value of the differences between each individual time value “5” partial image pixel, and the corresponding pixel in the time value “6” partial image. Here this “sum of differences” is then compared to a similarity criteria value “s”, and the final result is tested to see if it is greater than zero or not. If the final result is zero or negative, the two partial images would be considered to be “too similar”. If the final result is greater than zero, the two partial images are considered to be OK (not too similar). This function is shown mathematically in equation 1 below:
Equation 1:
Many other alternative similarity functions may also be used. For example, similarity functions can be absolute values, root mean functions, root mean square functions, polynomial functions, etc. More elaborate functions where different pixels are differentially weighed, or nearby pixels compared, can also be done. For example, pixels near the center of the image may be more heavily weighted than pixels on the sides of the image.
In the following examples, the similarity value “s” is set to be 1. Note that with this scheme, identical partial fingerprint images would give a similarity value of −1, and partial fingerprint images that differ by only one pixel would give a similarity value of zero.
As shown in table 1 below, the partial image at time row “6” is not “too similar” to the reference partial image at time row “5” because the similarity function gives a value of 4 (difference sum of 5 minus an “s” of 1=4), which is greater than zero.
However at the next step in
Thus the similarity is the sum of the differences (0) minus “s” (1), or −1, which is less than or equal to zero. Thus comparison algorithm (508) concludes that partial image time row “7” suffers from stiction. As a result, partial image time row “7” is redacted. Here the redaction process is simple because, in this example, partial image time row “7” is not saved (510), and in this example would thus be considered to be deleted or redacted. The algorithm continues to use partial image time “6” as the reference partial image (503).
When the algorithm then advances forward and uses partial image time “8” as the new partial image, it will again see that the difference is too small, and again partial image time “8” will be redacted (not saved (510)) as well, and partial image time “6” will continue to serve as the reference image. This state of affairs continues through partial images “9”, “10”, “11”, “12”, “13”, and “14” as well. The comparison algorithm (532), (534) inside the redaction loop (530) (which can be the same algorithm as used in (508), concludes that these images all suffer from stiction, and they are all redacted (510) or “not saved”.
However when the algorithm finally advances to examine new partial image time (15), a different result is obtained. Here reference partial image time row “6” is still being used as the reference partial image (503), and the new image (504) is now the partial image from time row “15”. Here the comparison algorithm once again detects that this particular combination is not “too similar”.
Here the similarity value is 5−1 or 4, which is greater than zero. Thus the test (534) concludes that the two partial images are not too similar, and the algorithm exits the redaction loop (530), (536) at exit point (534). Partial image time row “15” is saved (not redacted) (510), and partial image time row “15” also becomes the new reference image (512), (503).
After the region of stiction has passed, the rest of the image is processed without any stiction redaction. Here is an example of what happens in the next partial image time row “16”.
This region is also sufficiently different from reference image time “15”, so again new partial image time “16” is saved, and new partial image time “16” becomes the new partial image reference image. The algorithm also examines partial images time “17” to “29”, determines that none of the rest of these partial images suffer from stiction, and thus saves (510) the remainder of these images.
In order to help another processor later use these saved partial images to produce a distortion free complete fingerprint image, the algorithm may optionally also correct any location data or time data linked to the various partial images. For example, the location information that is linked to the various partial images could be corrected for errors caused by the stiction event.
In this example, after the algorithm of
To reduce the severity of this problem, the algorithm may optionally make corrections to the location information associated with the various partial images in order to compensate for this effect. Here, this correction would be to simply subtract the apparent distance of the redacted partial images from the location data of the images after the redaction.
One advantage of this method is that it requires a relatively minimal amount of memory, processing circuitry, and electrical power to implement. Instead of storing the complete, and potentially quite large, fingerprint scan in memory, only a small portion of the fingerprint scan need to be stored in memory. Instead of performing image analysis on a large number of finger print data pixels or sensor elements, the processor or circuitry need only to compare a small number of neighboring partial fingerprint images for approximate similarity.
Because the memory, processing circuitry, and algorithms are relatively simple, the stiction redaction method of the present invention may be implemented on a relatively simple circuit. In one embodiment, the processing circuitry needed to implement this stiction reduction method may be implemented or placed on the same integrated circuit chip that is used to drive the partial fingerprint sensor.
This embodiment simplifies overall circuit design, because the circuitry needed to implement stiction reduction need only operate when the fingerprint sensor itself is operated. In one configuration, the stiction reduction circuitry can intercept the partial fingerprint data as it is generated by the partial fingerprint sensor IC driver chip, perform the needed stiction corrections, and the output from the partial fingerprint sensor can then automatically provide the corrected fingerprint data. This simplifies the task of any subsequent processor that might do analysis on the fingerprint image because this image has been “cleaned up”. This in turn creates a faster, lower power, and cheaper system.
In operation, the fingerprint scan IC can correct stiction artifacts in a manner that is almost transparent to any additional processors residing outside of the IC. Stiction artifacts can be quickly and economically removed, and then output (730). Subsequent fingerprint recognition systems will be simplified because there is a higher assurance that distortion free fingerprints will be output by scanner chip (712). Scanner chip (712) can also receive commands (732) from outside circuitry to change its particular image correction algorithms, or even pass (transmit) raw (uncorrected) partial fingerprint image data directly, as needed.
In use, a user finger (728) is swiped across the sensing elements (716), (718), (720), and the fingerprint sensor module (714) of the chip (712) retrieves the data, in this case in either a time-sequential or all at once manner. Here time sequential means that only a part of the biometric data (such as a portion of the finger) is obtained at any one instant of time, and various partial portions of the biometric data can then be assembled to construct the full set of data. Here, for example, partial fingerprint data can be obtained over the time course of the finger swipe and later assembled to construct a complete fingerprint.
Number | Name | Date | Kind |
---|---|---|---|
4151512 | Rigannati et al. | Apr 1979 | A |
4310827 | Asi | Jan 1982 | A |
4353056 | Tsikos | Oct 1982 | A |
4405829 | Rivest et al. | Sep 1983 | A |
4525859 | Bowles et al. | Jun 1985 | A |
4550221 | Mabusth | Oct 1985 | A |
4580790 | Doose | Apr 1986 | A |
4758622 | Gosselin | Jul 1988 | A |
4817183 | Sparrow | Mar 1989 | A |
5076566 | Kriegel | Dec 1991 | A |
5109427 | Yang | Apr 1992 | A |
5140642 | Hsu et al. | Aug 1992 | A |
5305017 | Gerpheide | Apr 1994 | A |
5319323 | Fong | Jun 1994 | A |
5325442 | Knapp | Jun 1994 | A |
5420936 | Fitzpatrick et al. | May 1995 | A |
5422807 | Mitra et al. | Jun 1995 | A |
5456256 | Schneider et al. | Oct 1995 | A |
5543591 | Gillespie et al. | Aug 1996 | A |
5569901 | Bridgelall et al. | Oct 1996 | A |
5623552 | Lane | Apr 1997 | A |
5627316 | De Winter et al. | May 1997 | A |
5650842 | Maase et al. | Jul 1997 | A |
5717777 | Wong et al. | Feb 1998 | A |
5781651 | Hsiao et al. | Jul 1998 | A |
5801681 | Sayag | Sep 1998 | A |
5818956 | Tuli | Oct 1998 | A |
5838306 | O'Connor et al. | Nov 1998 | A |
5848176 | Hara et al. | Dec 1998 | A |
5850450 | Schweitzer et al. | Dec 1998 | A |
5852670 | Setlak et al. | Dec 1998 | A |
5864296 | Upton | Jan 1999 | A |
5887343 | Salatino et al. | Mar 1999 | A |
5892824 | Beatson et al. | Apr 1999 | A |
5903225 | Schmitt et al. | May 1999 | A |
5915757 | Tsuyama et al. | Jun 1999 | A |
5920384 | Borza | Jul 1999 | A |
5920640 | Salatino et al. | Jul 1999 | A |
5940526 | Setlak et al. | Aug 1999 | A |
5999637 | Toyoda et al. | Dec 1999 | A |
6002815 | Immega et al. | Dec 1999 | A |
6016355 | Dickinson et al. | Jan 2000 | A |
6052475 | Upton | Apr 2000 | A |
6067368 | Setlak et al. | May 2000 | A |
6073343 | Petrick et al. | Jun 2000 | A |
6076566 | Lowe | Jun 2000 | A |
6088585 | Schmitt et al. | Jul 2000 | A |
6098175 | Lee | Aug 2000 | A |
6134340 | Hsu et al. | Oct 2000 | A |
6157722 | Lerner et al. | Dec 2000 | A |
6161213 | Lofstrom | Dec 2000 | A |
6175407 | Santor | Jan 2001 | B1 |
6182076 | Yu et al. | Jan 2001 | B1 |
6182892 | Angelo et al. | Feb 2001 | B1 |
6185318 | Jain et al. | Feb 2001 | B1 |
6234031 | Suga | May 2001 | B1 |
6241288 | Bergenek et al. | Jun 2001 | B1 |
6259108 | Antonelli et al. | Jul 2001 | B1 |
6289114 | Mainguet | Sep 2001 | B1 |
6317508 | Kramer et al. | Nov 2001 | B1 |
6320394 | Tartagni | Nov 2001 | B1 |
6332193 | Glass et al. | Dec 2001 | B1 |
6333989 | Borza | Dec 2001 | B1 |
6337919 | Dunton | Jan 2002 | B1 |
6346739 | Lepert et al. | Feb 2002 | B1 |
6347040 | Fries et al. | Feb 2002 | B1 |
6360004 | Akizuki | Mar 2002 | B1 |
6362633 | Tartagni | Mar 2002 | B1 |
6392636 | Ferrari et al. | May 2002 | B1 |
6399994 | Shobu | Jun 2002 | B2 |
6400836 | Senior | Jun 2002 | B2 |
6408087 | Kramer | Jun 2002 | B1 |
6473072 | Comiskey et al. | Oct 2002 | B1 |
6509501 | Eicken et al. | Jan 2003 | B2 |
6539101 | Black | Mar 2003 | B1 |
6580816 | Kramer et al. | Jun 2003 | B2 |
6597289 | Sabatini | Jul 2003 | B2 |
6643389 | Raynal et al. | Nov 2003 | B1 |
6672174 | Deconde et al. | Jan 2004 | B2 |
6738050 | Comiskey et al. | May 2004 | B2 |
6741729 | Bjorn et al. | May 2004 | B2 |
6757002 | Oross et al. | Jun 2004 | B1 |
6766040 | Catalano et al. | Jul 2004 | B1 |
6785407 | Tschudi et al. | Aug 2004 | B1 |
6838905 | Doyle | Jan 2005 | B1 |
6886104 | McClurg et al. | Apr 2005 | B1 |
6897002 | Teraoka et al. | May 2005 | B2 |
6898299 | Brooks | May 2005 | B1 |
6924496 | Manansala | Aug 2005 | B2 |
6937748 | Schneider et al. | Aug 2005 | B1 |
6941001 | Bolle et al. | Sep 2005 | B1 |
6941810 | Okada | Sep 2005 | B2 |
6950540 | Higuchi | Sep 2005 | B2 |
6959874 | Bardwell | Nov 2005 | B2 |
6963626 | Shaeffer et al. | Nov 2005 | B1 |
6970584 | O'Gorman et al. | Nov 2005 | B2 |
6980672 | Saito et al. | Dec 2005 | B2 |
6983882 | Cassone | Jan 2006 | B2 |
7013030 | Wong et al. | Mar 2006 | B2 |
7020591 | Wei et al. | Mar 2006 | B1 |
7030860 | Hsu et al. | Apr 2006 | B1 |
7035443 | Wong | Apr 2006 | B2 |
7042535 | Katoh et al. | May 2006 | B2 |
7043061 | Hamid et al. | May 2006 | B2 |
7043644 | DeBruine | May 2006 | B2 |
7046230 | Zadesky et al. | May 2006 | B2 |
7064743 | Nishikawa | Jun 2006 | B2 |
7099496 | Benkley | Aug 2006 | B2 |
7110577 | Tschud | Sep 2006 | B1 |
7113622 | Hamid | Sep 2006 | B2 |
7126389 | McRae et al. | Oct 2006 | B1 |
7129926 | Mathiassen et al. | Oct 2006 | B2 |
7136514 | Wong | Nov 2006 | B1 |
7146024 | Benkley | Dec 2006 | B2 |
7146026 | Russon et al. | Dec 2006 | B2 |
7146029 | Manansala | Dec 2006 | B2 |
7190816 | Mitsuyu et al. | Mar 2007 | B2 |
7194392 | Tuken et al. | Mar 2007 | B2 |
7194393 | Wei | Mar 2007 | B2 |
7197168 | Russo | Mar 2007 | B2 |
7200250 | Chou | Apr 2007 | B2 |
7251351 | Mathiassen et al. | Jul 2007 | B2 |
7258279 | Schneider et al. | Aug 2007 | B2 |
7260246 | Fujii | Aug 2007 | B2 |
7289649 | Walley et al. | Oct 2007 | B1 |
7290323 | Deconde et al. | Nov 2007 | B2 |
7308121 | Mathiassen et al. | Dec 2007 | B2 |
7308122 | McClurg et al. | Dec 2007 | B2 |
7321672 | Sasaki et al. | Jan 2008 | B2 |
7356169 | Hamid | Apr 2008 | B2 |
7360688 | Harris | Apr 2008 | B1 |
7369685 | DeLeon | May 2008 | B2 |
7379569 | Chikazawa et al. | May 2008 | B2 |
7409876 | Ganapathi et al. | Aug 2008 | B2 |
7412083 | Takahashi | Aug 2008 | B2 |
7424618 | Roy et al. | Sep 2008 | B2 |
7447339 | Mimura et al. | Nov 2008 | B2 |
7447911 | Chou et al. | Nov 2008 | B2 |
7460697 | Erhart et al. | Dec 2008 | B2 |
7463756 | Benkley | Dec 2008 | B2 |
7505611 | Fyke | Mar 2009 | B2 |
7505613 | Russo | Mar 2009 | B2 |
7565548 | Fiske et al. | Jul 2009 | B2 |
7574022 | Russo | Aug 2009 | B2 |
7643950 | Getzin et al. | Jan 2010 | B1 |
7646897 | Fyke | Jan 2010 | B2 |
7681232 | Nordentoft et al. | Mar 2010 | B2 |
7689013 | Shinzaki | Mar 2010 | B2 |
7706581 | Drews et al. | Apr 2010 | B2 |
7733697 | Picca et al. | Jun 2010 | B2 |
7751601 | Benkley | Jul 2010 | B2 |
7843438 | Onoda | Nov 2010 | B2 |
7899216 | Watanabe et al. | Mar 2011 | B2 |
7953258 | Dean et al. | May 2011 | B2 |
8005276 | Dean et al. | Aug 2011 | B2 |
8031916 | Abiko et al. | Oct 2011 | B2 |
8077935 | Geoffroy et al. | Dec 2011 | B2 |
20010026636 | Mainguet | Oct 2001 | A1 |
20010030644 | Allport | Oct 2001 | A1 |
20010036299 | Senior | Nov 2001 | A1 |
20010043728 | Kramer et al. | Nov 2001 | A1 |
20020025062 | Black | Feb 2002 | A1 |
20020061125 | Fujii | May 2002 | A1 |
20020064892 | Lepert et al. | May 2002 | A1 |
20020067845 | Griffis | Jun 2002 | A1 |
20020073046 | David | Jun 2002 | A1 |
20020089044 | Simmons et al. | Jul 2002 | A1 |
20020089410 | Janiak et al. | Jul 2002 | A1 |
20020096731 | Wu et al. | Jul 2002 | A1 |
20020122026 | Bergstrom | Sep 2002 | A1 |
20020126516 | Jeon | Sep 2002 | A1 |
20020133725 | Roy et al. | Sep 2002 | A1 |
20020181749 | Matsumoto et al. | Dec 2002 | A1 |
20030002717 | Hamid | Jan 2003 | A1 |
20030002719 | Hamid et al. | Jan 2003 | A1 |
20030021495 | Cheng | Jan 2003 | A1 |
20030035570 | Benkley | Feb 2003 | A1 |
20030068072 | Hamid | Apr 2003 | A1 |
20030076301 | Tsuk et al. | Apr 2003 | A1 |
20030076303 | Huppi | Apr 2003 | A1 |
20030095096 | Robbin et al. | May 2003 | A1 |
20030102874 | Lane et al. | Jun 2003 | A1 |
20030123714 | O'Gorman et al. | Jul 2003 | A1 |
20030123715 | Uchida | Jul 2003 | A1 |
20030141959 | Keogh et al. | Jul 2003 | A1 |
20030147015 | Katoh et al. | Aug 2003 | A1 |
20030161510 | Fujii | Aug 2003 | A1 |
20030161512 | Mathiassen et al. | Aug 2003 | A1 |
20030169228 | Mathiassen et al. | Sep 2003 | A1 |
20030174871 | Yoshioka et al. | Sep 2003 | A1 |
20030186157 | Teraoka et al. | Oct 2003 | A1 |
20030209293 | Sako et al. | Nov 2003 | A1 |
20030224553 | Manansala | Dec 2003 | A1 |
20040012773 | Puttkammer | Jan 2004 | A1 |
20040022001 | Chu et al. | Feb 2004 | A1 |
20040042642 | Bolle et al. | Mar 2004 | A1 |
20040050930 | Rowe | Mar 2004 | A1 |
20040066613 | Leitao | Apr 2004 | A1 |
20040076313 | Bronstein et al. | Apr 2004 | A1 |
20040081339 | Benkley | Apr 2004 | A1 |
20040096086 | Miyasaka et al. | May 2004 | A1 |
20040113956 | Bellwood et al. | Jun 2004 | A1 |
20040120400 | Linzer | Jun 2004 | A1 |
20040125993 | Zhao et al. | Jul 2004 | A1 |
20040129787 | Saito | Jul 2004 | A1 |
20040136612 | Meister et al. | Jul 2004 | A1 |
20040172339 | Snelgrove et al. | Sep 2004 | A1 |
20040179718 | Chou | Sep 2004 | A1 |
20040184641 | Nagasaka et al. | Sep 2004 | A1 |
20040190761 | Lee | Sep 2004 | A1 |
20040208346 | Baharav et al. | Oct 2004 | A1 |
20040208347 | Baharav et al. | Oct 2004 | A1 |
20040208348 | Baharav et al. | Oct 2004 | A1 |
20040213441 | Tschudi | Oct 2004 | A1 |
20040215689 | Dooley et al. | Oct 2004 | A1 |
20040228505 | Sugimoto | Nov 2004 | A1 |
20040228508 | Shigeta | Nov 2004 | A1 |
20040240712 | Rowe et al. | Dec 2004 | A1 |
20040252867 | Lan et al. | Dec 2004 | A1 |
20050031174 | Ryhanen et al. | Feb 2005 | A1 |
20050036665 | Higuchi | Feb 2005 | A1 |
20050047485 | Khayrallah et al. | Mar 2005 | A1 |
20050100196 | Scott et al. | May 2005 | A1 |
20050109835 | Jacoby et al. | May 2005 | A1 |
20050110103 | Setlak | May 2005 | A1 |
20050111708 | Chou | May 2005 | A1 |
20050123176 | Ishii et al. | Jun 2005 | A1 |
20050136200 | Durell et al. | Jun 2005 | A1 |
20050139656 | Arnouse | Jun 2005 | A1 |
20050162402 | Watanachote | Jul 2005 | A1 |
20050169503 | Howell et al. | Aug 2005 | A1 |
20050210271 | Chou et al. | Sep 2005 | A1 |
20050219200 | Weng | Oct 2005 | A1 |
20050231213 | Chou et al. | Oct 2005 | A1 |
20050238212 | Du et al. | Oct 2005 | A1 |
20050244038 | Benkley, III | Nov 2005 | A1 |
20050244039 | Geoffroy et al. | Nov 2005 | A1 |
20050249386 | Juh | Nov 2005 | A1 |
20050258952 | Utter et al. | Nov 2005 | A1 |
20050269402 | Spitzer et al. | Dec 2005 | A1 |
20060006224 | Modi | Jan 2006 | A1 |
20060055500 | Burke et al. | Mar 2006 | A1 |
20060066572 | Yumoto et al. | Mar 2006 | A1 |
20060078176 | Abiko et al. | Apr 2006 | A1 |
20060083411 | Benkley | Apr 2006 | A1 |
20060110537 | Huang et al. | May 2006 | A1 |
20060140461 | Kim et al. | Jun 2006 | A1 |
20060144953 | Takao | Jul 2006 | A1 |
20060170528 | Funushige et al. | Aug 2006 | A1 |
20060187200 | Martin | Aug 2006 | A1 |
20060210082 | Devadas et al. | Sep 2006 | A1 |
20060214512 | Iwata | Sep 2006 | A1 |
20060239514 | Watanabe et al. | Oct 2006 | A1 |
20060249008 | Luther | Nov 2006 | A1 |
20060259873 | Mister | Nov 2006 | A1 |
20060261174 | Zellner et al. | Nov 2006 | A1 |
20060271793 | Devadas et al. | Nov 2006 | A1 |
20060287963 | Steeves et al. | Dec 2006 | A1 |
20070031011 | Erhart et al. | Feb 2007 | A1 |
20070036400 | Watanabe et al. | Feb 2007 | A1 |
20070057763 | Blattner et al. | Mar 2007 | A1 |
20070067828 | Bychkov | Mar 2007 | A1 |
20070076926 | Schneider et al. | Apr 2007 | A1 |
20070076951 | Tanaka et al. | Apr 2007 | A1 |
20070086634 | Setlak et al. | Apr 2007 | A1 |
20070090312 | Stallinga et al. | Apr 2007 | A1 |
20070138299 | Mitra | Jun 2007 | A1 |
20070198141 | Moore | Aug 2007 | A1 |
20070198435 | Siegal et al. | Aug 2007 | A1 |
20070228154 | Tran | Oct 2007 | A1 |
20070237366 | Maletsky | Oct 2007 | A1 |
20070248249 | Stoianov | Oct 2007 | A1 |
20080002867 | Mathiassen et al. | Jan 2008 | A1 |
20080013805 | Sengupta et al. | Jan 2008 | A1 |
20080019578 | Saito et al. | Jan 2008 | A1 |
20080049987 | Champagne et al. | Feb 2008 | A1 |
20080049989 | Iseri et al. | Feb 2008 | A1 |
20080063245 | Benkley et al. | Mar 2008 | A1 |
20080069412 | Champagne et al. | Mar 2008 | A1 |
20080126260 | Cox et al. | May 2008 | A1 |
20080169345 | Keane et al. | Jul 2008 | A1 |
20080170695 | Adler et al. | Jul 2008 | A1 |
20080175450 | Scott | Jul 2008 | A1 |
20080178008 | Takahashi et al. | Jul 2008 | A1 |
20080179112 | Qin et al. | Jul 2008 | A1 |
20080185429 | Saville | Aug 2008 | A1 |
20080201265 | Hewton | Aug 2008 | A1 |
20080205714 | Benkley et al. | Aug 2008 | A1 |
20080219521 | Benkley et al. | Sep 2008 | A1 |
20080222049 | Loomis et al. | Sep 2008 | A1 |
20080223925 | Saito et al. | Sep 2008 | A1 |
20080226132 | Gardner | Sep 2008 | A1 |
20080240523 | Benkley et al. | Oct 2008 | A1 |
20080244277 | Orsini et al. | Oct 2008 | A1 |
20080267462 | Nelson et al. | Oct 2008 | A1 |
20080279373 | Erhart et al. | Nov 2008 | A1 |
20090130369 | Huang et al. | May 2009 | A1 |
20090153297 | Gardner | Jun 2009 | A1 |
20090155456 | Benkley et al. | Jun 2009 | A1 |
20090169071 | Bond et al. | Jul 2009 | A1 |
20090174974 | Huang et al. | Jul 2009 | A1 |
20090237135 | Ramaraju et al. | Sep 2009 | A1 |
20090252384 | Dean et al. | Oct 2009 | A1 |
20090252385 | Dean et al. | Oct 2009 | A1 |
20090252386 | Dean et al. | Oct 2009 | A1 |
20090279742 | Abiko | Nov 2009 | A1 |
20090319435 | Little et al. | Dec 2009 | A1 |
20090324028 | Russo | Dec 2009 | A1 |
20100026451 | Erhart et al. | Feb 2010 | A1 |
20100045705 | Vertegaal et al. | Feb 2010 | A1 |
20100083000 | Kesanupalli et al. | Apr 2010 | A1 |
20100119124 | Satyan | May 2010 | A1 |
20100123675 | Ippel | May 2010 | A1 |
20100127366 | Bond et al. | May 2010 | A1 |
20100176823 | Thompson et al. | Jul 2010 | A1 |
20100176892 | Thompson et al. | Jul 2010 | A1 |
20100177940 | Thompson et al. | Jul 2010 | A1 |
20100180136 | Thompson et al. | Jul 2010 | A1 |
20100189314 | Benkley et al. | Jul 2010 | A1 |
20100208953 | Gardner et al. | Aug 2010 | A1 |
20100244166 | Shibuta et al. | Sep 2010 | A1 |
20100272329 | Benkley | Oct 2010 | A1 |
20100284565 | Benkley et al. | Nov 2010 | A1 |
20110002461 | Erhart et al. | Jan 2011 | A1 |
20110102567 | Erhart | May 2011 | A1 |
20110102569 | Erhart | May 2011 | A1 |
20110182486 | Valfridsson et al. | Jul 2011 | A1 |
20110214924 | Perezselsky et al. | Sep 2011 | A1 |
20110267298 | Erhart et al. | Nov 2011 | A1 |
20110298711 | Dean et al. | Dec 2011 | A1 |
20110304001 | Erhart et al. | Dec 2011 | A1 |
Number | Date | Country |
---|---|---|
2213813 | Oct 1973 | DE |
0929028 | Jan 1998 | EP |
0905646 | Mar 1999 | EP |
0973123 | Jan 2000 | EP |
1018697 | Jul 2000 | EP |
1139301 | Oct 2001 | EP |
1531419 | May 2005 | EP |
1533759 | May 2005 | EP |
1538548 | Jun 2005 | EP |
1624399 | Feb 2006 | EP |
1939788 | Jul 2008 | EP |
2331613 | May 1999 | GB |
2480919 | Dec 2011 | GB |
04158434 | Jun 1992 | JP |
2005242856 | Sep 2005 | JP |
WO 9003620 | Apr 1990 | WO |
WO 9858342 | Dec 1998 | WO |
WO 9928701 | Jun 1999 | WO |
WO9943258 | Sep 1999 | WO |
WO 0122349 | Mar 2001 | WO |
WO 0194902 | Dec 2001 | WO |
WO 0194902 | Dec 2001 | WO |
WO 0195304 | Dec 2001 | WO |
WO 0211066 | Feb 2002 | WO |
WO 0247018 | Jun 2002 | WO |
WO 0247018 | Jun 2002 | WO |
WO 02061668 | Aug 2002 | WO |
WO 02077907 | Oct 2002 | WO |
WO 03063054 | Jul 2003 | WO |
WO 03075210 | Sep 2003 | WO |
WO 2004066194 | Aug 2004 | WO |
WO 2004066693 | Aug 2004 | WO |
WO 20050104012 | Nov 2005 | WO |
WO 2005106774 | Nov 2005 | WO |
WO 2005106774 | Nov 2005 | WO |
WO 2006040724 | Apr 2006 | WO |
WO 2006041780 | Apr 2006 | WO |
WO 2007011607 | Jan 2007 | WO |
WO 2008033264 | Mar 2008 | WO |
WO 2008033264 | Mar 2008 | WO |
WO 2008033265 | Jun 2008 | WO |
WO 2008033265 | Jun 2008 | WO |
WO 2008137287 | Nov 2008 | WO |
WO 2009002599 | Dec 2008 | WO |
WO 2009002599 | Dec 2008 | WO |
WO 2009029257 | Jun 2009 | WO |
WO 2009079219 | Jun 2009 | WO |
WO 2009079221 | Jun 2009 | WO |
WO 2009079257 | Jun 2009 | WO |
WO 2009079262 | Jun 2009 | WO |
WO 2010034036 | Mar 2010 | WO |
WO 2010036445 | Apr 2010 | WO |
WO 2010143597 | Dec 2010 | WO |
WO 2011053797 | May 2011 | WO |
Number | Date | Country | |
---|---|---|---|
20090154779 A1 | Jun 2009 | US |