Biometric detection using spatial, temporal, and/or spectral techniques

Information

  • Patent Grant
  • 8355545
  • Patent Number
    8,355,545
  • Date Filed
    Thursday, April 10, 2008
    16 years ago
  • Date Issued
    Tuesday, January 15, 2013
    11 years ago
Abstract
A biometric feature detection system is provided according to one embodiment. The system may include a target, an illumination source, a color filter array, a light detector, and a computational unit. The illumination source illuminates a portion of the target with monochromatic light at a large angle of incidence measured from the normal of the target surface. The light detector may be configured to receive light from the target through the color-filter array and provide an image of the target surface. Each pixel of the light detector may correspond to one of a plurality of color filter mosaics such that each pixel detects light associated with a corresponding color filter. The computational unit may be interfaced with at least the light detector. The computational unit may include instructions to monitor the levels of monochromatic light and make proximity, presence and segmentation determinations based on the detected levels of monochromatic light.
Description
BACKGROUND OF THE INVENTION

This application relates generally to biometric feature detection. More specifically, this application relates to spectral, spatial, and/or temporal detection of a biometric feature.


Fingerprint-based biometric sensors are used across a broad range of applications, from law enforcement and civil identification to commercial access control. They are even used in some consumer devices such as laptops and cellular telephones. In at least some of these applications, there is a general need in the art to know when a biometric feature is placed on the sensor and to detect when a foreign object, other than a biometric feature, is placed on the sensor or where a biometric feature is located relative to a target location.


SUMMARY

A biometric feature detection system is provided according to one embodiment. The biometric feature detection system may include a target, an illumination source, a color filter array, a light detector, and a computational unit. In one embodiment, the target includes a target surface configured to receive a biometric feature at the target surface. In one embodiment, the illumination source is configured to illuminate at least a portion of a target surface from below the target surface with monochromatic light at a large angle of incidence measured from the normal of the target surface. The illumination source may include a plurality of illumination sources, such as LEDs. The illumination source may also provide blue light according to another embodiment. According to one embodiment, the color filter array may include a plurality of color mosaics arrayed across the color filter array. The color filter array, for example, may include a Bayer filter or the like. According to one embodiment, the light detector is configured to receive light from the target through the color-filter array and provide an image of the target surface. Moreover, each pixel of the light detector corresponds to one of the plurality of color mosaics such that each pixel detects light associated with corresponding color mosaic. According to one embodiment, the computational unit is interfaced with the light detector. The computation unit may also include instructions to monitor the levels of blue light and compute a mathematical function on the levels of blue light in proportion with levels of other light.


A method for detecting the proximity of a potential biometric feature to a target is also provided according to another embodiment. The method may include illuminating at least a portion of the target and a portion of the area above the target surface from below the target surface with monochromatic light at a large angle of incidence measured from normal of the target surface. Light may then be received at a light detector. This light may include ambient light. A determination is then made whether a potential biometric feature is near the surface target based on the intensity of the monochromatic light received at the light detector relative to the intensity of light other than the monochromatic light received at the light detector. According to another embodiment, the method may also monitor the relative intensity of the monochromatic light relative to the intensity of light other than the monochromatic light, filtering light into a plurality of wavelength bands prior to receiving light at the light detector. According to another embodiment, the method may also filter the light using a color filter array or the like.


A method is provided for identifying when a purported biometric feature is placed on a target surface according to another embodiment. At least a portion of a target surface and a portion of the area above the target surface may be illuminated from below the target surface with monochromatic light. For example, the monochromatic light may be blue light or light less than about 600 nm according to embodiments. The monochromatic light may include a wavelength that is absorbed by blood. Light may be received at a light detector that includes monochromatic light reflected from the purported biometric feature at or near the target surface. The intensity of the monochromatic light received at the light detector is monitored. The method may then identify when the purported biometric feature is placed on the target surface by determining when the intensity of monochromatic light received at the light detector increases due to blood loss in portions of the purported biometric feature.


A biometric feature presence detector is also provided according to another embodiment. The detector may include a target, a monochromatic light source, a light detector and a computational unit. In one embodiment, the target includes a target surface configured to receive a biometric feature at the target surface. In one embodiment, the illumination source is configured to illuminate at least a portion of a target surface from below the target surface with monochromatic light at a large angle of incidence measured from the normal of the target surface. The illumination source may include a plurality of illumination sources, such as LEDs. The illumination source may also provide blue light according to another embodiment. According to one embodiment, the light detector is disposed below the target surface and configured to receive light from the target surface and a portion of the area above the target surface. In one embodiment the computational unit is interfaced with the light detector. The computational unit may perform a variety of functions according to a variety of embodiments. In one embodiment the computational unit includes instructions to monitor the intensity of monochromatic light received at the light detector. In one embodiment the computational unit includes instructions to identify when the purported biometric feature is placed on the target surface by determining when the intensity of monochromatic light received at the light detector increases due to blood loss in portions of the purported biometric feature.


A method for segmenting a biometric feature from the background is also provided according to one embodiment. The method includes illuminating a biometric feature with monochromatic light and filtering light reflected from the biometric feature with a color filter array into a plurality of wavelength bands. In one embodiment, one of said plurality of wavelength bands corresponds with the wavelength of the monochromatic light illuminating the biometric feature. Light reflected from the biometric feature may be filtered by the color filter array and detected at a detector. Each pixel of the detector receives light corresponding to a wavelength band and providing an image of the target. Pixels corresponding with the wavelength of the monochromatic light are compared. Portions of the image corresponding to a purported biometric feature may then be segmented from portions of the image corresponding to the background by comparing the levels of a wavelength band corresponding to monochromatic light with the levels of another wavelength band.


A biometric feature presence detector is also provided according to another embodiment. The biometric feature presence detector may include a target, one or more monochromatic light sources; a color filter array, a light detector and a computational unit. In one embodiment, the target may include a target surface adapted to receive a potential biometric feature. In one embodiment, the monochromatic light source may be disposed below the target surface and illuminates at least a portion of the target surface and a portion of the area above the target surface with monochromatic light. In one embodiment, the light detector is disposed below the color filter array and the target surface. The light detector may be configured to receive light from the target surface and a portion of the area above the target surface through the color filter array. The computational unit may be interfaced with the light detector and may include instructions to segment portions of the image corresponding to a purported biometric feature from portions of the image corresponding to the background by comparing the levels of a wavelength band corresponding to monochromatic light with the levels of another wavelength band.


A biometric sensor is provided according to another embodiment. The biometric sensor may include a target, at least two illumination sources separated by a distance, an imager and a computational device. In one embodiment, the at least two illumination sources alternately illuminate the target and a portion of the area above the target. In one embodiment the imager may be configured to receive light reflected off a purported biometric feature from the at least two illumination sources as the purported biometric feature approaches the target. In one embodiment, the computational device may be coupled with the imager and may include instructions to process images created under each of the alternating illumination schemes. In one embodiment, the computational may also include instructions to make a comparison between the images to determine whether a biometric feature is at the target surface.


A method for determining the proximity of a purported biometric feature relative to a target surface is also provided according to another embodiment. The method may include all or any of the following in any combination: 1) illuminating at least a portion of the target and a portion of the area above the target surface with a first illumination source; 2) capturing a first image of the target area; 3) illuminating at least a portion of the target and a portion of the area above the target surface with a second illumination source, wherein the first and second illumination sources are separated by a distance; 4) capturing a second image of the target area; 5) comparing the intensity profile across a portion of the first image with a portion of the second image; and 6) determining the proximity of the purported biometric feature from the peak intensity separation in the first image and the second image.





BRIEF DESCRIPTION OF THE DRAWINGS


FIG. 1A shows a finger approaching a biometric sensor under stereoscopic illumination according to one embodiment.



FIGS. 1B-1D show graphs of the finger as recorded by the biometric sensor under the illumination configuration shown in FIG. 1A according to one embodiment.



FIG. 2A shows a finger approaching a biometric sensor under stereoscopic illumination according to one embodiment.



FIGS. 2B-2D show graphs of the finger as recorded by the biometric sensor under the illumination configuration shown in FIG. 2A according to one embodiment.



FIG. 3A shows a finger approaching a biometric sensor under stereoscopic illumination according to one embodiment.



FIGS. 3B-3D show graphs of the finger as recorded by the biometric sensor under the illumination configuration shown in FIG. 3A according to one embodiment.



FIG. 4A shows a finger approaching a biometric sensor under stereoscopic illumination according to one embodiment.



FIGS. 4B-4D show graphs of the finger as recorded by the biometric sensor under the illumination configuration shown in FIG. 4A according to one embodiment.



FIG. 5 shows a finger with two viewable cross sections shown according to one embodiment.



FIG. 6 shows a graph comparing the results of a biometric feature compared with a flat feature according to one embodiment.



FIG. 7 shows a flowchart showing a method to determine whether an expected biometric feature is on the platen according to one embodiment.



FIG. 8 shows another flowchart showing a method of determining whether an object on or near the biometric site is a biometric feature according to one embodiment.



FIG. 9A shows a block diagram of a biometric sensor system according to one embodiment.



FIG. 9B is a schematic representation of a computer system that may be used to manage functionality of various embodiments of the invention.



FIG. 10 shows another biometric detection system 1000 according one embodiment of the invention.



FIG. 11 shows a Bayer color filter array.



FIGS. 12A-12C show an example of how a Bayer filter array works under ideal conditions.



FIG. 13 shows a typical color response curve for a typical Bayer filter array.



FIG. 14 shows a flowchart depicting an embodiment of the invention for determining the proximity of a biometric feature to a target.



FIG. 15 shows a flowchart depicting an embodiment of the invention for determining whether a biometric feature is pressed with the surface of a target.



FIG. 16 shows a flowchart depicting an embodiment of the invention for segmenting the portions of an image associated with a biometric feature from background features.



FIG. 17 shows a flowchart combining the embodiments shown in FIGS. 14-16 according to another embodiment.



FIG. 18A-18D shows a system for performing proximity, presence and/or segmentation of a biometric feature according to one embodiment.





DETAILED DESCRIPTION OF THE INVENTION

Embodiments are provided throughout this disclosure that provide proximity, presence, and/or segmentation detection of a biometric feature at a target. The various embodiments may be used in conjunction, for example, with a fingerprint verification and/or identification system, retinal scanners, handprint verification and/or identification system, and/or the like. Proximity detection refers generally to determining the relative nearness of a biometric feature, such as a skin site on a finger, to a target surface. Presence detection refers generally to determining whether a biometric feature is present at a target surface. Segmentation refers generally to segmenting the portions of a biometric feature from the background and/or other ambient features in an image including a biometric feature. Segmentation may occur at anytime. In some embodiments, segmentation may occur in real time, that is, prior to, during or immediately after image capture.


A biometric system is disclosed according to one embodiment that includes a target configured to receive a biometric feature for identification. The target may be a transparent plate, platen, prism, glass, plastic, etc. The target may include a surface designed to receive a biometric feature, such as a skin site. One or more monochromatic blue light sources may be arranged circumferentially around the target. Various other light sources may be used with various wavelengths. The light sources may illuminate the target surface at an angle of about 60° from the normal of the target. The light sources may be below the target surface. A detector coupled with a color filter array, such as, for example, a Bayer filter may be used to filter light received from the target. Such light may be ambient light and/or monochromatic light reflected from objects near the target. The detector in conjunction with the color filter array may provide spectral data across the detector array. The spectral data may be used, according to some embodiments of the invention, to determine the proximity, presence and/or segmentation of a biometric feature. Temporal and/or spatial information may also be used to determine the presence, proximity and/or segmentation of a biometric feature in relation to the target.


Skin sites applicable to the biometric detection described herein include all surfaces and all joints of the fingers and thumbs, the fingernails and nail beds, the palms, the backs of the hands, the wrists and forearms, the face, the eyes, the ears, and all other external surfaces of the body. While the discussion below sometimes makes specific reference to “fingers” in providing examples of specific embodiments, it should be understood that these embodiments are merely exemplary and that other embodiments may use skin sites at other body parts.


According to one embodiment of the invention the proximity and/or presence of a biometric feature relative to a target may be determined by measuring the convergence of intensity peaks in images of a biometric feature from separately illuminated light sources as the biometric feature approaches the biometric sensor. Another embodiment compares biometric images with expected biometric feature characteristics to determine whether a recorded image is from an expected biometric feature or from a rogue source.



FIG. 1A shows a biometric sensor 100 with stereoscopic illumination according to one embodiment. The biometric sensor 100 includes, at least, a body 110, two illumination sources 130 separated by some distance, an imager and a platen 120 that may have a surface for receiving a biometric feature such as finger. A biometric sensor so configured may be used to determine the proper time to record a biometric image and determine whether an expected biometric feature is placed on the platen 120 or is approaching the platen 120.


In the embodiment shown in FIG. 1A, the biometric sensor may be used to read fingerprints from a finger 145 when the finger 145 is placed on the platen 120. While a fingerprint system is described below, those skilled in the art will recognize that the system may be applicable to other biometric systems. As shown, the finger 145 is approaching the platen 120 from some distance. As the finger 145 approaches the platen 120 the two illumination sources 130 alternate illuminating the finger 145. The imager 140 records images of the finger 145 under the alternating illumination schemes. A cross section from each of the alternating images may be compared against each other. More than two illumination sources may be used.



FIGS. 1B and 1C show graphs of the light intensity over a cross section of a biometric feature under illumination from each of the two illumination sources 130 shown in FIG. 1A. FIG. 1B shows a peak intensity towards the right of the cross section and FIG. 1C shows a peak in intensity towards the left of the cross section. The separation in peak intensity is due to the different perspectives from which the illumination sources 130 illuminate the finger 145. FIG. 1D shows the graphs of FIGS. 1B and 1C overlaid showing the peak intensity separation. The separation in intensity peaks shown in FIGS. 1B and 1C are a result of the distance the finger 145 is from the platen 120.



FIG. 2A shows the finger 145 at a position closer to the biometric sensor 100. FIGS. 2B-2D show the corresponding graphs of intensity verses cross section of the finger 145. In FIG. 2D plots of the two cross section images are overlaid and show the two intensity peaks closer together as the finger moves closer to the platen 120. FIG. 3A shows the finger 145 at a position even closer to the biometric sensor 100. FIGS. 3B-3D show the corresponding graphs of intensity verses cross section of the finger 145. In FIG. 3D the two intensity peaks are much closer together.


Finally, in FIG. 4A the finger 145 is resting on the platen 120 of the biometric sensor 145. FIGS. 4B and 4C show intensity peaks under the two illumination schemes. At this point the intensity peaks have substantially converged as shown in FIG. 4D. As the finger 145 approaches the platen 120 of the biometric sensor 100, the stereoscopic nature of the dual illumination sources provides data regarding the position of the finger 145. Accordingly, a system may determine when a biometric feature is at the platen 120 for recording by measuring the convergence of the peak intensity over a cross section of the biometric feature.


In some embodiments, depending on the distance between each of the illumination sources 130 and/or the distances between the illumination sources 130 and the platen 120, the amount of divergence required to know when to record an image may vary. Those skilled in the art will recognize that a simple calibration may be required to determine the ideal amount of convergence that may be used for a specific biometric sensor configuration to show that a biometric feature is on the platen 120. Moreover, calibration may be used to determine the distance of a biometric feature from the surface of the target. For instance, in non-contact biometric schemes, a biometric feature does not need to be in contact with the biometric feature. Accordingly, the proximity detection embodiments described throughout this disclosure may be used to determine the distance of a biometric feature from the surface of the target.



FIG. 5 shows a finger 145 with two exemplary cross sections 510, 520 that may be used by embodiments of the invention. The imager may record an image of the entire finger 145 and then use only the cross section of the finger 145 to determine the position of the finger 145 and/or whether a purported biometric feature on the platen 120 is a biometric feature or not. These cross sections may also be used to determine whether a biometric feature is properly placed on the platen. Using the cross section information, the system may alert a user that the biometric feature must be moved in order to be more properly imaged.


Another embodiment determines whether an object on the platen 120 is an expected biometric feature, such as a finger, a retina, a hand, a face, etc, or a rogue feature. FIG. 6 shows a graph 600 comparing two cross sections of two different objects placed on the platen 120. The flat graph line 620 shows the cross section when a flat object, such as a piece of paper is placed on the platen 120. The peaked graph line 610 shows a curved object, such as a finger 145. By comparing a cross section of the intensity one can determine whether an object on the platen 120 is the expected object or not. For example, a fingerprint sensor would compare cross sections with a general finger cross section to determine whether the object placed on the platen is the expected object. Other biometric feature detection systems may also make similar comparisons.


In another embodiment, multispectral data may be recoded by the imager. The multispectral information may be compared with an expected object's known multispectral response to white light or other illumination sources. For example, the skin on a finger absorbs and reflects light of various unique wavelengths. This known spectral response may be used to determine whether an object on a platen is or is not a finger as expected. Using this information, the biometric sensor will know whether or not to record an image. Accordingly, various embodiments of the invention provide imaging that may or may not use illumination sources other than the illumination sources used for proximity detection, presence detection and/or segmentation.


The imager (or detector) used in embodiments of the invention may comprise a silicon imaging array, such as a CCD or CMOS array, an InGaAs array, or other detector arrays as known in the art. In some instances, the imager may also comprise an optical filter. The optical filter may be a short-wavelength pass filter, which substantially blocks light of wavelengths longer than the illumination wavelength range. Such a may be used in the presence of bright, broad-band ambient lighting, since wavelengths of light longer than approximately 580 nm may substantially traverse the finger. In bright sunlight, this long wavelength light may saturate the detector array preventing the acquisition of an image. Blocking such long-wavelength light with filter while passing all desired illumination wavelengths may thus be beneficial.



FIG. 7 shows a flowchart showing a method to determine whether an expected biometric feature is on the platen 120 according to one embodiment. As shown in the flowchart, the method uses two illumination sources to stereoscopically determine when an object is resting on the platen 120 by following the convergence of histograms of cross sections of the object as recorded by the imager.


The surface of a target may be illuminated with a first illumination source at block 705. A first image is recorded at block 710. The surface of the target may then be illuminated with a second illumination source at block 715. The first and second illumination sources may be located at a distance from each other. The first and second illumination sources may, according to one embodiment, provide light that is incident on the target surface at an angle greater than 60° measured from the normal of the target surface. In another embodiment, the incident angle is 50°, 55°, 65°, 70°, 80°, 85°, etc. The incident angle is measured from the normal of the target surface and the line between the center of the illumination source the center of the illumination source.



FIG. 8 shows another flowchart showing a method of determining whether an object on or near the biometric site is an expected biometric feature according to one embodiment. The method captures a multispectral image of an object at the target surface. The captured image is then compared with predetermined features of an expected biometric feature to determine whether the object is an expected biometric feature or not.


A multispectral image of a purported biometric feature is captured at block 805. The multispectral image may then be compared with general predetermined characteristics associated with the biometric feature at block 810. For example, as shown in FIG. 6, a finger has a different response from the convergence of stereoscopic illumination sources than a flat object, such as paper or rubber with a finger print fabrication. Also, for example, human fingers have different multispectral characteristics than paper or rubber fingerprint fabrication. An eye also has general multispectral characteristics that may be used to distinguish fakes from real eyes.



FIG. 9A shows a block diagram 900 of a biometric sensor system according to one embodiment. The block diagram 900 may be used to implement embodiments of the invention. Illumination sources and/or an imager are coupled with the processor. The processor 905 may receive images from the imager 915 and analyze cross sections to determine whether an object is approaching a platen 120 or whether an object on the platen 120 is an expected object. Data storage 920 may be used to store images received from the imager 915. Characteristics of various biometric features and/or biometric identifiers of individuals may also be stored in the data storage 920. Any type of data storage device may be used. The processor 905 may also be used to control the illumination sources 910. The processor 905 may determine when to turn the illumination sources 910 “on” or “off”. In embodiments where the illumination sources are alternately illuminated, the processor 905 may determine and/or dictate the frequency of illumination and/or the duration of a cycle. The processor 905 may also be used to control illumination sources used for imaging and/or biometric identification.


In various embodiments, a biometric sensor, whether it be a non-contact, contact, or texture sensor of any of the types described above, may be operated by a computational system to implement biometric functionality. This computation system, for example, may be the processor or part of the processor shown in FIG. 9A.



FIG. 9B broadly illustrates how individual system elements may be implemented in a separated or more integrated manner. The computational device 901 is shown comprised of hardware elements that are electrically coupled via bus 926, which is also coupled with the biometric sensor 956. The hardware elements include a processor 902, an input device 904, an output device 906, a storage device 908, a computer-readable storage media reader 912a, a communications system 914, a processing acceleration unit 916 such as a DSP or special-purpose processor, and a memory 918. The computer-readable storage media reader 912a is further connected to a computer-readable storage medium 912b, the combination comprehensively representing remote, local, fixed, and/or removable storage devices plus storage media for temporarily and/or more permanently containing computer-readable information. The communications system 914 may comprise a wired, wireless, modem, and/or other type of interfacing connection and permits data to be exchanged with external devices.


The computational device 901 also may also comprise software elements, shown as being currently located within working memory 919, including an operating system 924 and other code 922, such as a program designed to implement methods of the invention. It will be apparent to those skilled in the art that substantial variations may be used in accordance with specific requirements. For example, customized hardware might also be used and/or particular elements might be implemented in hardware, software (including portable software, such as applets), or both. Further, connection to other computing devices such as network input/output devices may be employed.



FIG. 10 shows another biometric detection system 1000 according one embodiment of the invention. The system includes a target 120 adapted to receive a biometric feature, for example, a finger 145. The target 120 includes a target surface 121 for receiving a biometric feature, such as a finger 145. The target may be any transparent material, such as, for example, glass, plastic, etc. The target may also be multifaceted. As shown in the figure, two illumination sources 130, 135 illuminate the target 120 from beneath the target. In the example describing this embodiment, the illumination sources operate at less than 600 nm. The illumination sources also illuminate the finger 145 placed on the target. In various embodiments of the invention, the illumination sources 130, 135 are monochromatic sources. In other embodiments, the illumination sources 130, 135 provide light with a wavelength less than about 600 nm. In another embodiment, the illumination sources provide blue light. For example, blue light emitting LEDs may be used. As another example, a white light source with a blue light filter may also be used. The illumination sources, according to other embodiments of the invention, may be quasi-monochromatic sources, solid-state LEDs, organic LEDs, laser diodes, other kinds of lasers and/or quasi-monochromatic sources. The sources 507 may further comprise lenses, mirrors, optical diffusers, optical filters and other such optical elements.


Moreover, in other embodiments, more than two illumination sources may be used. The dashed lines show light from the illumination sources reflecting off portions of the finger 145. The dotted lines 1060 show ambient light passing through the target.


The biometric detection system also includes a color filter array 1005, such as, for example, a Bayer filter, between the detector 140 and the target. In some embodiment, the color filter array is on the detector or part of the detector Each pixel of the detector corresponds to a color pattern on the color filter array 1005. Various other color filter arrays may be used, such as, for example, a RGBE, CYMG, RGBW, RGBW #1, RGBW #2, RGBW #3, etc. In certain embodiments, the color filter array filters light by wavelength range, such that the separate filtered intensities include information about the color of light. For example, as shown in FIG. 11, a Bayer filter 1100 gives information about the intensity of light in red 115, green 1110, and blue 1105 wavelength regions. Each mosaic is associated with a pixel on the sensor 140. Thus, spectral information may be gathered corresponding to locations on the detector array. Turning back to FIG. 10, various functions may be performed on the detected images. For example, segmentation may be determined by performing a function 1025 on the light filtered by the blue filter 1015 and the light filtered by the green filter 1010. According to one embodiment, proximity and/or presence detection may, for example, use light filtered by the blue filter 1015 in reference to light filtered by the red and/or green filters 1005, 1010. In one embodiment, the light filtered by the red filter 1005 may be used to measure the ambient transmissions, which may be used as a quality metric. Various other functions may be performed using the filtered image data. For example, the ratio, difference, product, sum, or other linear or non-linear function may be used to determine the presence, proximity and/or segmentation of a biometric feature.



FIGS. 12A-12C show an example of how a Bayer filter array works under ideal conditions. Starting with FIG. 12A, red light 1210, blue light 1220, and green light 1215 are incident on a red mosaic 1245 of a Bayer filter. Red light passes from the red mosaic 1245 to the detector array 1205. Ideally only those portions or pixels of a detector array 1205 below the red mosaics 1210 register when red light is incident on the red mosaic resulting in a red Bayer pattern 1225. Similarly, as shown in FIGS. 12B and 12C, green light 1215 and blue light 1220 incident on green and blue mosaics 1250, 1255 produce green and red Bayer patterns 1225. Of course, in a non-ideal situation there may be some cross talk between pixels in the array. That is, red light may also trigger pixels under the green and/or blue mosaics and vice versa. As seen from the color response curve for a typical Bayer filter array shown in FIG. 13, there may be some overlap in the spectral ranges of the red 1315, green 1310, and blue 1305 transmission characteristics of the filter elements.



FIG. 14 shows a flowchart depicting an embodiment of the invention for determining the proximity of a biometric feature to a target. A target is illuminated with a monochromatic light source at block 1405. The light source may illuminate the target and/or target area with a large angle of incidence, for example, greater than 60° from the normal of the surface of the target. The light source, in various embodiments, may be a monochromatic blue light source or, in another embodiment, the light source may emit monochromatic light less than about 600 nm. The light source or sources illuminate not only the target but also the area immediately above the target from beneath the target. A color filter array filters light from the target area at block 1410 prior to the light being incident on a light detector at block 1415. The color filter array may be any color filter array described in the various embodiments described herein. The color filter array filters light according to wavelength bands. Thus, the relative intensity of a wavelength band may be compared with other wavelength bands. As a purported biometric feature approaches the target, monochromatic light is reflected from the surface of the biometric feature. Accordingly, the relative intensity of the wavelength band containing the wavelength of the monochromatic light increases relative to other frequency bands as the purported biometric feature approaches the target.


Accordingly, in the embodiment shown in the flowchart, the intensity of blue light is monitored relative to the intensity of red and/or green light at block 1420. The method then determines if the intensity of blue light relative to the intensity of red and/or green light increases at block 1425. If the blue light intensity increases, then a purported biometric feature is proximate to the target at block 1430. If the blue light intensity does not increase enough, than there is no purported biometric feature proximate to the target and the system continues to monitor the intensity levels of various wavelength bands.



FIG. 15 shows a flowchart depicting an embodiment of the invention for determining whether a biometric feature is pressed with the surface of a target. Blocks 1405, 1410, 1415 are similar to those shown in FIG. 14. At block 1520, the blue light intensity is monitored. When a biometric feature, such as a finger, is pressed against surface, the blood is forced from the general area of contact. Blood absorbs blue light. Therefore, as blood is forced from the area of contact, blue light absorption decreases and blue light reflectivity increases at the area of contact. Accordingly, at a light detector, the intensity of blue light may increase as the blood is forced from the area of contact. Once contact has been initiated and maintained the intensity of blue light may plateau. Accordingly, at block 1525, the intensity of blue light is monitored to check if it has peaked and/or plateaued. If the intensity of blue light does not peak and/or plateau, then the system continues to monitor the blue light intensity levels at block 1520. If the intensity of blue light has peaked and/or plateaued, then a purported biometric feature has been pressed with the target surface at block 1530. Such embodiments may also be used to determine whether a purported biometric feature is or is not a non-biological counterfeit.


While blue light is used to describe this embodiment, other wavelengths may be used without deviating from the scope and/or spirit of the invention. For instance, other wavelengths of light may be used to determine whether a biometric feature is pressed against a target surface forcing blood from the area of contact.



FIG. 16 shows a flowchart depicting an embodiment of the invention for segmenting the portions of an image associated with a biometric feature from background features. Blocks 1405, 1410, 1415 are similar to those shown in FIG. 14. At block 1620, the intensity of blue light pixels are compared with the intensity of adjacent green light pixels. Areas where the intensity of blue light is high relative to the intensity of the green light correspond to areas where a biometric feature is located on an image from purported biometric feature at a target surface. In one embodiment, such a determination can be made on a pixel-by-pixel basis. In another embodiment, such a determination can be made based on groups of pixels. Any of various mathematical functions may be used to compare the two intensity levels. For example, the ratio of the blue light intensity versus the adjacent green light intensities may be calculated. As another example, the difference of the two intensities may be calculated. In yet another embodiment of the invention various other wavelength bands may be used besides blue and green to make a segmentation determination. Moreover, these bands may be determined based on the type of color filter array used.



FIG. 17 shows a flowchart combining the embodiments shown in FIGS. 14-16 according to another embodiment. According to this embodiment, blocks 1405, 1410, 1415, 1420 are similar to those shown in FIG. 14. That is, the target is illuminated with monochromatic blue light 1405, light is filtered at a color filter array 1410, and then detected at the light detector 1415. The intensity of blue light relative to the red and/or green light is monitored across the detector array at block 1420. Similar variations discussed in previous embodiments may be applied in association with flowchart shown in FIG. 17. If the blue light intensity is not sufficiently greater than the green and/or red light intensities as determined at block 1425, then the system continues to monitor these intensities at block 1420. In some embodiments, a threshold value may be used to see if the blue light intensity is sufficiently greater than the red or green light intensities. If the blue light intensity is sufficiently greater than the red and/or green light intensities, then a purported biometric feature is within the proximity of the target.


The system may then monitor the blue light intensity to see if it peaks or plateaus at block 1525 according to one embodiment. If it does, then the biometric feature has made contact with the target. If it does not, then the system continues to monitor the blue light intensity. Once the biometric feature has made contact with the target surface, the system may then segment the portions of the detector array corresponding to the biometric feature from the portion of the detector array corresponding to background features by mathematically comparing the intensity of blue light pixels with the intensity of adjacent green light pixels at block 1625. The various other embodiments described in association with FIGS. 14-16 may also be applied to this embodiment of the invention.



FIGS. 18A-18D show various views of a system for performing proximity, presence and/or segmentation of a biometric feature according to one embodiment. FIG. 18A shows a perspective view of the system. FIG. 18B shows a top view and FIGS. 18C and 18D show side views. A target 117 is shown with a human finger 145 placed thereon for biometric detection. The system, according to this embodiment, includes six illumination sources 135 arrayed around the center of the target. In other embodiments of the invention, various other numbers of illumination sources may be used. For example, 2, 3, 4, 5, 7, 8, 9, or 10 illumination sources may be used. Each illumination source 135 may include various optical elements 1805 such as lenses as well as a monochromatic light source 1810. In one embodiment, the light source 1810 includes an LED. As shown in FIG. 18C, the illumination sources 135 are positioned such that they produce light that is incident on the target at a small angle relative to the target surface. This angle may be less than about 30°. In some embodiments, the angle is between about 20° and 25°. The system also includes a detector 140 with a color array filter 1005. The optical system also includes a number of optional optical elements 1815, 1820, 1825 above the detector 140.

Claims
  • 1. A biometric feature detection system comprising: a target including a target surface configured to receive a biometric feature at the target surface;an illumination source configured to illuminate at least a portion of a target surface from below the target surface with monochromatic blue light at a large angle of incidence measured from the normal of the target surface;a color filter array comprising a plurality of color mosaics arrayed across the color filter array;a light detector configured to receive light from the target through the color-filter array and provide an image of the target surface, wherein each pixel of the light detector corresponds to one of the plurality of color mosaics such that each pixel detects light associated with corresponding color filter; anda computational unit interfaced with the light detector, wherein the computation unit includes instructions to monitor the levels of blue light and compute a mathematical function on the levels of blue light in proportion with levels of other light.
  • 2. The biometric feature detection system according to claim 1, wherein the color filter array comprises a Bayer filter.
  • 3. The biometric feature detection system according to claim 1, wherein the mathematical function determines whether a potential biometric feature is near the surface target based on the intensity of the blue light received at the light detector relative to the intensity of light other than the blue light.
  • 4. The biometric feature detection system according to claim 1, wherein the mathematical function identifies when the purported biometric feature is placed on the target surface by determining when the intensity of blue light received at the light detector increases due to blood loss in portions of the purported biometric feature.
  • 5. The biometric feature detection system according to claim 1, wherein the mathematical function segments portions of the image corresponding to a purported biometric feature from portions of the image corresponding to the background by comparing the levels of a wavelength band corresponding to blue light with the levels of another wavelength band.
  • 6. A method for detecting the proximity of a potential biometric feature to a target, the method comprising: illuminating at least a portion of the target surface and a portion of the area above the target surface from below the target surface with monochromatic light at a large angle of incidence measured from normal of the target surface;receiving light at a light detector, wherein the light received at the light detector includes ambient light from above the target surface; anddetermining whether a potential biometric feature is near the surface target based on the intensity of the monochromatic light received at the light detector relative to the intensity of light other than the monochromatic light received at the light detector.
  • 7. The method according to claim 6, further comprising monitoring the relative intensity of the monochromatic light relative to the intensity of light other than the monochromatic light.
  • 8. The method according to claim 6, wherein the large angle comprises an angle greater than or equal to about 60°.
  • 9. The method according to claim 6, wherein the monochromatic light comprises blue light, and the light other than the monochromatic light comprises red light.
  • 10. The method according to claim 6, wherein the monochromatic light comprises blue light, and the light other than the monochromatic light comprises green light.
  • 11. The method according to claim 6, wherein the monochromatic light comprises light with a wavelength less than 600 nm.
  • 12. The method according to claim 6, wherein the monochromatic light is illuminated from a plurality of light sources arranged radially about the target.
  • 13. The method according to claim 6, wherein the light received at the light detector includes light reflected from a potential biometric feature in the proximity of the target.
  • 14. The method according to claim 6, further comprising filtering light into a plurality of wavelength bands prior to receiving light at the light detector.
  • 15. A biometric proximity detector, comprising: a target including a target surface adapted to receive a potential biometric feature;at least one monochromatic light source configured to illuminate at least a portion of the target surface and a portion of the area above the target surface from below the target surface with monochromatic light at a large angle of incidence measured from normal of the target surface;a light detector configured to receive light from the target surface and a portion of the area above the target surface, wherein the light received at the light detector includes ambient light from above the target; andmeans for determining whether a potential biometric feature is near the surface target based on the intensity of the monochromatic light received at the light detector relative to the intensity of light other than the monochromatic light received at the light detector.
  • 16. The biometric proximity detector according to claim 15, further comprising a color filter array, wherein the color filter array filters light into spectral bands before the light is received at the light detector.
  • 17. The biometric proximity detector according to claim 16, wherein the color filter array filters light into red, blue and green spectral bands.
  • 18. The biometric proximity detector according to claim 15, further comprising a Bayer filter that filters light into a pattern of red, green and blue light.
  • 19. The biometric proximity detector according to claim 15, wherein the monochromatic light is blue light.
  • 20. The biometric proximity detector according to claim 15, wherein the at least one monochromatic light source comprises at least four monochromatic light sources arranged radially about the surface target.
  • 21. The biometric proximity detector according to claim 15, wherein the large angle comprises an angle greater than or equal to about 60°.
  • 22. The biometric proximity detector according to claim 15, wherein the light source comprises an LED.
  • 23. A method for identifying when a purported biometric feature is placed on a target surface, the method comprising: illuminating at least a portion of the target surface and a portion of the area above the target surface from below the target surface with monochromatic light, wherein the monochromatic light is absorbed by blood;receiving light at a light detector, wherein the light received at the light detector includes monochromatic light reflected from the purported biometric feature at or near the target surface;monitoring the intensity of the monochromatic light received at the light detector; andidentifying when the purported biometric feature is placed on the target surface by determining when the intensity of monochromatic light received at the light detector increases due to blood loss in portions of the purported biometric feature.
  • 24. The method according to claim 23, wherein the monochromatic light comprises blue light.
  • 25. The method according to claim 23, wherein the monochromatic light comprises light with a wavelength less than about 600 nm.
  • 26. A biometric feature presence detector comprising: a target including a target surface adapted to receive a potential biometric feature;a monochromatic light source disposed below the target surface, wherein the monochromatic light source illuminates at least a portion of the target surface and a portion of the area above the target surface with monochromatic light;a light detector disposed below the target surface and configured to receive light from the target surface and a portion of the area above the target surface; anda computational unit interfaced with the light detector and having: instructions to monitor the intensity of monochromatic light received at the light detector; andinstructions to identify when the purported biometric feature is placed on the target surface by determining when the intensity of monochromatic light received at the light detector increases due to blood loss in portions of the purported biometric feature.
  • 27. The biometric feature presence detector according to claim 26, wherein the monochromatic light comprises light with a wavelength less than or equal to about 600 nm.
  • 28. A biometric feature presence detector comprising: a target including a target surface adapted to receive a potential biometric feature;means for illuminating at least a portion of the target surface and a portion of the area above the target surface with monochromatic light originating from below the target surface;a light detector configured to receive light from the target surface and a portion of the area above the target surface; andmeans for identifying when the purported biometric feature is placed on the target surface by determining when the intensity of monochromatic light received at the light detector increases due to blood loss in portions of the purported biometric feature.
  • 29. The biometric feature presence detector according to claim 28, wherein the monochromatic light comprises light with a wavelength less than or equal to about 600 nm.
  • 30. A method for segmenting a biometric feature from the background, the method comprising: illuminating the biometric feature with monochromatic light;filtering light reflected from the biometric feature and ambient light with a color filter array into a plurality of wavelength bands, wherein one of said plurality of wavelength bands corresponds with the wavelength of the monochromatic light illuminating the biometric feature;detecting light reflected from the biometric feature and ambient light filtered by the color filter array, wherein each pixel of the detector receives light corresponding to a wavelength band and providing an image of the target;comparing pixels corresponding with the wavelength of the monochromatic light;segmenting portions of the image corresponding to a purported biometric feature from portions of the image corresponding to the background by comparing the levels of a wavelength band corresponding to the monochromatic light with the levels of another wavelength band.
  • 31. The method according to claim 30, wherein the monochromatic light comprises blue light and the another wavelength band comprises green light.
  • 32. A biometric feature presence detector comprising: a target including a target surface adapted to receive a potential biometric feature;a monochromatic light source disposed below the target surface, wherein the monochromatic light source illuminates at least a portion of the target surface and a portion of the area above the target surface with monochromatic light;a color filter array;a light detector disposed below the color filter array and the target surface, and configured to receive light from the target surface and ambient light from an area above the target surface through the color filter array; anda computational unit interfaced with the light detector and having: instructions to segment portions of the image corresponding to a purported biometric feature from portions of the image corresponding to the background by comparing the levels of a wavelength band corresponding to monochromatic light with the levels of another wavelength band.
CROSS-REFERENCES TO RELATED APPLICATIONS

This application is a non-provisional, and claims the benefit, of commonly assigned U.S. Provisional Application No. 60/911,007, filed Apr. 10, 2007, entitled “Spatial And Temporal Biometric Detection,” the entirety of which is herein incorporated by reference for all purposes.

US Referenced Citations (354)
Number Name Date Kind
3508830 Hopkins et al. Apr 1970 A
3619060 Johnson Nov 1971 A
3854319 Burroughs et al. Dec 1974 A
3872443 Ott Mar 1975 A
3910701 Henderson et al. Oct 1975 A
RE29008 Ott Oct 1976 E
4035083 Woodriff et al. Jul 1977 A
4142797 Astheimer Mar 1979 A
4169676 Kaiser Oct 1979 A
4170987 Anselmo et al. Oct 1979 A
4260220 Whitehead Apr 1981 A
4322163 Schiller Mar 1982 A
4427889 Muller Jan 1984 A
4537484 Fowler Aug 1985 A
4598715 Machler et al. Jul 1986 A
4653880 Sting et al. Mar 1987 A
4654530 Dybwad Mar 1987 A
4655225 Dahne et al. Apr 1987 A
4656562 Sugino Apr 1987 A
4657397 Oehler et al. Apr 1987 A
4661706 Messerschmidt et al. Apr 1987 A
4684255 Ford Aug 1987 A
4699149 Rice Oct 1987 A
4712912 Messerschmidt Dec 1987 A
4730882 Messerschmidt Mar 1988 A
4747147 Sparrow May 1988 A
4787013 Sugino et al. Nov 1988 A
4787708 Whitehead Nov 1988 A
4830496 Young May 1989 A
4853542 Milosevic et al. Aug 1989 A
4857735 Noller Aug 1989 A
4859064 Messerschmidt et al. Aug 1989 A
4866644 Shenk et al. Sep 1989 A
4867557 Takatani et al. Sep 1989 A
4882492 Schlager Nov 1989 A
4883953 Koashi et al. Nov 1989 A
4936680 Henkes et al. Jun 1990 A
4937764 Komatsu et al. Jun 1990 A
4944021 Hoshino et al. Jul 1990 A
4975581 Robinson et al. Dec 1990 A
5015100 Doyle May 1991 A
5019715 Sting et al. May 1991 A
5028787 Rosenthal et al. Jul 1991 A
5051602 Sting et al. Sep 1991 A
5055658 Cockburn Oct 1991 A
5068536 Rosenthal Nov 1991 A
5070874 Barnes et al. Dec 1991 A
5077803 Kato et al. Dec 1991 A
5088817 Igaki et al. Feb 1992 A
5109428 Igaki et al. Apr 1992 A
5146102 Higuchi et al. Sep 1992 A
5158082 Jones Oct 1992 A
5163094 Prokoski et al. Nov 1992 A
5177802 Fujimoto et al. Jan 1993 A
5178142 Harjunmaa et al. Jan 1993 A
5179951 Knudson Jan 1993 A
5204532 Rosenthal Apr 1993 A
5222495 Clarke et al. Jun 1993 A
5222496 Clarke et al. Jun 1993 A
5223715 Taylor Jun 1993 A
5225678 Messerschmidt Jul 1993 A
5230702 Lindsay et al. Jul 1993 A
5237178 Rosenthal et al. Aug 1993 A
5243546 Maggard Sep 1993 A
5257086 Fateley et al. Oct 1993 A
5258922 Grill Nov 1993 A
5267152 Yang et al. Nov 1993 A
5268749 Weber et al. Dec 1993 A
5291560 Daugman Mar 1994 A
5299570 Hatschek Apr 1994 A
5303026 Strobl et al. Apr 1994 A
5311021 Messerschmidt May 1994 A
5313941 Braig et al. May 1994 A
5321265 Block Jun 1994 A
5331958 Oppenheimer Jul 1994 A
5335288 Faulkner Aug 1994 A
5348003 Caro Sep 1994 A
5351686 Steuer et al. Oct 1994 A
5355880 Thomas et al. Oct 1994 A
5360004 Purdy et al. Nov 1994 A
5361758 Hall et al. Nov 1994 A
5366903 Lundsgaard et al. Nov 1994 A
5372135 Mendelson et al. Dec 1994 A
5379764 Barnes et al. Jan 1995 A
5402778 Chance Apr 1995 A
5405315 Khuri et al. Apr 1995 A
5413098 Benaron et al. May 1995 A
5419321 Evans May 1995 A
5435309 Thomas et al. Jul 1995 A
5441053 Lodder et al. Aug 1995 A
5452723 Wu et al. Sep 1995 A
5459317 Small et al. Oct 1995 A
5459677 Kowalski et al. Oct 1995 A
5460177 Purdy et al. Oct 1995 A
5483335 Tobias Jan 1996 A
5494032 Robinson et al. Feb 1996 A
5505726 Meserol Apr 1996 A
5507723 Keshaviah Apr 1996 A
5515847 Braig et al. May 1996 A
5518623 Keshaviah et al. May 1996 A
5523054 Switalski et al. Jun 1996 A
5533509 Koashi et al. Jul 1996 A
5537208 Bertram et al. Jul 1996 A
5539207 Wong et al. Jul 1996 A
5552997 Massart Sep 1996 A
5559504 Itsumi et al. Sep 1996 A
5568251 Davies et al. Oct 1996 A
5596992 Haaland et al. Jan 1997 A
5606164 Price et al. Feb 1997 A
5613014 Eshera et al. Mar 1997 A
5630413 Thomas et al. May 1997 A
5636633 Messerschmidt et al. Jun 1997 A
5655530 Messerschmidt Aug 1997 A
5672864 Kaplan Sep 1997 A
5672875 Block et al. Sep 1997 A
5677762 Ortyn et al. Oct 1997 A
5681273 Brown Oct 1997 A
5708593 Saby et al. Jan 1998 A
5719399 Alfano et al. Feb 1998 A
5719950 Osten et al. Feb 1998 A
5724268 Sodickson et al. Mar 1998 A
5729619 Puma Mar 1998 A
5737439 Lapsley et al. Apr 1998 A
5743262 Lepper, Jr. et al. Apr 1998 A
5747806 Khalil May 1998 A
5750994 Schlager May 1998 A
5751835 Topping et al. May 1998 A
5751836 Wildes et al. May 1998 A
5761330 Stoianov et al. Jun 1998 A
5782755 Chance et al. Jul 1998 A
5792050 Alam et al. Aug 1998 A
5792053 Skladnev et al. Aug 1998 A
5793881 Stiver et al. Aug 1998 A
5796858 Zhou et al. Aug 1998 A
5808739 Turner et al. Sep 1998 A
5818048 Sodickson et al. Oct 1998 A
5823951 Messerschmidt et al. Oct 1998 A
5828066 Messerschmidt Oct 1998 A
5830132 Robinson Nov 1998 A
5830133 Osten et al. Nov 1998 A
5850623 Carman, Jr. et al. Dec 1998 A
5853370 Chance et al. Dec 1998 A
5857462 Thomas et al. Jan 1999 A
5859420 Borza Jan 1999 A
5860421 Eppstein et al. Jan 1999 A
5867265 Thomas Feb 1999 A
5886347 Inoue et al. Mar 1999 A
5902033 Levis et al. May 1999 A
5914780 Turner et al. Jun 1999 A
5929443 Alfano et al. Jul 1999 A
5933792 Andersen et al. Aug 1999 A
5935062 Messerschmidt et al. Aug 1999 A
5945676 Khalil Aug 1999 A
5949543 Bleier et al. Sep 1999 A
5957841 Maruo et al. Sep 1999 A
5961449 Toida et al. Oct 1999 A
5963319 Jarvis et al. Oct 1999 A
5978495 Thomopoulos et al. Nov 1999 A
5987346 Benaron et al. Nov 1999 A
5999637 Toyoda et al. Dec 1999 A
6005722 Butterworth et al. Dec 1999 A
6016435 Maruo et al. Jan 2000 A
6025597 Sterling et al. Feb 2000 A
6026314 Amerov et al. Feb 2000 A
6028773 Hundt Feb 2000 A
6031609 Funk et al. Feb 2000 A
6034370 Messerschmidt Mar 2000 A
6040578 Malin et al. Mar 2000 A
6041247 Weckstrom et al. Mar 2000 A
6041410 Hsu et al. Mar 2000 A
6043492 Lee et al. Mar 2000 A
6044285 Chaiken et al. Mar 2000 A
6045502 Eppstein et al. Apr 2000 A
6046808 Fateley Apr 2000 A
6049727 Crothall Apr 2000 A
6056738 Marchitto et al. May 2000 A
6057925 Anthon May 2000 A
6061581 Alam et al. May 2000 A
6061582 Small et al. May 2000 A
6066847 Rosenthal May 2000 A
6069689 Zeng et al. May 2000 A
6070093 Oosta et al. May 2000 A
6073037 Alam et al. Jun 2000 A
6081612 Gutkowicz-Krusin et al. Jun 2000 A
6088605 Griffith et al. Jul 2000 A
6088607 Diab et al. Jul 2000 A
6097035 Belongie et al. Aug 2000 A
6100811 Hsu et al. Aug 2000 A
6115484 Bowker et al. Sep 2000 A
6115673 Malin et al. Sep 2000 A
6122042 Wunderman et al. Sep 2000 A
6122394 Neukermans et al. Sep 2000 A
6122737 Bjorn et al. Sep 2000 A
6125192 Bjorn et al. Sep 2000 A
6141101 Bleier et al. Oct 2000 A
6147749 Kubo et al. Nov 2000 A
6148094 Kinsella Nov 2000 A
6152876 Robinson et al. Nov 2000 A
6154658 Caci Nov 2000 A
6157041 Thomas et al. Dec 2000 A
6159147 Lichter et al. Dec 2000 A
6172743 Kley et al. Jan 2001 B1
6175407 Sartor Jan 2001 B1
6181414 Raz et al. Jan 2001 B1
6181958 Steuer et al. Jan 2001 B1
6188781 Brownlee Feb 2001 B1
6193153 Lambert Feb 2001 B1
6208749 Gutkowicz-Krusin Mar 2001 B1
6212424 Robinson Apr 2001 B1
6226541 Eppstein et al. May 2001 B1
6229908 Edmonds et al. May 2001 B1
6230034 Messerschmidt et al. May 2001 B1
6236047 Malin et al. May 2001 B1
6240306 Rohrscheib et al. May 2001 B1
6240309 Yamashita et al. May 2001 B1
6241663 Wu et al. Jun 2001 B1
6256523 Diab et al. Jul 2001 B1
6272367 Chance Aug 2001 B1
6280381 Malin et al. Aug 2001 B1
6282303 Brownlee Aug 2001 B1
6285895 Ristolainen et al. Sep 2001 B1
6292576 Brownlee Sep 2001 B1
6301375 Choi Oct 2001 B1
6301815 Sliwa Oct 2001 B1
6304767 Soller et al. Oct 2001 B1
6307633 Mandella et al. Oct 2001 B1
6309884 Cooper et al. Oct 2001 B1
6317507 Dolfing et al. Nov 2001 B1
6324310 Brownlee Nov 2001 B1
6330346 Peterson et al. Dec 2001 B1
6404904 Einighammer et al. Jun 2002 B1
6419361 Cabib et al. Jul 2002 B2
6483929 Murakami et al. Nov 2002 B1
6504614 Messerschmidt et al. Jan 2003 B1
6537225 Mills Mar 2003 B1
6560352 Rowe et al. May 2003 B2
6574490 Abbink et al. Jun 2003 B2
6597945 Marksteiner Jul 2003 B2
6606509 Schmitt Aug 2003 B2
6628809 Rowe et al. Sep 2003 B1
6631199 Topping et al. Oct 2003 B1
6741729 Bjorn et al. May 2004 B2
6749115 Gressel et al. Jun 2004 B2
6799275 Bjorn Sep 2004 B1
6799726 Stockhammer Oct 2004 B2
6816605 Rowe et al. Nov 2004 B2
6825930 Cronin et al. Nov 2004 B2
6853444 Haddad Feb 2005 B2
6898299 Brooks May 2005 B1
6928181 Brooks Aug 2005 B2
6937885 Lewis et al. Aug 2005 B1
6958194 Hopper et al. Oct 2005 B1
6995384 Lee et al. Feb 2006 B2
7047419 Black May 2006 B2
7084415 Iwai Aug 2006 B2
7147153 Rowe et al. Dec 2006 B2
7254255 Dennis Aug 2007 B2
7263213 Rowe Aug 2007 B2
7287013 Schneider et al. Oct 2007 B2
7347365 Rowe Mar 2008 B2
7366331 Higuchi Apr 2008 B2
7386152 Rowe et al. Jun 2008 B2
7394919 Rowe et al. Jul 2008 B2
7397943 Merbach et al. Jul 2008 B2
7440597 Rowe Oct 2008 B2
7460696 Rowe Dec 2008 B2
7508965 Rowe et al. Mar 2009 B2
7515252 Hernandez Apr 2009 B2
7539330 Rowe May 2009 B2
7545963 Rowe Jun 2009 B2
7627151 Rowe Dec 2009 B2
7668350 Rowe Feb 2010 B2
7735729 Rowe Jun 2010 B2
7751594 Rowe et al. Jul 2010 B2
7801338 Rowe Sep 2010 B2
7801339 Sidlauskas et al. Sep 2010 B2
7804984 Sidlauskas et al. Sep 2010 B2
7819311 Rowe et al. Oct 2010 B2
7831072 Rowe Nov 2010 B2
7835554 Rowe Nov 2010 B2
7899217 Uludag et al. Mar 2011 B2
7995808 Rowe et al. Aug 2011 B2
20020009213 Rowe et al. Jan 2002 A1
20020065468 Ultzinger et al. May 2002 A1
20020101566 Elsner et al. Aug 2002 A1
20020111546 Cook et al. Aug 2002 A1
20020138768 Murakami et al. Sep 2002 A1
20020171834 Rowe et al. Nov 2002 A1
20020183624 Rowe et al. Dec 2002 A1
20030025897 Iwai Feb 2003 A1
20030044051 Fujieda Mar 2003 A1
20030078504 Rowe Apr 2003 A1
20030095525 Lavin et al. May 2003 A1
20030128867 Bennett Jul 2003 A1
20030163710 Ortiz et al. Aug 2003 A1
20030223621 Rowe et al. Dec 2003 A1
20040003295 Elderfield et al. Jan 2004 A1
20040008875 Linares Jan 2004 A1
20040022421 Endoh et al. Feb 2004 A1
20040042642 Bolle et al. Mar 2004 A1
20040047493 Rowe et al. Mar 2004 A1
20040114783 Spycher et al. Jun 2004 A1
20040120553 Stobbe Jun 2004 A1
20040125994 Engels et al. Jul 2004 A1
20040179722 Moritoki et al. Sep 2004 A1
20040240712 Rowe et al. Dec 2004 A1
20040240713 Hata Dec 2004 A1
20040264742 Zhang et al. Dec 2004 A1
20050007582 Villers et al. Jan 2005 A1
20050125339 Tidwell et al. Jun 2005 A1
20050169504 Black Aug 2005 A1
20050180620 Takiguchi Aug 2005 A1
20050185847 Rowe Aug 2005 A1
20050205667 Rowe Sep 2005 A1
20050265585 Rowe Dec 2005 A1
20050265586 Rowe et al. Dec 2005 A1
20050265607 Chang Dec 2005 A1
20050271258 Rowe Dec 2005 A1
20060002597 Rowe Jan 2006 A1
20060002598 Rowe et al. Jan 2006 A1
20060045330 Marion Mar 2006 A1
20060062438 Rowe Mar 2006 A1
20060110015 Rowe May 2006 A1
20060115128 Mainguet Jun 2006 A1
20060171571 Chan et al. Aug 2006 A1
20060173256 Ridder et al. Aug 2006 A1
20060202028 Rowe Sep 2006 A1
20060210120 Rowe Sep 2006 A1
20060244947 Rowe Nov 2006 A1
20060274921 Rowe Dec 2006 A1
20070014437 Sato Jan 2007 A1
20070030475 Rowe et al. Feb 2007 A1
20070052827 Hiltunen Mar 2007 A1
20070116331 Rowe et al. May 2007 A1
20070153258 Hernandez Jul 2007 A1
20070165903 Munro et al. Jul 2007 A1
20080008359 Beenau et al. Jan 2008 A1
20080013806 Hamid Jan 2008 A1
20080025579 Sidlauskas et al. Jan 2008 A1
20080025580 Sidlauskas et al. Jan 2008 A1
20080192988 Uludag et al. Aug 2008 A1
20080232653 Rowe Sep 2008 A1
20080260211 Bennett et al. Oct 2008 A1
20080298649 Ennis et al. Dec 2008 A1
20090046903 Corcoran et al. Feb 2009 A1
20090080709 Rowe et al. Mar 2009 A1
20090092290 Rowe Apr 2009 A1
20090148005 Rowe Jun 2009 A1
20090245591 Rowe et al. Oct 2009 A1
20100067748 Rowe Mar 2010 A1
20100246902 Rowe et al. Sep 2010 A1
20110085708 Martin et al. Apr 2011 A1
20110211055 Martin et al. Sep 2011 A1
20110235872 Rowe et al. Sep 2011 A1
Foreign Referenced Citations (66)
Number Date Country
1307711 Aug 2001 CN
1402183 Mar 2003 CN
1509454 Jun 2004 CN
10153808 May 2003 DE
0 280 418 Aug 1988 EP
0 372 748 Jun 1990 EP
0 426 358 May 1991 EP
0 449 335 Oct 1991 EP
0 573 137 Dec 1993 EP
0 631 137 Dec 1994 EP
0 670 143 Sep 1995 EP
0 681 166 Nov 1995 EP
0 757 243 Feb 1997 EP
0 788 000 Aug 1997 EP
0 801 297 Oct 1997 EP
0 836 083 Apr 1998 EP
0 843 986 May 1998 EP
0 869 348 Oct 1998 EP
0 897 164 Feb 1999 EP
0 897 691 Feb 1999 EP
0 317 121 May 1999 EP
0 924 656 Jun 1999 EP
0 982 583 Mar 2000 EP
0 990 945 Apr 2000 EP
1 353 292 Oct 2003 EP
1 434 162 Jun 2004 EP
2761180 Jan 1998 FR
61182174 Aug 1986 JP
3016160 Jan 1991 JP
7075629 Mar 1995 JP
10-127585 May 1998 JP
2001-033381 Feb 2001 JP
2001-112742 Apr 2001 JP
2001-184490 Jul 2001 JP
2002-133402 May 2002 JP
2002-517835 Jun 2002 JP
2003-050993 Feb 2003 JP
2003-511101 Mar 2003 JP
2003-308520 Oct 2003 JP
WO 9200513 Jan 1992 WO
WO 9217765 Oct 1992 WO
WO 9300855 Jan 1993 WO
WO 9307801 Apr 1993 WO
WO 9927848 Jun 1999 WO
WO 0030530 Jun 2000 WO
WO 0046739 Aug 2000 WO
WO 0115596 Mar 2001 WO
WO 0118332 Mar 2001 WO
WO 0120538 Mar 2001 WO
WO 0127882 Apr 2001 WO
WO 0152180 Jul 2001 WO
WO 0152726 Jul 2001 WO
WO 0153805 Jul 2001 WO
WO 0165471 Sep 2001 WO
WO 0169520 Sep 2001 WO
WO 02054337 Jul 2002 WO
WO 02084605 Oct 2002 WO
WO 02099393 Dec 2002 WO
WO 03010510 Feb 2003 WO
WO 03096272 Nov 2003 WO
WO 2004068388 Aug 2004 WO
WO 2004068394 Aug 2004 WO
WO 2004090786 Oct 2004 WO
WO 2006049394 May 2006 WO
WO 2006077446 Jul 2006 WO
WO 2006093508 Sep 2006 WO
Non-Patent Literature Citations (49)
Entry
Chinese Patent Application No. 2006/80038597.4, First Office Action mailed on Mar. 23, 2011, 7 pages.
European Patent Application No. 10166537.0, Extended European Search Report mailed on Jun. 1, 2011, 7 pages.
Author Unknown, “Improve the Clinical Outcome of Every Patient,” In Line Diagnostics, published on or before Oct. 30, 1997, 2 pages.
Anderson, C. E. et al., “Fundamentals of Calibration Transfer Through Procrustes Analysis,” Appln. Spectros., vol. 53, No. 10, 1999, pp. 1268-1276.
Ashboum, Julian, Biometrics; Advanced Identity Verification, Springer, 2000, pp. 63-64.
Bantle, John P. et al., “Glucose Measurement in Patients With Diabetes Mellitus With Dermal Interstitial Fluid,” Mosby-Year Book, Inc., 1997, 9 pages.
Berkoben, Michael S. et al., “Vascular Access for Hemodialysis,” Clinical Dialysis, Third Edition, 1995, pp. 2 cover pages and 26-45.
Blank, T.B. et al., “Transfer of Near-Infrared Multivariate Calibrations Without Standards,” Anal. Chem., vol. 68, 1996, p. 2987.
Bleyer, Anthony J. et al., “The Costs of Hospitalizations Due to Hemodialysis Access Management,” Nephrology News & Issues, Jan. 1995, pp. 19, 20 and 22.
Brasunas, John C. et al., “Uniform Time-Sampling Fourier Transform Spectroscopy,” Applied Optics, vol. 36, No. 10, Apr. 1, 1997, pp. 2206-2210.
Brault, James W., “New Approach to High-Precision Fourier Transform Spectrometer Design,” Applied Optics, Vo. 35, No. 16, Jun. 1, 1996, pp. 2891-2896.
Cassarly, W.J. et al., “Distributed Lighting Systems: Uniform Light Delivery,” Source Unknown, 1995, pp. 1698-1702.
Chang, Chong-Min et al., “An Uniform Rectangular Illuminating Optical System for Liquid Crystal Light Valve Projectors,” Euro Display '96, 1996, pp. 257-260.
Coyne, Lawrence J. et al., “Distributive Fiber Optic couplers Using Rectangular Lightguides as Mixing Elements,” Information Gatekeepers, Inc. Brookline, MA, 1979, pp. 160-164.
Daugirdas, JT et al., “Comparison of Methods to Predict the Equilibrated Kt/V (eKt/V) in the Hemo Study,” National Institutes of Health, Aug. 20, 1996, pp. 1-28.
De Noord, Onno E., “Multivariate Calibration Standardization,” Chemometrics and intelligent Laboratory Systems 25, 1994, pp. 85-97.
Demos, S. G. et al., “Optical Fingerprinting Using Polarisation Contrast Improvement,” Electronics Letters, vol. 33, No. 7, Mar. 27, 1997, pp. 582-584.
Depner, Thomas A. et al., “Clinical Measurement of Blood Flow in Hemodialysis Access Fistulae and Grafts by Ultrasound Dilution,” Division of Nephrology, University of California, published on or before Oct. 30, 1997, pp. M745-M748.
Despain, Alvin M. et al., “A Large-Aperture Field-Widened Interferometer-Spectrometer for Airglow Studies,” Aspen International Conference on Fourier Spectroscopy, 1970, pp. 293-300.
Faber, Nicolaas, “Multivariate Sensitivity for the Interpretation of the Effect of Spectral Pretreatment Methods on Near-Infrared Calibration Model Predictions,” Analytical Chemistry, vol. 71, No. 3, Feb. 1, 1999, pp. 557-565.
Fresenius USA, “Determination of Delivered Therapy Through Measurement of Effective Clearance,” Dec. 1994, 2 pages.
Geladi, Paul et al., A Multivariate NIR Study of Skin Alterations in Diabetic Patients as Compared to Control Subjects, Fresenius USA, “Determination of Delivered Therapy Through Measurement of Effective Clearance,” 2 pages, Dec. 1994, J. Near Infrared Spectrosc., vol. 8, 2000, pp. 217-227.
Hakim, Raymond M. et al., “Effects of Dose of Dialysis on Morbidity and Mortality,” American Journal of Kidney Diseases, vol. 23, No. 5, May 1994, pp. 661-669.
International Search Report and Written Opinion of PCT/US2008/066585 mailed Oct. 30, 2008, 10 pages.
International Search Report and Written Opinion of PCT/US2010/025463 mailed on Jun. 30, 2010, 6 pages.
International Search Report of PCT/US2010/046852 mailed Dec. 29, 2010, 5 pages.
Jacobs, Paul et al., “A Disposable Urea Sensor for Continuous Monitoring of Hemodialysis Efficiency,” ASAIO Journal, 1993, pp. M353-M358.
Keshaviah, Prakash R. et al., “On-Line Monitoring of the Delivery of the Hemodialysis Prescription,” Pediatric Nephrology, vol. 9, 1995, pp. S2-S8.
Krivitski, Nikolai M., “Theory and Validation of Access Flow Measurement by Dilution Technique During Hemodialysis,” Kidney International, vol. 48, 1995, pp. 244-250.
Lee et al., “Fingerprint Recognition Using Principal Gabor Basis Function”, Proceedings of 2001 International Symposium on Intelligent Multimedia, Video and Speech Processing, May 2-4, 2001, Sections 2-3.
Maltoni et al., “Handbook of Fingerprint Recognition,” 2005, pp. 58-61.
Marbach, Ralf, “Measurement Techniques for IR Spectroscopic Blood Glucose Determination,” Fortschritt Bericht, Series 8: Measurement and Control Technology, No. 346, Mar. 28, 1994, pp. cover and 1-158.
Mardia, K.V. et al., “Chapter 11—Discriminant Analysis,” Multivariate Analysis, 1979, pp. 2 cover pages and 300-325.
Nichols, Michael G. et al., “Design and Testing of a White-Light, Steady-State Diffuse Reflectance Spectrometer for Determination of Optical Properties of Highly Scattering Systems,” Applied Optics, vol. 36, No. 1, Jan. 1, 1997, pp. 93-104.
Nixon, Kristin A. et al., “Novel Spectroscopy-Based Technology for Biometric and Liveness Verification,” Technology for Human Identification. Proceedings oF SPIE, vol. 5404, No. 1, XP-002458441, Apr. 12-13, 2004, pp. 287-295 (ISSN: 0277-786x).
Pan et al., “Face Recognition in Hyperspectral Images,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 25, No. 12, Dec. 2003, pp. 1552-1560.
Ripley, B. D., “Chapter 3—Linear Discriminant Analysis,” Pattern Recognition and Neural Networks, 1996, pp. 3 cover pages and 91-120.
Ronco, C. et al., “On-Line Urea Monitoring : A Further Step Towards Adequate Dialysis Prescription and Delivery,” The International Journal of Artificial Organs, vol. 18, No. 9, 1995, pp. 534-543.
Ross et al., “A Hybrid Fingerprint Matcher,” Pattern Recognition 36, The Journal of the Pattern Recognition Society, 2003 Elsevier Science Ltd., pp. 1661-1673.
Rowe et al. “Multispectral Fingerprint Image Acquisition,” Advance in Biometrics, 2008, 22 pages.
Selvaraj et al., “Fingerprint Verification Using Wavelet Transform,” Proceedings of the Fifth International Conference on Computational Intelligence and Multimedia Applications, IEEE, 2003, 6 pages.
Service, F. John et al., “Dermal Interstitial Glucose As an Indicator of Ambient Glycemia,” Diabetes Care, vol. 20, No. 9, Aug. 1997, 8 pages.
Sherman, Richard A., “Chapter 4—Recirculation in the Hemodialysis Access,” Principles and Practice of Dialysis, pp. 2 cover pages and 38-46, 1994.
Sherman, Richard A., “The Measurement of Dialysis Access Recirculation,” American Journal of Kidney Diseases, vol. 22, No. 4, Oct. 1993, pp. 616-621.
Steuer, Robert R. et al., “A New Optical Technique for Monitoring Hematocrit and Circulating Blood Volume: Its Application in Renal Dialysis,” Dialysis & Transplantation, vol. 22, No. 5, May 1993, pp. 260-265.
Webb, Paul, “Temperatures of Skin, Subcutaneous Tissue, Muscle and Core in Resting Men in Cold, Comfortable and Hot Conditions,” European Journal of Applied Physiology, vol. 64, 1992, pp. 471-476.
Zavala, Albert et al., “Using Fingerprint Measures to Predict other Anthropometric Variables,” Human Factors, vol. 17, No. 6, 1975, pp. 591-602.
Chinese Patent Application No. 2006/80038579.4, First Office Action mailed on Mar. 23, 2011, 7 pages.
Rowe, “LumiGuard: A Novel Spectroscopic Sensor for Biometric Security Applications”, American Chemical Society 225th National Meeting, Mar. 25, 2003, 20 pages.
Related Publications (1)
Number Date Country
20090046903 A1 Feb 2009 US
Provisional Applications (1)
Number Date Country
60911007 Apr 2007 US