Biometric imaging using an optical adaptive interface

Information

  • Patent Grant
  • 8570149
  • Patent Number
    8,570,149
  • Date Filed
    Monday, March 14, 2011
    13 years ago
  • Date Issued
    Tuesday, October 29, 2013
    10 years ago
Abstract
Embodiments of the invention provide for a biometric system with an optically adaptive interface. In some embodiments, an optically adaptive interface changes optical characteristics in response to the placement of a finger on the optically adaptive interface. In some embodiments, the optically adaptive interface can include an active layer and a surface layer. The active layer and the surface layer can have different optical properties. For example, one layer may be opaque and the other transparent, the two layers may have complementary colors, the two layers may have orthogonal polarization reflectors, one layer may be reflective and the other absorptive, etc. Moreover, the active layer can be a fluid with either high or low viscosity. For example, the viscosity can be such that the active layer fluid is either completely displaced or not displaced in locations corresponding to finger valleys.
Description
BACKGROUND

“Biometrics” refers generally to the statistical analysis of characteristics of living bodies. One category of biometrics includes “biometric identification,” which commonly operates under one of two modes to provide automatic identification of people or to verify purported identities of people. Biometric sensing technologies measure the physical features or behavioral characteristics of a person and compare those features to similar prerecorded measurements to determine whether there is a match. Physical features that are commonly used for biometric identification include faces, irises, hand geometry, vein structure, and fingerprint patterns, which is the most prevalent of all biometric-identification features. Current methods for analyzing collected fingerprints include optical, capacitive, radio-frequency, thermal, ultrasonic, and several other less common techniques.


Most existing fingerprint sensors rely on relatively high-quality contact between the finger and the sensor to obtain images. Obtaining adequate contact is both finicky and time-consuming because of factors related to individual characteristics of users of the sensors, the quality of the skin, and environmental variability. Ease of consistent fingerprint capture limits the effectiveness and scope of applications that utilize fingerprint biometrics for identity management.


SUMMARY

An optically adaptive interface is provided according to various embodiments of the inventions. An optically adaptive interface, for example, can be placed on the surface of a platen or any other device. An optically adaptive interface can change its optical characteristics in response to contact (or the presence) of a finger at the optically adaptive interface. This change can provide a different look and feel to the device depending on whether a user is interacting with the device or not. An optically adaptive interface can include thin layer(s) disposed on the surface of a biometric sensor platen. An optically adaptive interface can prevent a user from looking into the biometric sensor and seeing the illumination and/or imaging optics. An optically adaptive interface can also cut down on ambient light entering the sensor that can cause image degradation, sensor saturation, and/or other damage to the imaging sensor or loss of quality in the resulting image. An optically adaptive interface can also prevent light from within the sensor (e.g., from illumination optics) from exiting the sensor. Such light can be annoying or disruptive to a user. The optically adaptive interface can be used to improve the aesthetics of the sensor surface by allowing the sensor surface to have any color.


According to some embodiments, a biometric sensor can include a platen, an optically adaptive interface, an illumination subsystem, and an imaging subsystem. The optically adaptive interface can be disposed on the platen and can include a contact state and a non-contact state. In the non-contact state the optically adaptive interface is opaque and in the contact state a portion of the optically adaptive interface is substantially transparent. The illumination subsystem can be configured to illuminate a skin site of a purported individual when placed on the optically adaptive interface when the optically adaptive interface is in the contact state. The imaging subsystem can be configured to image the skin site when the optically adaptive interface is in the contact state.


In some embodiments, the default state of the optically adaptive interface is the non-contact state, and the optically adaptive interface changes from the non-contact state to the contact state in response to contact by the skin site. In some embodiments, the portion of the optically adaptive interface that is substantially transparent in the contact state occurs where the skin site is in contact with the optically adaptive interface. In some embodiments, the optically adaptive interface blocks light from the illumination subsystem from being transmitted through the optically adaptive interface. In some embodiments, the optically adaptive interface comprises a surface layer and an active layer. In some embodiments, the active layer comprises a fluid-like material and the surface layer is a thin film.


In some embodiments, a biometric sensor can include a housing, a platen defining a surface of the housing, an optically adaptive interface disposed on the platen, an illumination subsystem disposed within the housing, and an imaging subsystem disposed within the housing. The optically adaptive interface can include an opaque surface when viewed from an environment external from the housing. A portion of the optically adaptive interface changes optical properties when a skin site of an individual is placed on the optically adaptive interface allowing the illumination and imaging subsystems to produce a biometric image of the skin site.


In some embodiments, the optically adaptive interface blocks a user from viewing both the illumination subsystem and the imaging subsystem. In some embodiments, the optically adaptive interface changes optical properties at the location where the skin site is in contact with the optically adaptive interface. In some embodiments the change in the optical properties of the optically adaptive interface is a change from being opaque to substantially transparent. In some embodiments, the optically adaptive interface comprises a thin surface layer and an fluid-like active layer.


According to some embodiments, a biometric sensor can include a platen, an optically adaptive interface disposed on the platen, an illumination subsystem configured to illuminate portions of the platen and the optically adaptive interface, and an imaging subsystem configured to collect light from a skin site of a purported individual when placed on the optically adaptive interface. The optically adaptive interface can include two states: a non-contact state with a first optical condition, and a contact state with a second optical condition.


In some embodiments, in the contact state only a portion of the optically adaptive interface includes the second optical condition. In some embodiments, the illumination subsystem includes an optical element that filters light according to either the first optical condition or the second optical condition. In some embodiments, the optically adaptive interface comprises a surface layer and a fluid-like active layer. In some embodiments, the first optical condition the optically adaptive interface absorbs light from the illumination subsystem, and in the second optical condition at least a portion of the optically adaptive interface reflects light from the illumination subsystem.


In some embodiments, in the first optical condition the optically adaptive interface reflects light from the illumination subsystem, and in the second optical condition at least a portion of the optically adaptive interface absorbs light from the illumination subsystem.


In some embodiments, in the first optical condition the optically adaptive interface reflects light having a first wavelength, and in the second optical condition at least a portion of the optically adaptive interface reflects light having a second wavelength distinct from the first wavelength.


In some embodiments, in the first optical condition the optically adaptive interface absorbs light having a first wavelength, and in the second optical condition at least a portion of the optically adaptive interface absorbs light having a second wavelength distinct from the first wavelength.


In some embodiments, in the first optical condition the optically adaptive interface is substantially opaque, and in the second optical condition at least a portion of the optically adaptive interface is substantially transparent.


In some embodiments, the first optical condition the optically adaptive interface reflects light with a first polarization condition, and in the second optical condition at least a portion of the optically adaptive interface reflects light having a second polarization condition orthogonal with the first polarization condition.





BRIEF DESCRIPTION OF THE DRAWINGS


FIGS. 1A, 1B, 1C, and 1D show an optically adaptive interface in a non-contact state and a contact state according to some embodiments of the invention.



FIG. 2 shows an optically adaptive interface according to some embodiments of the invention.



FIG. 3 shows an optically adaptive interface with a high viscosity active layer according to some embodiments of the invention.



FIG. 4 shows an optically adaptive interface with a low viscosity active layer according to some embodiments of the invention.



FIG. 5 is a block diagram of a biometric sensor computation unit according to some embodiments of the invention.



FIG. 6A is a TIR-imaging biometric sensor according to some embodiments of the invention.



FIG. 6B is a direct-imaging biometric sensor according to some embodiments of the invention.



FIG. 7A is a TIR-illumination biometric sensor according to some embodiments of the invention.



FIG. 7B is a direct-illumination biometric sensor according to some embodiments of the invention.





DETAILED DESCRIPTION

The ensuing description provides examples of embodiment(s) of the invention only, and is not intended to limit the scope, applicability or configuration of the disclosure. Rather, the ensuing description of the embodiment(s) will provide those skilled in the art with an enabling description for implementing these embodiments. Various changes may be made in the function and arrangement of elements without departing from the spirit and scope as set forth in the claims.


Introduction


An optically adaptive interface for biometric sensors is provided according to some embodiments of the invention. An optically adaptive interface, for example, can be placed on the surface of a platen, and can change its optical characteristics in response to contact (or the presence) of a finger at the optically adaptive interface. This change can provide a different look and feel to the device depending on whether a user is interacting with the device or not. An optically adaptive interface can include thin layer(s) disposed on the surface of a biometric sensor platen. An optically adaptive interface can prevent a user from looking into the biometric sensor and seeing the illumination and/or imaging optics. An optically adaptive interface can also cut down on ambient light entering the sensor that can cause image degradation, sensor saturation, and/or other damage to the imaging sensor or loss of quality in the resulting image. An optically adaptive interface can also prevent light from within the sensor (e.g., from illumination optics) from exiting the sensor. Such light can be annoying or disruptive to a user. The optically adaptive interface can be used to improve the aesthetics of the sensor surface by allowing the sensor surface to have any color.


The optically adaptive interface can be configured to have a contact state and a non-contact state that are optically distinct. These distinct optical responses can be complimentary. For example, the two states can reflect light at different wavelengths, absorb light at different wavelengths, one state can absorb light while another reflects light, and the two states can reflect light with distinct (orthogonal) polarization, and can have opposite light transmission properties. In some embodiments, the imaging and/or illumination subsystems can include one or more optical elements that can be used to differentiate between light received from the optically adaptive interface under the two different states.


An optically adaptive interface can be used to increase the contrast between the ambient environment and the skin site when the skin site is present at the sensor. In some configurations, the optically adaptive interface can also be used to enhance fingerprint patterns. Various other improvements or useful configurations can also be used with the various embodiments disclosed herein.


For purposes of this disclosure, the terms “finger,” “fingerprint,” and “fingerprint image” are meant to include sites and images collected from a single finger, multiple fingers, intermediate finger joints, the palm, the entire palmar surface of the hand, and/or any other skin site on the body, as well as other animate or inanimate objects such as documents, barcodes, credentials, and the like.


The terms “multispectral imaging,” “MSI,” and “multi-imaging” refer to methods and systems for acquiring multiple images of a finger during a single measurement session, wherein at least two of the multiple images are collected under different optical conditions. Different optical conditions may include, but not limited to, different illumination wavelengths, different illumination angles (both in azimuth and elevation and may include elevations on either side of the optical critical angle defined by the sensor imaging surface and the air or other surrounding medium), different illumination polarization conditions, different imaging angles (both in azimuth and elevation and may include elevations on either side of the optical critical angle defined by the sensor imaging surface and the air or other surrounding medium), different imaging focal planes, different imaging spatial resolutions, different imaging temporal resolutions, different imaging polarization conditions, and other such conditions that substantially alter the resulting images. Also, unless otherwise specified, the angle of incidence, angle of illumination, angle of imaging, etc. can refer to the elevation and/or azimuth angle. Unless otherwise specified, the elevation angle is measured relative to the normal of the incident surface.


The terms “total internal reflectance imaging” and “TIR imaging” refer to a method of imaging wherein the optical axis of the imaging system lies at an angle relative to the normal of the sensor imaging surface and that is greater than the optical critical angle of that surface. A block diagram showing TIR imaging is shown in FIG. 6A. In this diagram, imager 550 images light from platen 505 at angle θimaging that is greater than the critical angle, θcritical, as measured from normal 516 of facet 505. This is TIR imaging. Illumination sources 522 can be positioned at various elevation and azimuth angles and. FIG. 6B shows imager 550 that images light from platen 505 at angle θimaging that is less than the critical angle, θcritical.



FIG. 7A illustrates TIR illumination. In this example, illumination source 520 illuminates platen 505 at an angle, θillumination, that is greater than the critical angle, θcritical. FIG. 7B shows a non-TIR illumination system, with the illumination source 520 illuminating platen 505 at an angle, θillumination, that is less than the critical angle, θcritical.


Various configurations of illumination and imaging are shown in FIGS. 6A, 6B, 7A and 7B. Embodiments of the invention can incorporate any combination of TIR imaging, direct imaging, TIR illumination, and direct illumination. Moreover, multiple illumination sources and/or imagers can be at multiple angles of elevation and/or aziumuth.


The critical angle is a function of the index of refraction of the two media on either side of an interface and is approximately 42 degrees for a glass-air interface. Because the optical axis of the TIR imaging system lies beyond the critical angle of the sensor surface, the surface acts as a mirror (as seen by the imager) when untouched, and can cease to act as a mirror in those locations in which a material with suitable optical characteristics comes into direct contact with the sensor surface.


In locations where a finger or other material contacts a sensor surface, a new critical angle is established. However, for purposes of the present disclosure, the term “critical angle” will refer to the angle established by the sensor (i.e., the platen surface) and the surrounding environment, which is assumed to be air for most purposes. Also, as known in the art, light will change angles at boundaries between media due to phenomena such as refraction, reflection, diffraction and other such effects. When a ray angle is referred to in the present application as being greater than or less than the critical angle, for example, the statement refers to the angle of the ray at the operative boundary such as the sensor imaging surface rather than the angle of the same ray at any other boundary or media, unless explicitly stated as such.


The term “direct imaging” refers to a method of imaging wherein the optical axis of the imaging system lies at an angle relative to the sensor imaging surface that is less than the optical critical angle of that surface. For example, the system shown in FIG. 7B is one sample of a direct imaging system.


The term “opaque” refers to the physical properly of absorbing, scattering or reflecting light incident on a material. A material can be opaque in the visual spectrum, in a specific spectrum, or within a wavelength band. A material is considered opaque if 90% of the light incident thereon is scattered, reflected or absorbed by the material. The term “transparent” refers to the physical property of allowing light to pass through a material. A material can be transparent in the visual spectrum or a specific wavelength band. A material is considered transparent if 90% of the light incident thereon is passed through the material. The term “translucent” refers to the physical property of allowing light of a specific wavelength or in a wavelength band to pass through a material.


Optically Adaptive Interface


An optically adaptive interface can be any layer(s), material(s), interface, condition, optical elements(s), coating(s), etc. configured to change optical properties in response to contact from a finger or the proximity of a finger near the optically adaptive interface. In some embodiments, an optically adaptive interface can have optical properties that change when a finger is placed on the optically adaptive interface. In some embodiments, the entire optically adaptive interface can change optical properties. In other embodiments, a portion (or portions) of the optically adaptive interface can change optical properties.



FIGS. 1A, 1B, 1C, and 1D show an example of optically adaptive interface 102. Optically adaptive interface 102 is opaque to light incident from within and without optically adaptive interface 102 in the non-contact state shown in FIG. 1B. This opacity can block light from traversing optically adaptive interface 102 and/or platen 105 from the interior of the device to the exterior and/vice-versa. That is, the opacity blocks light from traversing the optically adaptive interface. This opacity can keep users from viewing the internal structure of the device (e.g. imaging and/or illumination optics). This opacity can keep illumination light from beneath the platen from being seen by a user or keep such light from shinning in a user's eyes. This opacity can keep ambient light from flooding and damaging the imaging sensor. Optically adaptive interface 102 can have an aesthetically pleasing color, texture, appearance, etc. when viewed by a user.


Optically adaptive interface 102 can then change from being opaque to transparent or partially transparent when finger 160 is placed in contact with optically adaptive interface 102 as shown in FIG. 1C. As shown in FIG. 1D, optically adaptive interface 102 can be transparent in contact region 150 where optically adaptive interface 102 is in contact with finger 160 or near where optically adaptive interface 102 is in contact with finger 160. This transparency can allow the illumination and imaging subsystems to then illuminate and image finger 160 through contact region 150. When finger 160 is in contact, TIR and/or direct illumination and/or imaging techniques can be used.


As another example, optically adaptive interface 102 can change from having an optical property in the non-contact state to having a complimentary optical property when contacted by finger 160 in the contact state. Complimentary optical properties can include a number of configurations. Complimentary optical properties can produce distinct images when illuminated and imaged under the same conditions. The following chart shows examples of complimentary optical properties for a contact state and a non-contact state of the optically adaptive interface. Various other complimentary optical properties can be used. Wavelength #1 and wavelength #2 can include complimentary colors such as, for example, red and green, orange and blue, purple and yellow, etc.













Contact State
Non-Contact state







Reflective at wavelength #1
Reflective at wavelength #2


Reflective
Absorptive


Absorptive
Reflective


Absorptive at wavelength #1
Absorptive at wavelength #2


Opaque
Transparent


Reflects polarized light
Reflects polarized light orthogonally



from light reflected in contact state


Reflects polarized light
Filters polarized light orthogonally



from light reflected in contact state









Two Layer Optically Adaptive Interface



FIG. 2 shows biometric sensor 100 with two layer optically adaptive interface 102 according to some embodiments of the invention. Optically adaptive interface 102 can include surface layer 120 that covers active layer 110. Optically adaptive interface 102 rests on platen 105. An individual can use biometric sensor 100 by placing a finger at imaging site 115 on the surface of optically adaptive interface 102. The sensor can include one or more illumination sources 130 and an imaging system 140. These illumination sources can be direct or TIR illumination sources. In some embodiments, one illumination source can illuminate the finger at an angle, θ1, while another illuminates the finger at an angle, θ2. Either or both of these illumination sources can be direct illumination sources such that θ1critical and/or θ2critical. Where θcritical is the optical critical angle. Or one of the two can be illumination sources that are used with total internal reflection proposes θ1critical or θ2critical. Three or more illumination sources can be use in any combination of direct or TIR illumination configurations. The number, angle, and/or location of the illumination sources may be selected to achieve certain levels of illumination, to provide for multiple illumination wavelengths, to provide for multiple polarization conditions, to meet packaging requirements, and to meet other structural constraints of biometric sensor 100. Moreover, multiple illumination conditions can be used to provide multispectral imaging.


Illumination light passes from illumination source 130 toward imaging site 115. In some embodiments, the light can pass through illumination optics that shape the illumination to a desired form, such as in the form of flood light, light lines, light points, and the like. The illumination optics can include any combination of one or more lenses, polarizers, collimators, one or more filters, one or more mirrors, and/or other optical elements. In some embodiments, the illumination optics may also comprise a scanner mechanism to scan the illumination light in a specified one-dimensional or two-dimensional pattern. Illumination source 130 may comprise a point source, a line source, an area source, or may comprise a series of such sources in different embodiments. Illumination source 130 may be narrow band sources such as monochromatic LED's and laser diodes or may be broad band sources such as white-light LED's or incandescent sources. In the case where illumination source 130 comprises a series of sources, the series of sources may be of the same wavelength or different wavelengths. Illumination source 130 may be configured identically or they may differ from each other.


After the light passes through the optional illumination optics it passes through platen 105 toward optically adaptive interface 102. Optically adaptive interface can include active layer 110 and surface layer 120. These layers can have any number of properties and/or configurations. Multiple additional layers may also be used. Finger 160 can be imaged when placed at imaging site 115 on optically adaptive interface 102 layers. Light from illumination source 130 can then illuminate finger 160 so that reflected light and/or scattered light can be directed to imaging system 140. In some embodiments, this can be done through imaging optics 150. Platen 105 may be configured in such a manner that illumination light entering the platen will traverse platen 105 at the desired angles to illuminate imaging site 115.


Imaging system 140 can include any type of imager, such as a color imager, a camera, video camera, digital imager, RF imager, etc. Imaging optics 150 can include any combination of one or more lenses, polarizers, filters, mirrors, color filter arrays, and/or other optical elements. In conjunction with illumination source 130, imaging system 140 can collect images of a finger placed at imaging site 115. These images can be collected under a number of different illumination and/or imaging conditions and can produce one or more multispectral images.


Biometric sensor 100 can include or be coupled with a biometric sensor computation system, like the one shown in FIG. 5. The biometric sensor computation system can be used to control imaging system 140 and/or illumination source 130. The biometric sensor computation system can include instructions, executable programs, routines, algorithms etc. that can be used to produce multispectral images, identify an individual, perform spoof detection, and/or check the liveliness of a finger from multispectral images. In particular, the biometric sensor computation system can derive spatially distributed multispectral characteristics from the one or more multispectral images.


In some embodiments, active layer 110 can include a fluid, fluid-like substance, paste, foam, dense liquid, gel, or gel-like substance. In some configurations, active layer 110 can a silicone, polyurethane, a copolymer, and/or a thermoplastic elastomer.


Surface layer 120 can include a thin elastic membrane, film or skin. Surface layer 120 can be a material made from an elastomer, acrylic, latex, artificial rubber, natural rubber, etc. In some embodiments, surface layer 120 can contain or hold a fluid or fluid-like active layer 110 and may be bound with platen 105. In some configurations, surface layer 120 can be painted on active layer 110. In some configurations, surface layer 120 can include a dissolved or diluted silicone, polyurethane, a copolymer, and/or a thermoplastic elastomer with a reflective pigment (e.g., metal paint such as silver, aluminum, bronze, titanium, titanium dioxide, etc.). This dissolved or diluted material can then be painted on active layer 110. Surface layer 120 can be made from a material that conforms with the ridges and valleys of a finger.


Opaque Active Layer and Transparent Surface Layer


In one example, active layer 110 can include an opaque fluid (or fluid like material or gel) and surface layer 120 can include a transparent thin film as shown in FIG. 3. When finger 160 is pressed on surface layer 120, the portion of active layer 110 is pushed aside and surface layer 120 contacts platen 105. Active layer 110 disperses under pressure from finger 160 when finger 160 is presses surface layer 120 onto platen 105. In such embodiments, because surface layer 120 is transparent and active layer 110 is opaque, imaging system 140 can see finger 160 through surface layer 120 when finger 160 is pressed against platen 105. This occurs because opaque active layer 110 has been displaced. That is, optically adaptive interface 102 changes from being opaque to being transparent (or partially transparent) in the area where finger 160 is pressing against platen. With this change, a direct image and/or a TIR image of finger 160 can be captured. In some configurations, active layer 110 can have a relatively high viscosity (e.g., greater than 1.43 centipoise at 18 degrees centigrade).


Furthermore, when finger 160 is pressed against platen 105 finger 160 is visible surrounded by opaque active layer 110. The opaque portions of active layer 110 blocks light from the ambient environment while in the contact state.


Active layer 110 can include inks, dyes, colored polymers, etc. or a combination thereof to produce an opaque material. Various inks and/or dyes can be used to color active layer 110 and/or surface layer 120. In some embodiments, the inks or dyes can be added to ensure that material is sufficiently opaque and/or transmissive with the small thicknesses used in these layers.


In some embodiments, active layer 110 can have a color that is complimentary with the color of skin to produce high contrast images between finger 160 and active layer 120. For example, active layer 110 can be a green or blue layer. In some embodiments, active layer 110 can have a color that is designed to compliment the external environment. In such embodiments, a user can see surface layer 120 and active layer matches 110, or compliments the external environment and/or the external components of biometric sensor 100 for aesthetic purposes.


Low Viscosity, Opaque Active Layer and Transparent Surface Layer


In another example, active layer 110 can include an opaque fluid or gel with a low viscosity and surface layer 120 can include a thin transparent film. The viscosity of the fluid can be low enough and/or the film thin and/or pliable enough that when finger 160 is pressed on surface layer 120, the fluid fills the valleys (or cracks) in finger 160 as shown in FIG. 4. That is, the stiffness of surface layer 120 and/or the fluid pressure of active layer 110 may be such that the active layer disperses only at points of contact (e.g., fingerprint ridges) but remain in the fine negative features (e.g., fingerprint valleys).


Imaging system 140 can then image ridges and valleys of finger 160 through surface layer 120. The valleys are not directly imaged. Instead, active layer 110 is imaged in the valleys of finger 160. Not only can active layer 110 be used to heighten the contrast between ridges and valleys, optically adaptive interface 102 can be useful for blocking the background environment from imaging in areas where finger 160 is not pressed against optically adaptive interface 102.


In some examples, active layer 110 can have a color that is complimentary with the color of skin. For example, active layer 110 can be a green or blue layer. In some embodiments, active layer 110 can have a color that is designed to compliment the external environment. In such embodiments, a user can see active layer 110 through transparent surface layer 120 and the color of active layer 110 matches, or compliments the external environment and/or the external components of biometric sensor 100.


Active layer 110 can include inks, dyes, colored polymers, etc. or a combination thereof to produce an opaque material. Various inks and/or dyes can be used to color active layer 110 and/or surface layer 120. In some embodiments, the inks or dyes can be added to ensure that material is sufficiently opaque and/or transmissive with the small thicknesses used in these layers.


Low Viscosity, Opaque Active Layer and Opaque Surface Layer


In some embodiments, active layer 110 and surface layer 120 can both be opaque and can be made of materials having complementary on contrasting colors. Complementary colors are colors that when mixed produced a neutral color (e.g., grey, white or black). Moreover, complementary colors can provide a high contrast images of the two layers. For example, a low viscosity active layer 110 can be used with a color complementary to the color of surface layer 120. During imaging, the valleys can be identified by the color of active layer 110 and the ridges by the color of surface layer 120. Because the two layers have complementary colors, the contrast between the two will be heightened. As another example, active layer 110 can be a low viscosity layer and translucent at a single wavelength or wavelength band. The wavelength or wavelength band can be complementary to the color of surface layer 120. The valleys will then show up as a neutral color based on the combination of translucent active layer 110 and the color of surface layer 120. The ridges will show up only as the color of active layer 110.


In some configurations, imaging optics can include a color filter array. The colors of active layer 110 and surface layer 120 can be selected to coincide with a pixel color of the color filter array. In such configurations, finger valleys can be identified by isolating pixels corresponding with the color of active layer 110 and finger ridges can be identified by isolating pixels corresponding with the color of surface layer 120.


Surface Layer and Active Layer with Orthogonal Polarity Reflectors


In some embodiments, active layer 110 and surface layer 120 can include polarity reflectors with orthogonal polarizations. In some configurations, active layer can include a low viscosity fluid. In such configurations, ridges and valleys in a finger can be identified based on the polarization of the received light. For example, valleys can correspond with active layer 110 reflecting light of a given polarization and ridges can correspond with surface layer 120 reflecting light with an orthogonal polarization. In other embodiments, illumination source 130 illuminates skin site with unpolarized light. Imaging system 140 can include one or more polarizers (e.g., optical elements 150) that can filter the light reflected from finger 160 based on the polarization of active layer 110 and/or surface layer 120. In some embodiments, the polarization of the polarity reflector in the surface layer is orthogonal with a polarizer used to polarize light prior to imaging.


Surface Layer and Active Layer with Opposite Reflective and Absorption Characteristics


In some embodiments, active layer 110 and surface layer 120 can have opposite reflective and absorption characteristics. Active layer 110 can have a low viscosity. For example, active layer 110 can be absorptive at a given wavelength (or wavelength band) and surface layer 120 can be reflective at the same wavelength (or wavelength band) or vice versa. In such embodiments, ridges and valleys in a skin site can be identified based on the reflective and/or absorption of light at the wavelength and/or wavelength band. For example, ridges will reflect light at the wavelength or in the wavelength band and valleys will absorb light at the wavelength or within the wavelength band. Illumination source 130 can illuminate finger 160 with light having the same wavelength or within the same wavelength band, for example, using a filter. Moreover, imaging system 140 can similarly image the wavelength or wavelength band, for example, using a filter or post process filter. A high contrast image will be produced with contrast between ridges and valleys.


Optically Adaptive Interface with Thermochromatic Material


In some embodiments, one or both of active layer 110 and surface layer 120 can include a thermochromatic material. A thermochromatic material is a material that changes optical properties in response to a change in temperature. A thermochromatic material can change optical properties in response to a change in temperature near the temperature of the human body. For example, the thermochromatic material can become opaque, transparent, change absorptive properties, or change reflective properties at specific wavelengths at temperatures between 93 and 130 degrees Fahrenheit. Thus, when finger 160 makes contact with a thermochromatic material, the material is heated by finger 160 to a higher temperature near the temperature of the human body. In response the material can become transparent, opaque, and/or change color.


In another configuration, an optically adaptive interface can include a single surface layer with thermochromatic properties disposed on a platen. In such configurations, the surface layer can normally be opaque, but can become transparent in response to contact with heat from finger 160. Direct and/or TIR imaging and/or illumination can be used to collect images of finger 160.


Optically Adaptive Interface with Conductive Materials


In some embodiments, either or both active layer 110 and surface layer 120 can include a conductive, semiconductive or resistive coating to act as mechanism for presence detection. The presence of a finger can be detected based on changes in capacitance, resistance, electric field effects or other means. Such a presence detection can provide a wake-up signal to other portions of the sensing system and/or other connected systems, for example, illumination system 130 and/or imaging system 140. For example, surface layer 120 and platen 105 may have conductive coating layers (e.g., transparent on at least the platen surface). When a finger touches the surface layer, some portions of surface layer 120 will directly contact the top of platen 105. This contact can complete an electrical circuit between the bottom of the surface layer 120 and the top of platen 105. This circuit can be monitored to provide a means to turn on the illumination, imaging, embedded processor, and/or other functions of biometric sensor system 100.


Biometric Sensor Computation System



FIG. 5 shows a block diagram of a biometric sensor computation system 500 including a computational device and peripheral devices according to one embodiment. The figure broadly illustrates how individual system elements may be implemented in a separated or more integrated manner. The biometric sensor computation system 500 is shown comprised of hardware elements that are electrically coupled via bus 530. Bus 530, depending on the configuration, may also be coupled with the one or more illumination systems 516 (e.g., illumination source 130), proximity sensor 512 (or presence sensor), and/or imager 518 (e.g., imaging system 140) according to various embodiments.


The hardware elements may include a central processing unit (CPU) 550, input/output device(s) 535, storage device 555, computer-readable storage 540, network interface card (NIC) 545, processing acceleration unit 548 such as a DSP or special-purpose processor, and memory 560. Computer-readable storage 540 may include a computer-readable storage medium and a computer readable medium reader, the combination comprehensively representing remote, local, fixed, and/or removable storage devices plus storage media for temporarily and/or more permanently containing computer-readable information. NIC 545 may comprise a wired, wireless, modem, and/or other type of interfacing connection and permits data to be exchanged with external devices.


Biometric sensor computation system 500 may also comprises software elements, shown as being currently located within memory 560, including an operating system 565 and other programs and/or code 570, such as a program or programs designed to implement methods described herein. Substantial variations in biometric sensor computation system 500 may be used in accordance with specific requirements. For example, customized hardware might also be used and/or particular elements might be implemented in hardware, software (including portable software, such as applets), or both. Further, connection to other computing devices such as network input/output devices may be employed.

Claims
  • 1. A biometric sensor comprising: a platen;an optically adaptive interface disposed on the platen comprising a contact state and a non-contact state, wherein in the non-contact state the optically adaptive interface is opaque and in the contact state a portion of the optically adaptive interface is substantially transparent;an illumination subsystem configured to illuminate a skin site of a purported individual when placed on the optically adaptive interface when the optically adaptive interface is in the contact state; andan imaging subsystem configured to image the skin site when the optically adaptive interface is in the contact state.
  • 2. The biometric sensor according to claim 1, wherein the default state of the optically adaptive interface is in the non-contact state, and the optically adaptive interface changes from the non-contact state to the contact state in response to contact by the skin site.
  • 3. The biometric sensor according to claim 1, wherein the portion of the optically adaptive interface that is substantially transparent in the contact state occurs where the skin site is in contact with the optically adaptive interface.
  • 4. The biometric sensor according to claim 1, wherein the optically adaptive interface blocks light from the illumination subsystem from being transmitted through the optically adaptive interface.
  • 5. The biometric sensor according to claim 1, wherein the optically adaptive interface comprises a surface layer and an active layer.
  • 6. The biometric sensor according to claim 5, wherein the active layer comprises a fluid-like material and the surface layer is a thin film.
  • 7. A biometric sensor comprising: a housing;a platen defining a surface of the housing;an optically adaptive interface disposed on the platen;an illumination subsystem disposed within the housing; andan imaging subsystem disposed within the housing,wherein the optically adaptive interface comprises an opaque surface when viewed from an environment external from the housing, and a portion of the optically adaptive interface changes optical properties when a skin site of an individual is placed on the optically adaptive interface allowing the illumination and imaging subsystems to produce a biometric image of the skin site.
  • 8. The biometric sensor according to claim 7, wherein the optically adaptive interface blocks a user from viewing both the illumination subsystem and the imaging subsystem.
  • 9. The biometric sensor according to claim 7, wherein the optically adaptive interface changes optical properties at the location where the skin site is in contact with the optically adaptive interface.
  • 10. The biometric sensor according to claim 7, wherein the change in the optical properties of the optically adaptive interface is a change from being opaque to substantially transparent.
  • 11. The biometric sensor according to claim 7, wherein the optically adaptive interface comprises a thin surface layer and an fluid-like active layer.
  • 12. A biometric sensor comprising: a platen;an optically adaptive interface disposed on the platen;an illumination subsystem configured to illuminate portions of the platen and the optically adaptive interface; andan imaging subsystem configured to collect light from a skin site of a purported individual when placed on the optically adaptive interface;wherein the optically adaptive interface comprises two states: a non-contact state with a first optical condition, and a contact state with a second optical condition.
  • 13. The biometric sensor according to claim 12, wherein in the contact state only a portion of the optically adaptive interface includes the second optical condition.
  • 14. The biometric sensor according to claim 12, wherein the illumination subsystem includes an optical element that filters light according to either the first optical condition or the second optical condition.
  • 15. The biometric sensor according to claim 12, wherein the optically adaptive interface comprises a surface layer and a fluid-like active layer.
  • 16. The biometric sensor according to claim 12, wherein in the first optical condition the optically adaptive interface absorbs light from the illumination subsystem, and in the second optical condition at least a portion of the optically adaptive interface reflects light from the illumination subsystem.
  • 17. The biometric sensor according to claim 12, wherein in the first optical condition the optically adaptive interface reflects light from the illumination subsystem, and in the second optical condition at least a portion of the optically adaptive interface absorbs light from the illumination subsystem.
  • 18. The biometric sensor according to claim 12, wherein in the first optical condition the optically adaptive interface reflects light having a first wavelength, and in the second optical condition at least a portion of the optically adaptive interface reflects light having a second wavelength distinct from the first wavelength.
  • 19. The biometric sensor according to claim 12, wherein in the first optical condition the optically adaptive interface absorbs light having a first wavelength, and in the second optical condition at least a portion of the optically adaptive interface absorbs light having a second wavelength distinct from the first wavelength.
  • 20. The biometric sensor according to claim 12, wherein in the first optical condition the optically adaptive interface is substantially opaque, and in the second optical condition at least a portion of the optically adaptive interface is substantially transparent.
  • 21. The biometric sensor according to claim 12, wherein in the first optical condition the optically adaptive interface reflects light with a first polarization condition, and in the second optical condition at least a portion of the optically adaptive interface reflects light having a second polarization condition orthogonal with the first polarization condition.
CROSS-REFERENCES TO RELATED APPLICATIONS

This application is a non-provisional, and claims the benefit, of commonly assigned U.S. Provisional Application No. 61/314,563, filed Mar. 16, 2010, entitled “Biometric Imaging Using An Optical Adaptive Interface,” the entirety of which is herein incorporated by reference for all purposes.

US Referenced Citations (356)
Number Name Date Kind
3508830 Hopkins et al. Apr 1970 A
3619060 Johnson Nov 1971 A
3854319 Burroughs et al. Dec 1974 A
3872443 Ott Mar 1975 A
3910701 Henderson et al. Oct 1975 A
RE29008 Ott Oct 1976 E
4035083 Woodriff et al. Jul 1977 A
4142797 Astheimer Mar 1979 A
4169676 Kaiser Oct 1979 A
4170987 Anselmo et al. Oct 1979 A
4260220 Whitehead Apr 1981 A
4322163 Schiller Mar 1982 A
4427889 Muller Jan 1984 A
4537484 Fowler Aug 1985 A
4598715 Machler et al. Jul 1986 A
4653880 Sting et al. Mar 1987 A
4654530 Dybwad Mar 1987 A
4655225 Dahne et al. Apr 1987 A
4656562 Sugino Apr 1987 A
4657397 Oehler et al. Apr 1987 A
4661706 Messerschmidt et al. Apr 1987 A
4684255 Ford Aug 1987 A
4699149 Rice Oct 1987 A
4712912 Messerschmidt Dec 1987 A
4730882 Messerschmidt Mar 1988 A
4747147 Sparrow May 1988 A
4787013 Sugino et al. Nov 1988 A
4787708 Whitehead Nov 1988 A
4830496 Young May 1989 A
4853542 Milosevic et al. Aug 1989 A
4857735 Noller Aug 1989 A
4859064 Messerschmidt et al. Aug 1989 A
4866644 Shenk et al. Sep 1989 A
4867557 Takatani et al. Sep 1989 A
4882492 Schlager Nov 1989 A
4883953 Koashi et al. Nov 1989 A
4936680 Henkes et al. Jun 1990 A
4937764 Komatsu et al. Jun 1990 A
4944021 Hoshino et al. Jul 1990 A
4975581 Robinson et al. Dec 1990 A
5015100 Doyle May 1991 A
5019715 Sting et al. May 1991 A
5028787 Rosenthal et al. Jul 1991 A
5051602 Sting et al. Sep 1991 A
5055658 Cockburn Oct 1991 A
5068536 Rosenthal Nov 1991 A
5070874 Barnes et al. Dec 1991 A
5077803 Kato et al. Dec 1991 A
5088817 Igaki et al. Feb 1992 A
5109428 Igaki et al. Apr 1992 A
5146102 Higuchi et al. Sep 1992 A
5158082 Jones Oct 1992 A
5163094 Prokoski et al. Nov 1992 A
5177802 Fujimoto et al. Jan 1993 A
5178142 Harjunmaa et al. Jan 1993 A
5179951 Knudson Jan 1993 A
5204532 Rosenthal Apr 1993 A
5222495 Clarke et al. Jun 1993 A
5222496 Clarke et al. Jun 1993 A
5223715 Taylor Jun 1993 A
5225678 Messerschmidt Jul 1993 A
5230702 Lindsay et al. Jul 1993 A
5237178 Rosenthal et al. Aug 1993 A
5243546 Maggard Sep 1993 A
5257086 Fateley et al. Oct 1993 A
5258922 Grill Nov 1993 A
5267152 Yang et al. Nov 1993 A
5268749 Weber et al. Dec 1993 A
5291560 Daugman Mar 1994 A
5299570 Hatschek Apr 1994 A
5303026 Strobl et al. Apr 1994 A
5311021 Messerschmidt May 1994 A
5313941 Braig et al. May 1994 A
5321265 Block Jun 1994 A
5331958 Oppenheimer Jul 1994 A
5335288 Faulkner Aug 1994 A
5348003 Caro Sep 1994 A
5351686 Steuer et al. Oct 1994 A
5355880 Thomas et al. Oct 1994 A
5360004 Purdy et al. Nov 1994 A
5361758 Hall et al. Nov 1994 A
5366903 Lundsgaard et al. Nov 1994 A
5372135 Mendelson et al. Dec 1994 A
5379764 Barnes et al. Jan 1995 A
5402778 Chance Apr 1995 A
5405315 Khuri et al. Apr 1995 A
5413098 Benaron et al. May 1995 A
5419321 Evans May 1995 A
5435309 Thomas et al. Jul 1995 A
5441053 Lodder et al. Aug 1995 A
5452723 Wu et al. Sep 1995 A
5459317 Small et al. Oct 1995 A
5459677 Kowalski et al. Oct 1995 A
5460177 Purdy et al. Oct 1995 A
5483335 Tobias Jan 1996 A
5494032 Robinson et al. Feb 1996 A
5505726 Meserol Apr 1996 A
5507723 Keshaviah Apr 1996 A
5515847 Braig et al. May 1996 A
5518623 Keshaviah et al. May 1996 A
5523054 Switalski et al. Jun 1996 A
5533509 Koashi et al. Jul 1996 A
5537208 Bertram et al. Jul 1996 A
5539207 Wong et al. Jul 1996 A
5552997 Massart Sep 1996 A
5559504 Itsumi et al. Sep 1996 A
5568251 Davies et al. Oct 1996 A
5596992 Haaland et al. Jan 1997 A
5606164 Price et al. Feb 1997 A
5613014 Eshera et al. Mar 1997 A
5630413 Thomas et al. May 1997 A
5636633 Messerschmidt et al. Jun 1997 A
5655530 Messerschmidt Aug 1997 A
5672864 Kaplan Sep 1997 A
5672875 Block et al. Sep 1997 A
5677762 Ortyn et al. Oct 1997 A
5681273 Brown Oct 1997 A
5708593 Saby et al. Jan 1998 A
5719399 Alfano et al. Feb 1998 A
5719950 Osten et al. Feb 1998 A
5724268 Sodickson et al. Mar 1998 A
5729619 Puma Mar 1998 A
5737439 Lapsley et al. Apr 1998 A
5743262 Lepper, Jr. et al. Apr 1998 A
5747806 Khalil May 1998 A
5750994 Schlager May 1998 A
5751835 Topping et al. May 1998 A
5751836 Wildes et al. May 1998 A
5761330 Stoianov et al. Jun 1998 A
5782755 Chance et al. Jul 1998 A
5792050 Alam et al. Aug 1998 A
5792053 Skladnev et al. Aug 1998 A
5793881 Stiver et al. Aug 1998 A
5796858 Zhou et al. Aug 1998 A
5808739 Turner et al. Sep 1998 A
5818048 Sodickson et al. Oct 1998 A
5823951 Messerschmidt et al. Oct 1998 A
5828066 Messerschmidt Oct 1998 A
5830132 Robinson Nov 1998 A
5830133 Osten et al. Nov 1998 A
5850623 Carman, Jr. et al. Dec 1998 A
5853370 Chance et al. Dec 1998 A
5857462 Thomas et al. Jan 1999 A
5859420 Borza Jan 1999 A
5860421 Eppstein et al. Jan 1999 A
5867265 Thomas Feb 1999 A
5886347 Inoue et al. Mar 1999 A
5902033 Levis et al. May 1999 A
5914780 Turner et al. Jun 1999 A
5929443 Alfano et al. Jul 1999 A
5933792 Anderson et al. Aug 1999 A
5935062 Messerschmidt et al. Aug 1999 A
5942761 Tuli Aug 1999 A
5945676 Khalil Aug 1999 A
5949543 Bleier et al. Sep 1999 A
5957841 Maruo et al. Sep 1999 A
5961449 Toida et al. Oct 1999 A
5963319 Jarvis et al. Oct 1999 A
5978495 Thomopoulos et al. Nov 1999 A
5987346 Benaron et al. Nov 1999 A
5999637 Toyoda et al. Dec 1999 A
6005722 Butterworth et al. Dec 1999 A
6016435 Maruo et al. Jan 2000 A
6025597 Sterling et al. Feb 2000 A
6026314 Amerov et al. Feb 2000 A
6028773 Hundt Feb 2000 A
6031609 Funk et al. Feb 2000 A
6034370 Messerschmidt Mar 2000 A
6040578 Malin et al. Mar 2000 A
6041247 Weckstrom et al. Mar 2000 A
6041410 Hsu et al. Mar 2000 A
6043492 Lee et al. Mar 2000 A
6044285 Chaiken et al. Mar 2000 A
6045502 Eppstein et al. Apr 2000 A
6046808 Fateley Apr 2000 A
6049727 Crothall Apr 2000 A
6056738 Marchitto et al. May 2000 A
6057925 Anthon May 2000 A
6061581 Alam et al. May 2000 A
6061582 Small et al. May 2000 A
6066847 Rosenthal May 2000 A
6069689 Zeng et al. May 2000 A
6070093 Oosta et al. May 2000 A
6073037 Alam et al. Jun 2000 A
6081612 Gutkowicz-Krusin et al. Jun 2000 A
6088605 Griffith et al. Jul 2000 A
6088607 Diab et al. Jul 2000 A
6097035 Belongie et al. Aug 2000 A
6100811 Hsu et al. Aug 2000 A
6115484 Bowker et al. Sep 2000 A
6115673 Malin et al. Sep 2000 A
6122042 Wunderman et al. Sep 2000 A
6122394 Neukermans et al. Sep 2000 A
6122737 Bjorn et al. Sep 2000 A
6125192 Bjorn et al. Sep 2000 A
6141101 Bleier et al. Oct 2000 A
6147749 Kubo et al. Nov 2000 A
6148094 Kinsella Nov 2000 A
6152876 Robinson et al. Nov 2000 A
6154658 Caci Nov 2000 A
6157041 Thomas et al. Dec 2000 A
6159147 Lichter et al. Dec 2000 A
6172743 Kley et al. Jan 2001 B1
6175407 Sartor Jan 2001 B1
6181414 Raz et al. Jan 2001 B1
6181958 Steuer et al. Jan 2001 B1
6188781 Brownlee Feb 2001 B1
6193153 Lambert Feb 2001 B1
6208749 Gutkowicz-Krusin Mar 2001 B1
6212424 Robinson Apr 2001 B1
6226541 Eppstein et al. May 2001 B1
6229908 Edmonds et al. May 2001 B1
6230034 Messerschmidt et al. May 2001 B1
6236047 Malin et al. May 2001 B1
6240306 Rohrscheib et al. May 2001 B1
6240309 Yamashita et al. May 2001 B1
6241663 Wu et al. Jun 2001 B1
6256523 Diab et al. Jul 2001 B1
6272367 Chance Aug 2001 B1
6280381 Malin et al. Aug 2001 B1
6282303 Brownlee Aug 2001 B1
6285895 Ristolainen et al. Sep 2001 B1
6292576 Brownlee Sep 2001 B1
6301375 Choi Oct 2001 B1
6301815 Sliwa Oct 2001 B1
6304767 Soller et al. Oct 2001 B1
6307633 Mandella et al. Oct 2001 B1
6309884 Cooper et al. Oct 2001 B1
6317507 Dolfing Nov 2001 B1
6324310 Brownlee Nov 2001 B1
6330346 Peterson et al. Dec 2001 B1
6404904 Einighammer et al. Jun 2002 B1
6419361 Cabib et al. Jul 2002 B2
6483929 Murakami et al. Nov 2002 B1
6504614 Messerschmidt et al. Jan 2003 B1
6537225 Mills Mar 2003 B1
6560352 Rowe et al. May 2003 B2
6574490 Abbink et al. Jun 2003 B2
6597945 Marksteiner Jul 2003 B2
6606509 Schmitt Aug 2003 B2
6628809 Rowe et al. Sep 2003 B1
6631199 Topping et al. Oct 2003 B1
6741729 Bjorn et al. May 2004 B2
6749115 Gressel et al. Jun 2004 B2
6799275 Bjorn Sep 2004 B1
6799726 Stockhammer Oct 2004 B2
6816605 Rowe et al. Nov 2004 B2
6825930 Cronin et al. Nov 2004 B2
6853444 Haddad Feb 2005 B2
6898299 Brooks May 2005 B1
6928181 Brooks Aug 2005 B2
6937885 Lewis et al. Aug 2005 B1
6958194 Hopper et al. Oct 2005 B1
6995384 Lee et al. Feb 2006 B2
7047419 Black May 2006 B2
7084415 Iwai Aug 2006 B2
7147153 Rowe et al. Dec 2006 B2
7254255 Dennis Aug 2007 B2
7263213 Rowe Aug 2007 B2
7287013 Schneider et al. Oct 2007 B2
7347365 Rowe Mar 2008 B2
7366331 Higuchi Apr 2008 B2
7386152 Rowe et al. Jun 2008 B2
7394919 Rowe et al. Jul 2008 B2
7397943 Merbach et al. Jul 2008 B2
7440597 Rowe Oct 2008 B2
7460696 Rowe Dec 2008 B2
7508965 Rowe et al. Mar 2009 B2
7515252 Hernandez Apr 2009 B2
7539330 Rowe May 2009 B2
7545543 Herloski et al. Jun 2009 B2
7545963 Rowe Jun 2009 B2
7627151 Rowe Dec 2009 B2
7668350 Rowe Feb 2010 B2
7735729 Rowe Jun 2010 B2
7751594 Rowe et al. Jul 2010 B2
7801338 Rowe Sep 2010 B2
7801339 Sidlauskas et al. Sep 2010 B2
7804984 Sidlauskas et al. Sep 2010 B2
7819311 Rowe et al. Oct 2010 B2
7831072 Rowe Nov 2010 B2
7835554 Rowe Nov 2010 B2
7899217 Uludag et al. Mar 2011 B2
7995808 Rowe et al. Aug 2011 B2
20020009213 Rowe et al. Jan 2002 A1
20020065468 Utzinger et al. May 2002 A1
20020101566 Elsner et al. Aug 2002 A1
20020111546 Cook et al. Aug 2002 A1
20020138768 Murakami et al. Sep 2002 A1
20020171834 Rowe et al. Nov 2002 A1
20020183624 Rowe et al. Dec 2002 A1
20030025897 Iwai Feb 2003 A1
20030044051 Fujieda Mar 2003 A1
20030078504 Rowe Apr 2003 A1
20030095525 Lavin et al. May 2003 A1
20030128867 Bennett Jul 2003 A1
20030163710 Ortiz et al. Aug 2003 A1
20030223621 Rowe et al. Dec 2003 A1
20040003295 Elderfield et al. Jan 2004 A1
20040008875 Linares Jan 2004 A1
20040022421 Endoh et al. Feb 2004 A1
20040042642 Bolle et al. Mar 2004 A1
20040047493 Rowe et al. Mar 2004 A1
20040114783 Spycher et al. Jun 2004 A1
20040120553 Stobbe Jun 2004 A1
20040125994 Engels et al. Jul 2004 A1
20040179722 Moritoki et al. Sep 2004 A1
20040240712 Rowe et al. Dec 2004 A1
20040240713 Hata Dec 2004 A1
20040264742 Zhang et al. Dec 2004 A1
20050007582 Villers et al. Jan 2005 A1
20050125339 Tidwell et al. Jun 2005 A1
20050169504 Black Aug 2005 A1
20050180620 Takiguchi Aug 2005 A1
20050185847 Rowe Aug 2005 A1
20050205667 Rowe Sep 2005 A1
20050265585 Rowe Dec 2005 A1
20050265586 Rowe et al. Dec 2005 A1
20050265607 Chang Dec 2005 A1
20050271258 Rowe Dec 2005 A1
20060002597 Rowe Jan 2006 A1
20060002598 Rowe et al. Jan 2006 A1
20060045330 Marion Mar 2006 A1
20060062438 Rowe Mar 2006 A1
20060110015 Rowe May 2006 A1
20060115128 Mainguet Jun 2006 A1
20060171571 Chan et al. Aug 2006 A1
20060173256 Ridder et al. Aug 2006 A1
20060202028 Rowe Sep 2006 A1
20060210120 Rowe Sep 2006 A1
20060244947 Rowe Nov 2006 A1
20060274921 Rowe Dec 2006 A1
20070014437 Sato Jan 2007 A1
20070030475 Rowe et al. Feb 2007 A1
20070052827 Hiltunen Mar 2007 A1
20070116331 Rowe et al. May 2007 A1
20070153258 Hernandez Jul 2007 A1
20070165903 Munro et al. Jul 2007 A1
20080008359 Beenau et al. Jan 2008 A1
20080013806 Hamid Jan 2008 A1
20080025579 Sidlauskas et al. Jan 2008 A1
20080025580 Sidlauskas et al. Jan 2008 A1
20080192988 Uludag et al. Aug 2008 A1
20080232653 Rowe Sep 2008 A1
20080260211 Bennett et al. Oct 2008 A1
20080298649 Ennis et al. Dec 2008 A1
20090046903 Corcoran et al. Feb 2009 A1
20090080709 Rowe et al. Mar 2009 A1
20090092290 Rowe Apr 2009 A1
20090148005 Rowe Jun 2009 A1
20090245591 Rowe et al. Oct 2009 A1
20100067748 Rowe Mar 2010 A1
20100246902 Rowe et al. Sep 2010 A1
20110085708 Martin et al. Apr 2011 A1
20110211055 Martin et al. Sep 2011 A1
20110235872 Rowe et al. Sep 2011 A1
Foreign Referenced Citations (66)
Number Date Country
1307711 Aug 2001 CN
1402183 Mar 2003 CN
1509454 Jun 2004 CN
10153808 May 2003 DE
0 280 418 Aug 1988 EP
0 372 748 Jun 1990 EP
0 426 358 May 1991 EP
0 449 335 Oct 1991 EP
0 573 137 Dec 1993 EP
0 631 137 Dec 1994 EP
0 670 143 Sep 1995 EP
0 681 166 Nov 1995 EP
0 757 243 Feb 1997 EP
0 788 000 Aug 1997 EP
0 801 297 Oct 1997 EP
0 836 083 Apr 1998 EP
0 843 986 May 1998 EP
0 869 348 Oct 1998 EP
0 897 164 Feb 1999 EP
0 897 691 Feb 1999 EP
0 317 121 May 1999 EP
0 924 656 Jun 1999 EP
0 982 583 Mar 2000 EP
0 990 945 Apr 2000 EP
1 353 292 Oct 2003 EP
1 434 162 Jun 2004 EP
2761180 Jan 1998 FR
61182174 Aug 1986 JP
3016160 Jan 1991 JP
7075629 Mar 1995 JP
10-127585 May 1998 JP
2001-112742 Apr 2001 JP
2001-184490 Jul 2001 JP
2001033381 Sep 2001 JP
2002-133402 May 2002 JP
2002-517835 Jun 2002 JP
2003050993 Feb 2003 JP
2003-511101 Mar 2003 JP
2003-308520 Oct 2003 JP
WO 9200513 Jan 1992 WO
WO 9217765 Oct 1992 WO
WO 9300855 Jan 1993 WO
WO 9307801 Apr 1993 WO
WO 9927848 Jun 1999 WO
WO 0030530 Jun 2000 WO
WO 0046739 Aug 2000 WO
WO 0115596 Mar 2001 WO
WO 0118332 Mar 2001 WO
WO 0120538 Mar 2001 WO
WO 0127882 Apr 2001 WO
WO 0152180 Jul 2001 WO
WO 0152726 Jul 2001 WO
WO 0153805 Jul 2001 WO
WO 0165471 Sep 2001 WO
WO 0169520 Sep 2001 WO
WO 02054337 Jul 2002 WO
WO 02084605 Oct 2002 WO
WO 02099393 Dec 2002 WO
WO 03010510 Feb 2003 WO
WO 03096272 Nov 2003 WO
WO 2004068388 Aug 2004 WO
WO 2004068394 Aug 2004 WO
WO 2004090786 Oct 2004 WO
WO 2006049394 May 2006 WO
WO 2006077446 Jul 2006 WO
WO 2006093508 Sep 2006 WO
Non-Patent Literature Citations (49)
Entry
Anderson, C. E. et al., “Fundamentals of Calibration Transfer Through Procrustes Analysis,” Appln. Spectros., vol. 53, No. 10, 1999, pp. 1268-1276.
Author Unknown, “Improve the Clinical Outcome of Every Patient,” In Line Diagnostics, published on or before Oct. 30, 1997, 2 pages.
Ashboum, Julian, Biometrics; Advanced Identity Verification, Springer, 2000, pp. 63-64.
Bantle, John P. et al., “Glucose Measurement in Patients With Diabetes Mellitus With Dermal Interstitial Fluid,” Mosby-Year Book, Inc., 1997, 9 pages.
Berkoben, Michael S. et al., “Vascular Access for Hemodialysis,” Clinical Dialysis, Third Edition, 1995, pp. 2 cover pages and 26-45.
Blank, T.B. et al., “Transfer of Near-Infrared Multivariate Calibrations Without Standards,” Anal. Chem., vol. 68, 1996, p. 2987.
Bleyer, Anthony J. et al., “The Costs of Hospitalizations Due to Hemodialysis Access Management,” Nephrology News & Issues, Jan. 1995, pp. 19, 20 and 22.
Brasunas, John C. et al., “Uniform Time-Sampling Fourier Transform Spectroscopy,” Applied Optics, vol. 36, No. 10, Apr. 1, 1997, pp. 2206-2210.
Brault, James W., “New Approach to High-Precision Fourier Transform Spectrometer Design,” Applied Optics, Vo. 35, No. 16, Jun. 1, 1996, pp. 2891-2896.
Cassarly, W.J. et al., “Distributed Lighting Systems: Uniform Light Delivery,” Source Unknown, 1995, pp. 1698-1702.
Chang, Chong-Min et al., “An Uniform Rectangular Illuminating Optical System for Liquid Crystal Light Valve Projectors,” Euro Display '96, 1996, pp. 257-260.
Coyne, Lawrence J. et al., “Distributive Fiber Optic couplers Using Rectangular Lightguides as Mixing Elements,” Information Gatekeepers, Inc. Brookline, MA, 1979, pp. 160-164.
Daugirdas, JT et al., “Comparison of Methods to Predict the Equilibrated Kt/V (eKt/V) in the Hemo Study,” National Institutes of Health, Aug. 20, 1996, pp. 1-28.
De Noord, Onno E., “Multivariate Calibration Standardization,” Chemometrics and intelligent Laboratory Systems 25, 1994, pp. 85-97.
Demos, S. G. et al., “Optical Fingerprinting Using Polarisation Contrast Improvement,” Electronics Letters, vol. 33, No. 7, Mar. 27, 1997, pp. 582-584.
Depner, Thomas A. et al., “Clinical Measurement of Blood Flow in Hemodialysis Access Fistulae and Grafts by Ultrasound Dilution,” Division of Nephrology, University of California, published on or before Oct. 30, 1997, pp. M745-M748.
Despain, Alvin M. et al., “A Large-Aperture Field-Widened Interferometer-Spectrometer for Airglow Studies,” Aspen International Conference on Fourier Spectroscopy, 1970, pp. 293-300.
Faber, Nicolaas, “Multivariate Sensitivity for the Interpretation of the Effect of Spectral Pretreatment Methods on Near-Infrared Calibration Model Predictions,” Analytical Chemistry, vol. 71, No. 3, Feb. 1, 1999, pp. 557-565.
Fresenius USA, “Determination of Delivered Therapy Through Measurement of Effective Clearance,” Dec. 1994, 2 pages.
Geladi, Paul et al., “A Multivariate NIR Study of Skin Alterations in Diabetic Patients as Compared to Control Subj ects,Fresenius USA, Determination of Delivered Therapy Through Measurement of Effective Clearance,” 2 pages, Dec. 1994, J. Near Infrared Spectrosc., vol. 8, 2000, pp. 217-227.
Hakim, Raymond M. et al., “Effects of Dose of Dialysis on Morbidity and Mortality,” American Journal of Kidney Diseases, vol. 23, No. 5, May 1994, pp. 661-669.
International Search Report and Written Opinion of PCT/US2008/066585 mailed Oct. 30, 2008, 10 pages.
International Search Report of PCT/US2010/046852 mailed Dec. 29, 2010, 5 pages.
Jacobs, Paul et al., “A Disposable Urea Sensor for Continuous Monitoring of Hemodialysis Efficiency,” ASAIO Journal, 1993, pp. M353-M358.
Keshaviah, Prakash R. et al., “On-Line Monitoring of the Delivery of the Hemodialysis Prescription,” Pediatric Nephrology, vol. 9, 1995, pp. S2-S8.
Krivitski, Nikolai M., “Theory and Validation of Access Flow Measurement by Dilution Technique During Hemodialysis,” Kidney International, vol. 48, 1995, pp. 244-250.
Lee et al., “Fingerprint Recognition Using Principal Gabor Basis Function”, Proceedings of 2001 International Symposium on Intelligent Multimedia, Video and Speech Processing, May 2-4, 2001, Sections 2-3.
Maltoni et al., “Handbook of Fingerprint Recognition,” 2005, pp. 58-61.
Marbach, Ralf, “Measurement Techniques for IR Spectroscopic Blood Glucose Determination,” Fortschritt Bericht, Series 8: Measurement and Control Technology, No. 346, Mar. 28, 1994, pp. cover and 1-158.
Mardia, K.V. et al., “Chapter 11—Discriminant Analysis,” Multivariate Analysis, 1979, pp. 2 cover pages and 300-325.
Nichols, Michael G. et al., “Design and Testing of a White-Light, Steady-State Diffuse Reflectance Spectrometer for Determination of Optical Properties of Highly Scattering Systems,” Applied Optics, vol. 36, No. 1, Jan. 1, 1997, pp. 93-104.
Nixon, Kristin A. et al., “Novel Spectroscopy-Based Technology for Biometric and Liveness Verification,” Technology for Human Identification. Proceedings of SPIE, vol. 5404, No. 1, XP-002458441, Apr. 12-13, 2004, pp. 287-295 (ISSN: 0277-786x).
Pan et al., “Face Recognition in Hyperspectral Images,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 25, No. 12, Dec. 2003 pp. 1552-1560.
Ripley, B. D., “Chapter 3—Linear Discriminant Analysis,” Pattern Recognition and Neural Networks, 1996, pp. 3 cover pages and 91-120.
Ronco, C. et al., “On-Line Urea Monitoring : A Further Step Towards Adequate Dialysis Prescription and Delivery,” The International Journal of Artificial Organs, vol. 18, No. 9, 1995, pp. 534-543.
Ross et al., “A Hybrid Fingerprint Matcher,” Pattern Recognition 36, The Journal of the Pattern Recognition Society, 2003 Elsevier Science Ltd., pp. 1661-1673.
Selvaraj et al., “Fingerprint Verification Using Wavelet Transform,” Proceedings of the Fifth International Conference on Computational Intelligence and Multimedia Applications, IEEE, 2003, 6 pages.
Service, F. John et al., “Dermal Interstitial Glucose as an Indicator of Ambient Glycemia,” Diabetes Care, vol. 20, No. 9, Aug. 1997, 8 pages.
Sherman, Richard A., “Chapter 4—Recirculation in the Hemodialysis Access,” Principles and Practice of Dialysis, pp. 2 cover pages and 38-46, 1994.
Sherman, Richard A., “The Measurement of Dialysis Access Recirculation,” American Journal of Kidney Diseases, vol. 22, No. 4, Oct. 1993, pp. 616-621.
Steuer, Robert R. et al., “A New Optical Technique for Monitoring Hematocrit and Circulating Blood Volume: Its Application in Renal Dialysis,” Dialysis & Transplantation, vol. 22, No. 5, May 1993, pp. 260-265.
Webb, Paul, “Temperatures of Skin, Subcutaneous Tissue, Muscle and Core in Resting Men in Cold, Comfortable and Hot Conditions,” European Journal of Applied Physiology, vol. 64, 1992, pp. 471-476.
Zavala, Albert et al., “Using Fingerprint Measures to Predict Other Anthropometric Variables,” Human Factors, vol. 17, No. 6, 1975, pp. 591-602.
Chinese Patent Application No. 2006/80038597.4, First Office Action mailed on Mar. 23, 2011, 7 pages.
European Patent Application No. 10166537.0, Extended European Search Report mailed on Jun. 1, 2011, 7 pages.
International Search Report and Written Opinion of PCT/US2010/025463 mailed on Jun. 30, 2010, 12 pages.
Rowe, et al. “Multispectral Fingerprint Image Acquisition,” Advance in Biometrics, 2008, 22 pages.
Chinese Patent Application No. 2006/80038579.4, First Office Action mailed on Mar. 23, 2011, 7 pages.
Rowe, “LumiGuard: A Novel Spectroscopic Sensor for Biometric Security Applications”, American Chemical Society 225th National Meeting, Mar. 25, 2003, 20 pages.
Related Publications (1)
Number Date Country
20120062364 A1 Mar 2012 US
Provisional Applications (1)
Number Date Country
61314563 Mar 2010 US