Biometric object sensor and method

Information

  • Patent Grant
  • 9697411
  • Patent Number
    9,697,411
  • Date Filed
    Wednesday, September 9, 2015
    9 years ago
  • Date Issued
    Tuesday, July 4, 2017
    7 years ago
Abstract
A biometric image sensor system and method is disclosed which may comprise a first light source having a first wavelength λ1, a second light source having a second wavelength λ2, a photodetector configured and positioned to receive light of the first wavelength λ1 and light of the second wavelength λ2 reflecting from a biometric object being imaged and to produce a first output indicative of the amount of reflectance of the light of the first wavelength λ1 and a second output indicative of the amount of reflectance of the light of the second wavelength λ2, and a computing device configured to compare the difference between the first output and the second output with an authenticity threshold.
Description
BACKGROUND OF THE INVENTION

In the field of biometric sensors, such as fingerprint sensors, efforts are made to fool of “spoof” an image sensor/collector and image producing system and method, e.g., by forming a fake finger, such as out of a material like so-called “silly putty” and forming on the fake finger an image of a fingerprint in an effort to falsely obtain access to an electronic device protected by a biometric image sensor and comparison system and method. There exists in the art a need for a system and method for authenticating that the biometric object being imaged for comparison is, in fact, part of a human body from which the biometric sensor is detecting the image.


One example of a fingerprint swipe sensor is described in U.S. Pat. No. 6,289,114 entitled FINGERPRINT-READING SYSTEM. This patent describes a system in which the surface area of the sensor is far smaller than the surface area of the fingerprint to be read. The reading is done when the sensor and the finger are in contact and in a relative motion of sliding of the sensor and the finger with respect to each other. The system reconstitutes a complete image of the fingerprint from the partial images given by the sensor during this motion, utilizing pressure and/or temperature from the finger being sensed. The manner in which the system reconstitutes a complete image of the fingerprint from the partial images given by the sensor is not described. Another example of a swiped sensor utilizing capacitive coupling through features of the biometric being imaged, e.g., fingerprint ridges and valleys can be found in U.S. Pat. No. 7,099,496, entitled SWIPED APERTURE CAPACITIVE FINGERPRINT SENSING SYSTEMS AND METHODS.


As discussed in U.S. Pat. No. 5,940,526 to Setlak et al., issued on Aug. 17, 1999, entitled, ELECTRIC FIELD FINGERPRINT SENSOR HAVING ENHANCED FEATURES AND RELATED METHODS, an integrated circuit fingerprint sensor including an array of RF sensing electrodes can provide an image of the fingerprint ridges and valleys. The RF sensing permits imaging of live tissue, including just below the surface of the skin, and in that sense itself is a primative form of reducing spoofing. Setlak et al. is incorporated herein by reference.


SUMMARY OF THE INVENTION

A biometric image sensor system and method is disclosed which may comprise a first light source having a first wavelength λ1, a second light source having a second wavelength λ2, a photodetector configured and positioned to receive light of the first wavelength λ1 and light of the second wavelength λ2 reflecting from a biometric object being imaged and to produce a first output indicative of the amount of reflectance of the light of the first wavelength λ1 and a second output indicative of the amount of reflectance of the light of the second wavelength λ2, and a computing device configured to compare the difference between the first output and the second output with an authenticity threshold. The biometric object may be a finger. The finger may be being imaged to obtain a fingerprint image. The wavelengths λ1 and λ2 may be selected to coincide with a section of relatively high change in the reflectance vs. wavelength curve for the biometric object. The biometric object may comprise human skin. The wavelength λ1 may be in the visible range of yellow light and the wavelength λ2 may be in the visible range of one of orange light and red light.


A biometric image sensor system and method is disclosed which may comprise one or more light sources wherein the one or more light sources is configurable to have a first wavelength λ1, and a second wavelength λ2, one or more photodetectors configurable and positionable to receive one or more electromagnetic radiation wavelengths reflected from a surface in response to the first wavelength λ1 and the second wavelength λ2 wherein the reflected electromagnetic radiation is reflecting from a biometric object being imaged to produce a first output indicative of the amount of reflectance of the light of the first wavelength λ1 and a second output indicative of the amount of reflectance of the light of the second wavelength λ2, and a computing device configured to compare the difference between the first output and the second output with an authenticity threshold. The biometric object may be a finger. The finger may be being imaged to obtain a fingerprint image. The wavelengths λ1 and λ2 may be selected to coincide with a section of relatively high change in the reflectance vs. wavelength curve for the biometric object. The biometric object may comprise human skin. The wavelength λ1 may be in the visible range of yellow light and the wavelength λ2 may be in the visible range of one of orange light and red light.


INCORPORATION BY REFERENCE

All publications, patents, and patent applications mentioned in this specification are herein incorporated by reference to the same extent as if each individual publication, patent, or patent application was specifically and individually indicated to be incorporated by reference, for all purposes, and to the same extent as if fully reproduced in the present application.





BRIEF DESCRIPTION OF THE DRAWINGS

The novel features of the invention are set forth with particularity in the appended claims. A better understanding of the features and advantages of the present invention will be obtained by reference to the following detailed description that sets forth illustrative embodiments, in which the principles of the invention are utilized, and the accompanying drawings of which:



FIG. 1 illustrates a reflectivity vs. wavelength curve for a plurality of human fingers from different individuals;



FIG. 2 illustrates the differences in reflectivity for “spoofing” materials that, e.g., generally increase in reflectivity as wavelength increases or decrease in reflectivity as wavelength increases;



FIG. 3 shows in schematic and block diagram form a circuit for distinguishing between a “spoofing” material and a human finger being imaged;



FIG. 4 shows an illustration of absorption varying with wavelength for oxygenated blood and de-oxygenated blood;



FIG. 5 illustrates a system and method for detecting a pulse from the presence or absence of oxygenated blood;



FIG. 6 illustrates schematically the detection of light reflecting from the biogenic object being sensed, such as a human finger.





DETAILED DESCRIPTION OF THE INVENTION

Turning now to FIG. 1 there is shown a plurality of reflectivity vs. wavelength curves 10 for a plurality of human fingers from different individuals. It will be noted that, while reflectivities vary as much as about 10% to 20% between individuals, all human skin exhibits a relative steep rate of increase in reflectivity for wavelengths from about 575 in the yellow visible light range 12 to about 650 in the red visible light range 14, with the intervening orange visible light range 16. That is, the reflectivity (amount or percentage of reflectance) increases from about 45% to 60% to about 100%, i.e., the reflectance about doubles in this range of light wavelengths. It can be seen that a similar relatively steep increase occurs from the wavelengths of about 440-510, though not quite as pronounced an increase as for the 575-650 range, covering yellow light to red light and the intervening orange light.


Turning now to FIG. 2 there is illustrated the differences in reflectivity for “spoofing” materials that, e.g., as illustrated in curve 20 generally increase in reflectivity as wavelength increases or, as illustrated in curve 30 generally decrease in reflectivity as wavelength increases. As can be seen from the representative human finger response curve 10 in FIG. 2, a difference between the response to light in the yellow range 12, e.g., at about 575 nm wavelength, i.e., about 50% reflectivity and light in the red range, e.g., at about 675 nm, i.e., about 95% gives a difference of about 45%. On the other hand, the difference for the spoofing material A, i.e., on curve 20, is between about 83% and about 95% and the difference for the spoofing material B, i.e., curve 30 is between about 40% and 53%. Thus, the absolute value of the differences for the spoofing materials A, curve 20, and B, curve 30, at the noted wavelengths, is about one fourth the difference for actual human skin. Therefore, an authenticity threshold can be selected, such as greater than around a 1.5× ratio, being indicative of a reflection of the two wavelengths of light being from human skin, and less than the authenticity threshold not being a reflection of the two light wavelengths from human skin.



FIG. 3 shows in schematic and block diagram form a circuit 50 for distinguishing between a “spoofing” material and a human finger being imaged. In an aspect of the disclosure, the circuit 50 can include a computing device, such as a microprocessor or microcontroller 52, that can serve to control one or more light sources, such as a pair of light sources comprising an orange or red light emitting diode (“LED”) 60 and a yellow LED 70. The microprocessor 52 may control the one or more light sources, or different LEDs 60, 70 to be on at different times while a biometric object of a user is being sensed, such as when a user is swiping a finger over a fingerprint sensor as part of a fingerprint imaging system and method. In another aspect, the circuit 50 can include a computing device, such as a microprocessor or microcontroller 52, that can serve to control a single light source capable of delivering or emitting more than one frequency of light. The microprocessor 52 may control the light source to be on emit two or more wavelengths of light at different times while a biometric object of a user is being sensed, such as when a user is swiping a finger over a fingerprint sensor as part of a fingerprint imaging system and method. As will be appreciated by those skilled in the art, one or more sensors can be configured to be used with one or more light sources, and one or more microprocessors or microcontrollers to provide redundancy to the system if desired without departing from the scope of the disclosure. Additionally, the circuit 50 can further be combined with one or more additional biometric sensors, such as fingerprint swipe sensors described in U.S. Pat. No. 6,289,114.


The LEDs could be, e.g., Red, 610<λ<760, Aluminium gallium arsenide (AlGaAs), Gallium arsenide phosphide (GaAsP), Aluminium gallium indium phosphide (AlGaInP), and Gallium(III) phosphide (GaP), Orange, 590<λ<610, Gallium arsenide phosphide (GaAsP), Aluminium gallium indium phosphide (AlGaInP), Gallium(III) phosphide (GaP) and Yellow, 570<λ<590, Gallium arsenide phosphide (GaAsP), Aluminium gallium indium phosphide (AlGaInP), and Gallium(III) phosphide (GaP) with suitable coating of a phosphor or scintillator. In other configurations a single light source capable of emitting two or more wavelengths can be used.


At the same time, a photodetector 80 may be attached to a voltage supply 82 causing a sensor current 84 to pass through the detector 80 which, e.g., accumulates charge from incident photons. The incident photons are being collected in the photodetector 80 and the photodetector 80 provides an analog output to an A/D converter 90, providing a digital output to the microprocessor 52. The digital output is alternatively representative of light of the respective wavelength from the respective yellow LED 70 and orange LED 60, as examples, which has reflected from a finger in the vicinity of the biometric object detector, e.g., reflected off of a finger being scanned for a fingerprint image.


It will be understood that in operation the circuit 50 provides a measurement of the reflectivity of the biometric object measured by the amount of the output signal from the photodiode 80 responsive to, e.g., the orange LED being turned on for a period of time and the e.g., yellow LED 70 being turned on for a time, e.g., both for the same amount of time at different time periods within a sampling period. In some embodiments it may be necessary, as shown in FIG. 3, to provide for some signal processing, such as in block 82, between the photodiode 80 and the A/D converter 90. Such signal processing, as is well known in the art, can provide, e.g., suitable filtering and amplification of the output of the photodiode 80 to improve, e.g., the gain and/or dynamic range of the output of the A/D converter, or both.


The absolute values of the amounts of reflected light collected for the, e.g., yellow light emitted by LED 70 and orange light emitted by LED 60 can then be compared in the microprocessor 52 against a selected authenticity threshold. As an example, if the indicated reflectivity of the orange light is more than, e.g., three times that of the yellow light, then the reflecting object is determined to be a human finger, and, if not it is rejected. Thus, difference of the reflections of “spoofing” material A, curve 20, i.e., about 10% or for “spoofing” material B, curve 30, i.e., about 12, each would fail to meet a threshold of, e.g., 1.5X.



FIG. 4 shows an illustration of absorption varying with wavelength for oxygenated blood and de-oxygenated blood. It can be seen that in the range of about yellow to red light, and most prominently within the red band, there is a relatively large difference between the reflectivity of oxygenated blood HbO2, curve 102, and non-oxygenated blood Hb, curve 104. Reflectivity here is the inverse of the illustrated absorption coefficient. Thus, by shining red light on the finger or other part of the body being imaged for a biometric identification, and measuring the reflectivity an output curve 120 from the photodetector, such as 80 in FIG. 3, can be utilized to detect a significant drop in the output, indicating less absorption (i.e., more reflectivity).



FIG. 5 illustrates a system and method for detecting a pulse from the presence or absence of oxygenated blood. It will be understood that the biometric image sensor may be operating over too short a time to detect sequential drops in absorption 22 and 124, and therefore measure a pulse rate for the individual. However, the individual could be prompted or otherwise directed to leave the biometric object in place, e.g., keep the finger on the object sensor, and, therefore, have the pulse of the user detected. Alternatively, the user could be prompted or otherwise directed to swipe the object several times, in order to detect at least one change 122, 124 in the photodetector output, as another “anti-spoofing” detector. That is, assuming the usual user has a resting pulse rate of 60, such a change in absorption 122, 124 will occur at a rate of once per second. In the case of a fingerprint image detector utilizing swiping of the finger over a sensor, and assuming that the swipe takes one half a second, on average for every two swipes the detector should see at least the onset or the discontinuance of one such change 122124. Detecting such a change, i.e., the onset of the indication of the decrease in absorption or the return to the higher level of absorption, i.e., the leading or trailing edges of the respective one of the blips 122, 124 can indicate a change from non-oxygenated blood to oxygenated blood or vice versa, a further indicator that the object being sensed is part of a human body.



FIG. 6 illustrates schematically the detection of light reflecting from the biogenic object being sensed, such as a human finger. As can be seen, the photodetector, such as 80 in FIG. 3 should be placed relatively adjacent the respective one or more light sources, illustrated as LED 60, in order to be sensitive to light 132, 134 that is actually reflecting from the biometric object, such as finger 130, as opposed to traveling into the finger 130 and then exiting the finger, such as light 136.


Depending upon the timing of the energization of the LEDs for the anti-snooping, one of the LEDs for the anti-snooping by detection of reflection from human skin may also be utilized for anti-snooping by detection of the periodic presence of oxygenated blood. Also, at least one of the LEDs could be utilized to detect the pulse rate of the user, by having the user maintain the biometric object, such as the finger over the sensor, i.e., not swipe the finger. Applicants' assignee has previously filed applications for waking the sensor from “sleep” when no biometric object has been detected in the vicinity of the sensor for some period of time, as a power saving measure, whereby the biometric object sensor comes out of a sleep mode and begins to detect image portions for assembling an image of all or most of a fingerprint, as is discussed in SYSTEM AND METHOD FOR MINIMIZING POWER CONSUMPTION FOR AN OBJECT SENSOR, U.S. Pat. No. 7,643,950, issued on Jan. 5, 2010, and ULTRA LOW POWER WAKE-ON-EVENT MODE FOR BIOMETRIC SYSTEMS, U.S. Pat. App. Pub. No. 2010/0180136, published on Jul. 15, 2010.


There are many ways to detect the increase in oxygenated blood, i.e., decrease in percentage absorption as illustrated in FIG. 5, representing a human pulse. The photodiode 80 in the circuit 50 of FIG. 3, or a separate photodiode (not shown), can be, as an example, sampled periodically over time, and discharged to ground after each sample. The photodiode 80 can then accumulate photons and be sampled at some regular sample timing period, so as to accumulate an amount of photons up to the next sample time (some photodiodes can be tuned to a given photon wavelength) so as to only respond to reflected light of the given wavelength. As can be seen from the curve of FIG. 5 the samples should be relatively un-variable over most of the period of the sampling, recognizing that some variability may be induced by noise in the system. Photodiodes can be any suitable photosensor or photodetector capable of sensing light or other electromagnetic energy.


However, as the sampling times approach the time of the occurrence of a pulse in the human subject, i.e., the decrease 122124 in the absorption, due to oxygenated blood in the finger, simply comparing successive samples can detect the change on the leading side of a respective blip 122, 24, a peak of the blip 122, 124 and the return of the trailing edge to the steady state value. Once again, looking at FIG. 5 and assuming that the blip from start to finish occupies about ⅛ of the time between pulses, and also that the pulse is about sixty, then the blip occupies about ⅛ of a second. for a swipe time of one half second, there would be about a 60% chance that the swipe would occur while at least some portion of the blip 122, 124 is occurring, i.e., leading edge, peak and trailing edge, and would, therefore, be detectable. Therefore, on average, if the first swipe fails to detect the blip 122, 124 indicative of the human pulse being present (even though not enough information is gathered to measure pulse rate and such, two swipes or at most three would most likely indicate the presence of a pulse cause by a human heart beat, and, therefore, another anti-spoof indicator.


The following is a disclosure by way of example of a computing device which may be used with the presently disclosed subject matter. The description of the various components of a computing device is not intended to represent any particular architecture or manner of interconnecting the components. Other systems that have fewer or more components may also be used with the disclosed subject matter. A communication device may constitute a form of a computing device and may at least emulate a computing device. The computing device may include an inter-connect (e.g., bus and system core logic), which can interconnect such components of a computing device to a data processing device, such as a processor(s) or microprocessor(s), or other form of partly or completely programmable or pre-programmed device, e.g., hard wired and/or application specific integrated circuit (“ASIC”) customized logic circuitry, such as a controller or microcontroller, a digital signal processor, or any other form of device that can fetch instructions, operate on pre-loaded/pre-programmed instructions, and/or follow instructions found in hard-wired or customized circuitry, to carry out logic operations that, together, perform steps of and whole processes and functionalities as described in the present disclosure.


In this description, various functions, functionalities and/or operations may be described as being performed by or caused by software program code to simplify description. However, those skilled in the art will recognize what is meant by such expressions is that the functions resulting from execution of the program code/instructions are performed by a computing device as described above, e.g., including a processor, such as a microprocessor, microcontroller, logic circuit or the like. Alternatively, or in combination, the functions and operations can be implemented using special purpose circuitry, with or without software instructions, such as using Application-Specific Integrated Circuit (ASIC) or Field-Programmable Gate Array (FPGA), which may be programmable, partly programmable or hard wired. The application specific integrated circuit (“ASIC”) logic may be such as gate arrays or standard cells, or the like, implementing customized logic by metalization(s) interconnects of the base gate array ASIC architecture or selecting and providing metalization(s) interconnects between standard cell functional blocks included in a manufacturers library of functional blocks, etc. Embodiments can thus be implemented using hardwired circuitry without program software code/instructions, or in combination with circuitry using programed software code/instructions.


Thus, the techniques are limited neither to any specific combination of hardware circuitry and software, nor to any particular tangible source for the instructions executed by the data processor(s) within the computing device. While some embodiments can be implemented in fully functioning computers and computer systems, various embodiments are capable of being distributed as a computing device including, e.g., a variety of forms and capable of being applied regardless of the particular type of machine or tangible computer-readable media used to actually effect the performance of the functions and operations and/or the distribution of the performance of the functions, functionalities and/or operations.


The interconnect may connect the data processing device to define logic circuitry including memory. The interconnect may be internal to the data processing device, such as coupling a microprocessor to on-board cache memory, or external (to the microprocessor) memory such as main memory, or a disk drive, or external to the computing device, such as a remote memory, a disc farm or other mass storage device(s), etc. Commercially available microprocessors, one or more of which could be a computing device or part of a computing device, include a PA-RISC series microprocessor from Hewlett-Packard Company, an 80×86 or Pentium series microprocessor from Intel Corporation, a PowerPC microprocessor from IBM, a Sparc microprocessor from Sun Microsystems, Inc, or a 68xxx series microprocessor from Motorola Corporation as examples.


The inter-connect in addition to interconnecting such as microprocessor(s) and memory may also interconnect such elements to a display controller and display device, and/or to other peripheral devices such as input/output (I/O) devices, e.g., through an input/output controller(s). Typical I/O devices can include a mouse, a keyboard(s), a modem(s), a network interface(s), printers, scanners, video cameras and other devices which are well known in the art. The inter-connect may include one or more buses connected to one another through various bridges, controllers and/or adapters. In one embodiment the I/O controller may include a USB (Universal Serial Bus) adapter for controlling USB peripherals, and/or an IEEE-1394 bus adapter for controlling IEEE-1394 peripherals.


The memory may include any tangible computer-readable media, which may include but are not limited to recordable and non-recordable type media such as volatile and non-volatile memory devices, such as volatile RAM (Random Access Memory), typically implemented as dynamic RAM (DRAM) which requires power continually in order to refresh or maintain the data in the memory, and non-volatile ROM (Read Only Memory), and other types of non-volatile memory, such as a hard drive, flash memory, detachable memory stick, etc. Non-volatile memory typically may include a magnetic hard drive, a magnetic optical drive, or an optical drive (e.g., a DVD RAM, a CD ROM, a DVD or a CD), or other type of memory system which maintains data even after power is removed from the system.


A server could be made up of one or more computing devices. Servers can be utilized, e.g., in a network to host a network database, compute necessary variables and information from information in the database(s), store and recover information from the database(s), track information and variables, provide interfaces for uploading and downloading information and variables, and/or sort or otherwise manipulate information and data from the database(s). In one embodiment a server can be used in conjunction with other computing devices positioned locally or remotely to perform certain calculations and other functions as may be mentioned in the present application.


At least some aspects of the disclosed subject matter can be embodied, at least in part, utilizing programmed software code/instructions. That is, the functions, functionalities and/or operations techniques may be carried out in a computing device or other data processing system in response to its processor, such as a microprocessor, executing sequences of instructions contained in a memory, such as ROM, volatile RAM, non-volatile memory, cache or a remote storage device. In general, the routines executed to implement the embodiments of the disclosed subject matter may be implemented as part of an operating system or a specific application, component, program, object, module or sequence of instructions usually referred to as “computer programs,” or “software.” The computer programs typically comprise instructions stored at various times in various tangible memory and storage devices in a computing device, such as in cache memory, main memory, internal or external disk drives, and other remote storage devices, such as a disc farm, and when read and executed by a processor(s) in the computing device, cause the computing device to perform a method(s), e.g., process and operation steps to execute an element(s) as part of some aspect(s) of the method(s) of the disclosed subject matter.


A tangible machine readable medium can be used to store software and data that, when executed by a computing device, causes the computing device to perform a method(s) as may be recited in one or more accompanying claims defining the disclosed subject matter. The tangible machine readable medium may include storage of the executable software program code/instructions and data in various tangible locations, including for example ROM, volatile RAM, non-volatile memory and/or cache. Portions of this program software code/instructions and/or data may be stored in any one of these storage devices. Further, the program software code/instructions can be obtained from remote storage, including, e.g., through centralized servers or peer to peer networks and the like. Different portions of the software program code/instructions and data can be obtained at different times and in different communication sessions or in a same communication session.


The software program code/instructions and data can be obtained in their entirety prior to the execution of a respective software application by the computing device. Alternatively, portions of the software program code/instructions and data can be obtained dynamically, e.g., just in time, when needed for execution. Alternatively, some combination of these ways of obtaining the software program code/instructions and data may occur, e.g., for different applications, components, programs, objects, modules, routines or other sequences of instructions or organization of sequences of instructions, by way of example. Thus, it is not required that the data and instructions be on a single machine readable medium in entirety at any particular instant of time.


In general, a tangible machine readable medium includes any tangible mechanism that provides (i.e., stores) information in a form accessible by a machine (i.e., a computing device), which may be included, e.g., in a communication device, a network device, a personal digital assistant, a mobile communication device, whether or not able to download and run applications from the communication network, such as the Internet, e.g., an I-phone, Blackberry, Droid or the like, a manufacturing tool, or any other device including a computing device, comprising one or more data processors, etc.


In one embodiment, a user terminal can be a computing device, such as in the form of or included within a PDA, a cellular phone, a notebook computer, a personal desktop computer, etc. Alternatively, the traditional communication client(s) may be used in some embodiments of the disclosed subject matter.


While some embodiments of the disclosed subject matter have been described in the context of fully functioning computing devices and computing systems, those skilled in the art will appreciate that various embodiments of the disclosed subject matter are capable of being distributed, e.g., as a program product in a variety of forms and are capable of being applied regardless of the particular type of computing device machine or computer-readable media used to actually effect the distribution.


The disclosed subject matter may be described with reference to block diagrams and operational illustrations of methods and devices to provide a system and methods according to the disclosed subject matter. It will be understood that each block of a block diagram or other operational illustration (herein collectively, “block diagram”), and combination of blocks in a block diagram, can be implemented by means of analog or digital hardware and computer program instructions. These computing device software program code/instructions can be provided to the computing device such that the instructions, when executed by the computing device, e.g., on a processor within the computing device or other data processing apparatus, the program software code/instructions cause the computing device to perform functions, functionalities and operations of a method(s) according to the disclosed subject matter, as recited in the accompanying claims, with such functions, functionalities and operations specified in the block diagram.


It will be understood that in some possible alternate implementations, the function, functionalities and operations noted in the blocks of a block diagram may occur out of the order noted in the block diagram. For example, the function noted in two blocks shown in succession can in fact be executed substantially concurrently or the functions noted in blocks can sometimes be executed in the reverse order, depending upon the function, functionalities and operations involved. Therefore, the embodiments of methods presented and described as a flowchart(s) in the form of a block diagram in the present application are provided by way of example in order to provide a more complete understanding of the disclosed subject matter. The disclosed flow and concomitantly the method(s) performed as recited in the accompanying claims are not limited to the functions, functionalities and operations illustrated in the block diagram and/or logical flow presented herein. Alternative embodiments are contemplated in which the order of the various functions, functionalities and operations may be altered and in which sub-operations described as being part of a larger operation may be performed independently or performed differently than illustrated or not performed at all.


Although some of the drawings may illustrate a number of operations in a particular order, functions, functionalities and/or operations which are not now known to be order dependent, or become understood to not be order dependent, may be reordered and other operations may be combined or broken out. While some reordering or other groupings may have been specifically mentioned in the present application, others will be or may become apparent to those of ordinary skill in the art and so the disclosed subject matter does not present an exhaustive list of alternatives. It should also be recognized that the aspects of the disclosed subject matter may be implemented in parallel or seriatim in hardware, firmware, software or any combination(s) thereof co-located or remotely located, at least in part, from each other, e.g., in arrays or networks of computing devices, over interconnected networks, including the Internet, and the like.


The disclosed subject matter is described in the present application with reference to one or more specific exemplary embodiments thereof. It will be evident that various modifications may be made to the disclosed subject matter without departing from the broader spirit and scope of the disclosed subject matter as set forth in the appended claims. The specification and drawings are, accordingly, to be regarded in an illustrative sense for explanation of aspects of the disclosed subject matter rather than a restrictive or limiting sense. It should be understood that various alternatives to the embodiments of the invention described herein may be employed in practicing the invention. It is intended that the following claims define the scope of the invention and that methods and structures within the scope of these claims and their equivalents be covered thereby.

Claims
  • 1. A biometric image sensor comprising: a first light source having a first wavelength λ1;a second light source having a second wavelength λ2;a photodetector configured and positioned to receive light of the first wavelength λ1 and light of the second wavelength λ2 reflecting from a biometric object being imaged and to produce a first reflectivity output indicative of the amount of reflectance of the light of the first wavelength λ1 and a second reflectivity output indicative of the amount of reflectance of the light of the second wavelength λ2; anda processor configured to compare the difference between the first reflectivity output and the second reflectivity output to detect one or more of a leading edge, peak or trailing edge of a change in absorption of oxygen in blood of the biometric object.
  • 2. The biometric image sensor of claim 1 wherein the biometric object is a finger.
  • 3. The biometric image sensor of claim 2 wherein the finger is being imaged to obtain a fingerprint image.
  • 4. The biometric image sensor of claim 3 wherein the processor is configured to compare successive samples to detect the change due to the presence of oxygenated blood in the finger.
  • 5. The biometric image sensor of claim 1 wherein the first wavelength λ1 and the second wavelength λ2 are selected to coincide with a section of relatively high change in a reflectance vs. wavelength curve for the biometric object.
  • 6. The biometric image sensor of claim 1 wherein the biometric object comprises human skin.
  • 7. The biometric image sensor of claim 1 wherein the first wavelength λ1 is in the visible range of yellow light and the second wavelength λ2 is in the visible range of one of orange light and red light.
  • 8. The biometric image sensor of claim 1, wherein the biometric object is a finger being imaged to obtain a fingerprint image, and wherein the processor is configured to measure a pulse rate from a periodic presence of oxygenated blood detected while the finger is kept on the biometric image sensor.
  • 9. A method of sensing an image of a biometric object comprising: directing a first light source having a first wavelength λ1 on the biometric object;directing a second light source having a second wavelength λ2 on the biometric object;detecting the light of the first wavelength λ1 and the light of the second wavelength λ2 reflecting from the biometric object being imaged and producing a first reflectivity output indicative of the amount of reflectance of the light of the first wavelength λ1 and a second reflectivity output indicative of the amount of reflectance of the light of the second wavelength λ2, andcomparing, via a processor, the difference between the first reflectivity output and the second reflectivity output to detect one or more of a leading edge, peak or trailing edge of a change in absorption of oxygen in blood of the biometric object.
  • 10. The method of claim 9 wherein the biometric object is a finger.
  • 11. The method of claim 10 wherein the finger is being imaged to obtain a fingerprint image.
  • 12. The method of claim 11 wherein successive samples are compared to detect the change due to a periodic presence of oxygenated blood in the finger.
  • 13. The method of claim 9 wherein the first wavelength λ1 and the second wavelength λ2 are selected to coincide with a section of relatively high change in a reflectance vs. wavelength curve for the biometric object.
  • 14. The method of claim 9 wherein the biometric object comprises human skin.
  • 15. The method of claim 9 wherein the first wavelength λ1 is in the visible range of yellow light and the second wavelength λ2 is in the visible range of one of orange light and red light.
  • 16. The method of claim 9, wherein the biometric object is a finger being imaged by a fingerprint sensor to obtain a fingerprint image, and wherein a pulse rate is measured from a periodic presence of oxygenated blood detected while the finger is kept on the fingerprint sensor.
  • 17. A biometric image sensor comprising: a light source configurable to emit a first wavelength λ1 of light and a second wavelength λ2 of light;a photodetector configurable and positionable to receive one or more electromagnetic radiation wavelengths reflected from a surface in response to the first wavelength λ1 and the second wavelength λ2 wherein the reflected electromagnetic radiation is reflecting from a biometric object being imaged to produce a first reflectivity output indicative of the amount of reflectance of the first wavelength λ1 of light and a second reflectivity output indicative of the amount of reflectance of the second wavelength λ2 of light; anda processor configured to compare the difference between the first reflectivity output and the second reflectivity output to detect one or more of a leading edge, peak or trailing edge of a change in absorption of oxygen in blood of the biometric object.
  • 18. The biometric image sensor of claim 17, wherein the biometric object is a finger being imaged to obtain a fingerprint image, wherein the processor is configured to compare successive samples to detect the change due to the presence of oxygenated blood in the finger.
  • 19. The biometric image sensor of claim 17, wherein the biometric object is a finger being imaged to obtain a fingerprint image, and wherein the processor is configured to measure a pulse rate from a periodic presence of oxygenated blood detected while the finger is kept on the biometric image sensor.
  • 20. The biometric image sensor of claim 17, wherein the first wavelength λ1 of light and the second wavelength λ2 of light are selected to coincide with a section of relatively high change in a reflectance vs. wavelength curve for the biometric object.
  • 21. The biometric image sensor of claim 1, wherein the biometric object is a finger being imaged to obtain a fingerprint image, and wherein detection of one of the leading edge, the peak or the trailing edge of the change in the absorption of oxygen indicates that the finger is from a living body part.
  • 22. The method of claim 9, wherein the biometric object is a finger being imaged to obtain a fingerprint image, and wherein detection of one of the leading edge, the peak or the trailing edge of the change in the absorption of oxygen indicates that the finger is from a living body part.
  • 23. The biometric image sensor of claim 17, wherein the biometric object is a finger being imaged to obtain a fingerprint image, and wherein detection of one of the leading edge, the peak or the trailing edge of the change in the absorption of oxygen indicates that the finger is from a living body part.
  • 24. The biometric image sensor of claim 17, wherein the leading edge of the change in the absorption of oxygen in the blood of the biometric object is detected.
  • 25. The biometric image sensor of claim 17, wherein the peak of the change in the absorption of oxygen in the blood of the biometric object is detected.
  • 26. The biometric image sensor of claim 17, wherein the trailing edge of the change in the absorption of oxygen in the blood of the biometric object is detected.
  • 27. The biometric object of claim 17, wherein the processor is configured to compare the difference between the first reflectivity output and the second reflectivity output over no more than three samples.
CROSS-REFERENCE

This application is a continuation of U.S. patent application Ser. No. 13/376,377, filed Feb. 8, 2013, issued Sep. 15, 2015, as U.S. Pat. No. 9,137,438, and this application claims the benefit of U.S. Provisional Application No. 61/616,112, filed Mar. 27, 2012. Each above identified patent and patent application are hereby incorporated herein by reference in their entirety.

US Referenced Citations (451)
Number Name Date Kind
4151512 Rigannati et al. Apr 1979 A
4225850 Chang et al. Sep 1980 A
4310827 Asi Jan 1982 A
4353056 Tsikos Oct 1982 A
4405829 Rivest et al. Sep 1983 A
4525859 Bowles et al. Jun 1985 A
4550221 Mabusth Oct 1985 A
4580790 Doose Apr 1986 A
4582985 Loftberg Apr 1986 A
4675544 Shrenk Jun 1987 A
4758622 Gosselin Jul 1988 A
4817183 Sparrow Mar 1989 A
5076566 Kriegel Dec 1991 A
5088817 Igaki et al. Feb 1992 A
5109427 Yang Apr 1992 A
5140642 Hau et al. Aug 1992 A
5305017 Gerpheide Apr 1994 A
5319323 Fang Jun 1994 A
5325442 Knapp Jun 1994 A
5359243 Norman Oct 1994 A
5420936 Fitzpatrick et al. May 1995 A
5422807 Mitra et al. Jun 1995 A
5429006 Tamori Jul 1995 A
5456256 Schneider et al. Oct 1995 A
5543591 Gillespie et al. Aug 1996 A
5569901 Bridgelall et al. Oct 1996 A
5623552 Lane Apr 1997 A
5627316 De Winter et al. May 1997 A
5650842 Maase et al. Jul 1997 A
5717777 Wong et al. Feb 1998 A
5781651 Hsiao et al. Jul 1998 A
5801681 Sayag Sep 1998 A
5818956 Tuli Oct 1998 A
5838306 O'Connor Nov 1998 A
5848176 Hara et al. Dec 1998 A
5850450 Schweitzer et al. Dec 1998 A
5852670 Setlak et al. Dec 1998 A
5864296 Upton Jan 1999 A
5887343 Salatino et al. Mar 1999 A
5892824 Beatson et al. Apr 1999 A
5903225 Schmitt et al. May 1999 A
5915757 Tsuyama et al. Jun 1999 A
5920384 Borza Jul 1999 A
5920640 Salatino et al. Jul 1999 A
5940526 Setlak et al. Aug 1999 A
5963679 Setlak Oct 1999 A
5995630 Borza Nov 1999 A
5999637 Toyoda et al. Dec 1999 A
6002815 Immega et al. Dec 1999 A
6011859 Kalnitsky et al. Jan 2000 A
6016355 Dickinson et al. Jan 2000 A
6052475 Upton Apr 2000 A
6067368 Setlak et al. May 2000 A
6073343 Petrick et al. Jun 2000 A
6076566 Lowe Jun 2000 A
6088585 Schmitt et al. Jul 2000 A
6098175 Lee Aug 2000 A
6118318 Fifield et al. Sep 2000 A
6134340 Hsu et al. Oct 2000 A
6157722 Lerner et al. Dec 2000 A
6161213 Lofstrom Dec 2000 A
6175407 Santor Jan 2001 B1
6182076 Yu et al. Jan 2001 B1
6182892 Angelo et al. Feb 2001 B1
6185318 Jain et al. Feb 2001 B1
6234031 Suga May 2001 B1
6241288 Bergenek et al. Jun 2001 B1
6259108 Antonelli et al. Jul 2001 B1
6289114 Mainguet Sep 2001 B1
6292272 Okauchi et al. Sep 2001 B1
6317508 Kramer et al. Nov 2001 B1
6320394 Tartagni Nov 2001 B1
6325285 Baratelli Dec 2001 B1
6327376 Harkin Dec 2001 B1
6330345 Russo et al. Dec 2001 B1
6332193 Glass et al. Dec 2001 B1
6333989 Borza Dec 2001 B1
6337919 Duton Jan 2002 B1
6343162 Saito et al. Jan 2002 B1
6346739 Lepert et al. Feb 2002 B1
6347040 Fries et al. Feb 2002 B1
6357663 Takahashi et al. Mar 2002 B1
6360004 Akizuki Mar 2002 B1
6362633 Tartagni Mar 2002 B1
6392636 Ferrari et al. May 2002 B1
6399994 Shobu Jun 2002 B2
6400836 Senior Jun 2002 B2
6408087 Kramer Jun 2002 B1
6459804 Mainguet Oct 2002 B2
6473072 Comiskey et al. Oct 2002 B1
6509501 Eicken et al. Jan 2003 B2
6525547 Hayes Feb 2003 B2
6525932 Ohnishi et al. Feb 2003 B1
6539101 Black Mar 2003 B1
6580816 Kramer et al. Jun 2003 B2
6597289 Sabatini Jul 2003 B2
6628812 Setlak et al. Sep 2003 B1
6631201 Dickinson et al. Oct 2003 B1
6643389 Raynal et al. Nov 2003 B1
6672174 Deconde et al. Jan 2004 B2
6710461 Chou et al. Mar 2004 B2
6738050 Comiskey et al. May 2004 B2
6741729 Bjorn et al. May 2004 B2
6757002 Oross et al. Jun 2004 B1
6766040 Catalano et al. Jul 2004 B1
6785407 Tschudi et al. Aug 2004 B1
6799275 Bjorn et al. Sep 2004 B1
6836230 Le Pailleur et al. Dec 2004 B2
6838905 Doyle Jan 2005 B1
6873356 Kanbe et al. Mar 2005 B1
6886104 McClurg et al. Apr 2005 B1
6897002 Teraoka et al. May 2005 B2
6898299 Brooks May 2005 B1
6924496 Manansala Aug 2005 B2
6937748 Schneider et al. Aug 2005 B1
6941001 Bolle et al. Sep 2005 B1
6941810 Okada Sep 2005 B2
6950540 Higuchi Sep 2005 B2
6959874 Bardwell Nov 2005 B2
6963626 Shaeffer et al. Nov 2005 B1
6970584 O'Gorman et al. Nov 2005 B2
6980672 Saito et al. Dec 2005 B2
6983882 Cassone Jan 2006 B2
7013030 Wong et al. Mar 2006 B2
7020591 Wei et al. Mar 2006 B1
7030860 Hsu et al. Apr 2006 B1
7031670 May Apr 2006 B2
7035443 Wong Apr 2006 B2
7042535 Katoh et al. May 2006 B2
7043061 Hamid et al. May 2006 B2
7043644 DeBruine May 2006 B2
7046230 Zadesky et al. May 2006 B2
7064743 Nishikawa Jun 2006 B2
7099496 Benkley Aug 2006 B2
7110574 Haruki et al. Sep 2006 B2
7110577 Tschud Sep 2006 B1
7113622 Hamid Sep 2006 B2
7126389 McRae et al. Oct 2006 B1
7129926 Mathiassen et al. Oct 2006 B2
7136514 Wong Nov 2006 B1
7146024 Benkley Dec 2006 B2
7146026 Russon et al. Dec 2006 B2
7146029 Manansala Dec 2006 B2
7184581 Johansen et al. Feb 2007 B2
7190209 Kang et al. Mar 2007 B2
7190816 Mitsuyu et al. Mar 2007 B2
7194392 Tuken et al. Mar 2007 B2
7197168 Russo Mar 2007 B2
7200250 Chou Apr 2007 B2
7251351 Mathiassen et al. Jul 2007 B2
7258279 Schneider et al. Aug 2007 B2
7260246 Fujii Aug 2007 B2
7263212 Kawabe Aug 2007 B2
7263213 Rowe Aug 2007 B2
7289649 Walley et al. Oct 2007 B1
7290323 Deconde et al. Nov 2007 B2
7308121 Mathiassen et al. Dec 2007 B2
7308122 McClurg et al. Dec 2007 B2
7321672 Sasaki et al. Jan 2008 B2
7356169 Hamid Apr 2008 B2
7360688 Harris Apr 2008 B1
7369685 Deleon May 2008 B2
7379569 Chikazawa et al. May 2008 B2
7408135 Fujeda Aug 2008 B2
7409876 Ganapathi et al. Aug 2008 B2
7412083 Takahashi Aug 2008 B2
7424618 Roy et al. Sep 2008 B2
7447339 Mimura et al. Nov 2008 B2
7447911 Chou et al. Nov 2008 B2
7460697 Erhart et al. Dec 2008 B2
7463756 Benkley Dec 2008 B2
7474772 Russo et al. Jan 2009 B2
7505611 Fyke Mar 2009 B2
7505613 Russo Mar 2009 B2
7565548 Fiske et al. Jul 2009 B2
7574022 Russo Aug 2009 B2
7596832 Hsieh et al. Oct 2009 B2
7599530 Boshra Oct 2009 B2
7616787 Boshra Nov 2009 B2
7643950 Getzin et al. Jan 2010 B1
7646897 Fyke Jan 2010 B2
7681232 Nordentoft et al. Mar 2010 B2
7689013 Shinzaki Mar 2010 B2
7706581 Drews et al. Apr 2010 B2
7733697 Picca et al. Jun 2010 B2
7751601 Benkley Jul 2010 B2
7826645 Cayen Nov 2010 B1
7843438 Onoda Nov 2010 B2
7848798 Martinsen et al. Dec 2010 B2
7899216 Watanabe et al. Mar 2011 B2
7953258 Dean et al. May 2011 B2
8005276 Dean et al. Aug 2011 B2
8031046 Franza et al. Oct 2011 B2
8031916 Abiko et al. Oct 2011 B2
8063734 Conforti Nov 2011 B2
8077935 Geoffroy et al. Dec 2011 B2
8107212 Nelson et al. Jan 2012 B2
8116540 Dean et al. Feb 2012 B2
8131026 Benkley et al. Mar 2012 B2
8165355 Benkley et al. Apr 2012 B2
8175345 Gardner May 2012 B2
8181031 Narayanaswami May 2012 B2
8204281 Satya et al. Jun 2012 B2
8224044 Benkley Jul 2012 B2
8229184 Benkley Jul 2012 B2
8276816 Gardner Oct 2012 B2
8278946 Thompson Oct 2012 B2
8290150 Erhart et al. Oct 2012 B2
8315444 Gardner Nov 2012 B2
8331096 Garcia Dec 2012 B2
8358815 Benkley et al. Jan 2013 B2
8374407 Benkley et al. Feb 2013 B2
8391568 Satyan Mar 2013 B2
8616451 Rao Dec 2013 B1
20010026636 Mainget Oct 2001 A1
20010030644 Allport Oct 2001 A1
20010036299 Senior Nov 2001 A1
20010043728 Kramer et al. Nov 2001 A1
20020025062 Black Feb 2002 A1
20020061125 Fujii May 2002 A1
20020064892 Lepert et al. May 2002 A1
20020067845 Griffis Jun 2002 A1
20020073046 David Jun 2002 A1
20020089044 Simmons et al. Jul 2002 A1
20020089410 Janiak et al. Jul 2002 A1
20020096731 Wu et al. Jul 2002 A1
20020122026 Bergstrom Sep 2002 A1
20020126516 Jean Sep 2002 A1
20020133725 Roy et al. Sep 2002 A1
20020152048 Hayes Oct 2002 A1
20020181749 Matsumoto et al. Dec 2002 A1
20030002717 Hamid Jan 2003 A1
20030002719 Hamid et al. Jan 2003 A1
20030021495 Cheng Jan 2003 A1
20030035570 Benkley Feb 2003 A1
20030044051 Fujieda Mar 2003 A1
20030063782 Acharya et al. Apr 2003 A1
20030068072 Hamid Apr 2003 A1
20030076301 Tsuk et al. Apr 2003 A1
20030076303 Huppi Apr 2003 A1
20030095096 Robbin et al. May 2003 A1
20030095690 Su et al. May 2003 A1
20030102874 Lane et al. Jun 2003 A1
20030123714 O'Gorman et al. Jul 2003 A1
20030123715 Uchida Jul 2003 A1
20030141959 Keogh et al. Jul 2003 A1
20030147015 Katoh et al. Aug 2003 A1
20030161510 Fuji Aug 2003 A1
20030161512 Mathiassen Aug 2003 A1
20030169228 Mathiassen et al. Sep 2003 A1
20030174871 Yoshioka et al. Sep 2003 A1
20030186157 Teraoka et al. Oct 2003 A1
20030209293 Sako et al. Nov 2003 A1
20030224553 Manansala Dec 2003 A1
20040012773 Puttkammer Jan 2004 A1
20040017934 Kocher et al. Jan 2004 A1
20040022001 Chu et al. Feb 2004 A1
20040042642 Bolle et al. Mar 2004 A1
20040050930 Rowe Mar 2004 A1
20040066613 Leitao Apr 2004 A1
20040076313 Bronstein et al. Apr 2004 A1
20040081339 Benkley Apr 2004 A1
20040096086 Miyasaka May 2004 A1
20040113956 Bellwood et al. Jun 2004 A1
20040120400 Linzer Jun 2004 A1
20040125993 Zhao et al. Jul 2004 A1
20040129787 Saito Jul 2004 A1
20040136612 Meister et al. Jul 2004 A1
20040155752 Radke Aug 2004 A1
20040172339 Snelgrove et al. Sep 2004 A1
20040179718 Chou Sep 2004 A1
20040184641 Nagasaka et al. Sep 2004 A1
20040188838 Okada et al. Sep 2004 A1
20040190761 Lee Sep 2004 A1
20040208346 Baharav et al. Oct 2004 A1
20040208347 Baharav et al. Oct 2004 A1
20040208348 Baharav et al. Oct 2004 A1
20040213441 Tschudi Oct 2004 A1
20040215689 Dooley et al. Oct 2004 A1
20040228505 Sugimoto Nov 2004 A1
20040228508 Shigeta Nov 2004 A1
20040240712 Rowe et al. Dec 2004 A1
20040252867 Lan et al. Dec 2004 A1
20050001711 Doughty et al. Jan 2005 A1
20050031174 Ryhanen et al. Feb 2005 A1
20050036665 Higuchi Feb 2005 A1
20050047485 Khayrallah et al. Mar 2005 A1
20050100196 Scott et al. May 2005 A1
20050100938 Hofmann et al. May 2005 A1
20050109835 Jacoby et al. May 2005 A1
20050110103 Setlak May 2005 A1
20050111708 Chou May 2005 A1
20050123176 Ishii et al. Jun 2005 A1
20050129291 Boshra Jun 2005 A1
20050136200 Durell et al. Jun 2005 A1
20050139656 Arnouse Jun 2005 A1
20050139685 Kozlay Jun 2005 A1
20050162402 Watanachote Jul 2005 A1
20050169503 Howell et al. Aug 2005 A1
20050174015 Scott et al. Aug 2005 A1
20050210271 Chou et al. Sep 2005 A1
20050219200 Weng Oct 2005 A1
20050220329 Payne et al. Oct 2005 A1
20050231213 Chou et al. Oct 2005 A1
20050238212 Du et al. Oct 2005 A1
20050244038 Benkley Nov 2005 A1
20050244039 Geoffroy et al. Nov 2005 A1
20050247559 Frey et al. Nov 2005 A1
20050249386 Juh Nov 2005 A1
20050258952 Utter et al. Nov 2005 A1
20050269402 Spitzer et al. Dec 2005 A1
20060006224 Modi Jan 2006 A1
20060055500 Burke et al. Mar 2006 A1
20060066572 Yumoto et al. Mar 2006 A1
20060078176 Abiko et al. Apr 2006 A1
20060083411 Benkley Apr 2006 A1
20060110537 Huang et al. May 2006 A1
20060140461 Kim et al. Jun 2006 A1
20060144953 Takao Jul 2006 A1
20060170528 Funushige et al. Aug 2006 A1
20060181521 Perrault et al. Aug 2006 A1
20060182319 Setlank et al. Aug 2006 A1
20060187200 Martin Aug 2006 A1
20060210082 Devadas et al. Sep 2006 A1
20060214512 Iwata Sep 2006 A1
20060214767 Carrieri Sep 2006 A1
20060239514 Watanabe et al. Oct 2006 A1
20060249008 Luther Nov 2006 A1
20060259873 Mister Nov 2006 A1
20060261174 Zellner et al. Nov 2006 A1
20060267125 Huang et al. Nov 2006 A1
20060267385 Steenwyk et al. Nov 2006 A1
20060271793 Devadas et al. Nov 2006 A1
20060285728 Leung et al. Dec 2006 A1
20060287963 Steeves et al. Dec 2006 A1
20070031011 Erhart et al. Feb 2007 A1
20070036400 Watanabe et al. Feb 2007 A1
20070057763 Blattner et al. Mar 2007 A1
20070058843 Theis et al. Mar 2007 A1
20070067828 Bychkov Mar 2007 A1
20070076926 Schneider et al. Apr 2007 A1
20070076951 Tanaka et al. Apr 2007 A1
20070086634 Setlak et al. Apr 2007 A1
20070090312 Stallinga et al. Apr 2007 A1
20070138299 Mitra Jun 2007 A1
20070154072 Taraba et al. Jul 2007 A1
20070160269 Kuo Jul 2007 A1
20070180261 Akkermans et al. Aug 2007 A1
20070196002 Choi et al. Aug 2007 A1
20070198141 Moore Aug 2007 A1
20070198435 Siegal et al. Aug 2007 A1
20070228154 Tran Oct 2007 A1
20070237366 Maletsky Oct 2007 A1
20070237368 Bjorn et al. Oct 2007 A1
20070248249 Stoianov Oct 2007 A1
20070290124 Neil et al. Dec 2007 A1
20080002867 Mathiassen et al. Jan 2008 A1
20080013805 Sengupta et al. Jan 2008 A1
20080019578 Saito et al. Jan 2008 A1
20080049987 Champagne et al. Feb 2008 A1
20080049989 Iseri et al. Feb 2008 A1
20080063245 Benkley et al. Mar 2008 A1
20080069412 Champagne et al. Mar 2008 A1
20080089605 Haven Apr 2008 A1
20080126260 Cox et al. May 2008 A1
20080169345 Keane et al. Jul 2008 A1
20080170695 Adler et al. Jul 2008 A1
20080175450 Scott et al. Jul 2008 A1
20080178008 Takahashi et al. Jul 2008 A1
20080179112 Qin et al. Jul 2008 A1
20080185429 Saville Aug 2008 A1
20080201265 Hewton Aug 2008 A1
20080205714 Benkley et al. Aug 2008 A1
20080219521 Benkley et al. Sep 2008 A1
20080222049 Loomis et al. Sep 2008 A1
20080223925 Saito et al. Sep 2008 A1
20080226132 Gardner Sep 2008 A1
20080240523 Benkley et al. Oct 2008 A1
20080240537 Yang et al. Oct 2008 A1
20080244277 Orsini et al. Oct 2008 A1
20080247607 Amano Oct 2008 A1
20080267462 Nelson et al. Oct 2008 A1
20080279373 Erhart et al. Nov 2008 A1
20080317290 Tazoe Dec 2008 A1
20090001999 Douglas Jan 2009 A1
20090130369 Huang et al. May 2009 A1
20090153297 Gardner Jun 2009 A1
20090154779 Satyan et al. Jun 2009 A1
20090155456 Benkley et al. Jun 2009 A1
20090169071 Bond et al. Jul 2009 A1
20090174974 Huang et al. Jul 2009 A1
20090212902 Haddock Aug 2009 A1
20090218698 Lam Sep 2009 A1
20090237135 Ramaraju et al. Sep 2009 A1
20090252384 Dean et al. Oct 2009 A1
20090252385 Dean et al. Oct 2009 A1
20090252386 Dean et al. Oct 2009 A1
20090279742 Abiko Nov 2009 A1
20090319435 Little et al. Dec 2009 A1
20090324028 Russo Dec 2009 A1
20100026451 Erhart et al. Feb 2010 A1
20100034433 Thiebot Feb 2010 A1
20100045705 Vertegaal et al. Feb 2010 A1
20100075631 Black et al. Mar 2010 A1
20100083000 Kesanupalli et al. Apr 2010 A1
20100110170 Bringer May 2010 A1
20100117794 Adams et al. May 2010 A1
20100119124 Satyan May 2010 A1
20100123675 Ippel May 2010 A1
20100127366 Bond et al. May 2010 A1
20100176823 Thompson et al. Jul 2010 A1
20100176892 Thompson et al. Jul 2010 A1
20100177940 Thompson et al. Jul 2010 A1
20100180136 Thompson et al. Jul 2010 A1
20100185068 Park Jul 2010 A1
20100189314 Benkley et al. Jul 2010 A1
20100208953 Gardner et al. Aug 2010 A1
20100244166 Shibuta et al. Sep 2010 A1
20100272329 Benkley Oct 2010 A1
20100284565 Benkley et al. Nov 2010 A1
20110002461 Erhart et al. Jan 2011 A1
20110018556 Le et al. Jan 2011 A1
20110082791 Baghdasaryan et al. Apr 2011 A1
20110083018 Kesanupalli et al. Apr 2011 A1
20110083170 Kesanupalli et al. Apr 2011 A1
20110083173 Baghdasaryan et al. Apr 2011 A1
20110090047 Patel Apr 2011 A1
20110102137 Schroter May 2011 A1
20110102567 Erhart May 2011 A1
20110102569 Erhart May 2011 A1
20110134251 Kim et al. Jun 2011 A1
20110175703 Benkley Jul 2011 A1
20110176037 Benkley Jul 2011 A1
20110182486 Valfridsson et al. Jul 2011 A1
20110214924 Perezselsky et al. Sep 2011 A1
20110267298 Erhart et al. Nov 2011 A1
20110298711 Dean et al. Dec 2011 A1
20110304001 Erhart et al. Dec 2011 A1
20120044639 Garcia Feb 2012 A1
20120062364 Rowe Mar 2012 A1
20120189166 Russo Jul 2012 A1
20120189172 Russo Jul 2012 A1
20120206586 Gardner Aug 2012 A1
20120230551 Hama Sep 2012 A1
20120256280 Erhart Oct 2012 A1
20120257032 Benkley Oct 2012 A1
20120308092 Benkley et al. Dec 2012 A1
20130021044 Thompson et al. Jan 2013 A1
20130094715 Benkley et al. Apr 2013 A1
20130279768 Boshra Oct 2013 A1
20140026208 Coons Jan 2014 A1
Foreign Referenced Citations (69)
Number Date Country
2213813 Oct 1973 DE
0929028 Jan 1998 EP
0905646 Mar 1999 EP
0973123 Jan 2000 EP
1018697 Jul 2000 EP
1139301 Oct 2001 EP
1531419 May 2005 EP
1533759 May 2005 EP
1538548 Jun 2005 EP
162439981 Feb 2006 EP
1775674 Apr 2007 EP
1939788 Jul 2008 EP
2331613 May 1999 GB
2480919 Dec 2011 GB
2487661 Aug 2012 GB
2489100 Sep 2012 GB
2490192 Oct 2012 GB
24749998 Feb 2013 GB
01094418 Apr 1989 JP
04158434 Jun 1992 JP
2005011002 Jan 2005 JP
2005242856 Sep 2005 JP
2007305097 Nov 2007 JP
200606745 Feb 2006 TW
200606746 Feb 2006 TW
200614092 May 2006 TW
200617798 Jun 2006 TW
200620140 Jun 2006 TW
200629167 Aug 2006 TW
WO 9003620 Apr 1990 WO
WO 9858342 Dec 1998 WO
WO 9928701 Jun 1999 WO
WO 9943258 Sep 1999 WO
WO 0122349 Mar 2001 WO
WO 0194902 Dec 2001 WO
WO 0194902 Dec 2001 WO
WO 0195304 Dec 2001 WO
WO 0211066 Feb 2002 WO
WO 0247018 Jun 2002 WO
WO 0247018 Jun 2002 WO
WO 02061668 Aug 2002 WO
WO 02077907 Oct 2002 WO
WO 03063054 Jul 2003 WO
WO 03075210 Sep 2003 WO
WO 2004066194 Aug 2004 WO
WO 2004066693 Aug 2004 WO
WO 2005104012 Nov 2005 WO
WO 2005106774 Nov 2005 WO
WO 2005106774 Nov 2005 WO
WO 2006040724 Apr 2006 WO
WO 2006041780 Apr 2006 WO
WO 2007011607 Jan 2007 WO
WO 2008033264 Mar 2008 WO
WO 2008033264 Mar 2008 WO
WO 2008033265 Jun 2008 WO
WO 2008033265 Jun 2008 WO
WO 2008137287 Nov 2008 WO
WO 2009002599 Dec 2008 WO
WO 2009002599 Dec 2008 WO
WO 2009029257 Jun 2009 WO
WO 2009079219 Jun 2009 WO
WO 2009079221 Jun 2009 WO
WO 2009079262 Jun 2009 WO
WO 2010034036 Mar 2010 WO
WO 2010036445 Apr 2010 WO
WO 2010143597 Dec 2010 WO
WO 2011088248 Jan 2011 WO
WO 2011088252 Jan 2011 WO
WO 2011053797 May 2011 WO
Non-Patent Literature Citations (20)
Entry
Matsumoto et al., Impact of Artificial “Gummy” Fingers on Fingerprint Systems, SPIE 4677 (2002), reprinted from cryptome.org.
Maltoni, “Handbook of Fingerprint Recognition”, XP002355942 Springer, New York, USA, Jun. 2003 (Jun. 2003) pp. 65-69.
Vermasan, et al., “A500 dpi AC Capacitive Hybrid Flip-Chip CMOS ASIC/Sensor Module for Fingerprint, Navigation, and Pointer Detection With On-Chip Data Processing”, IEEE Journal of Solid State Circuits, vol. 38, No. 12, Dec. 2003, pp. 2288-2294.
Ratha, et al. “Adaptive Flow Orientation Based Feature Extraction in Fingerprint Images,” Pattern Recognition, vol. 28 No. 11, 1657-1672, Nov. 1995.
Ratha, et al., “A Real Time Matching System for Large Fingerprint Databases,” IEEE, Aug. 1996.
Suh, et al., “Design and Implementation of the AEGIS Single-Chip Secure Processor Using Physical Random Functions”, Computer Architecture, 2005, ISCA '05, Proceedings, 32nd International Symposium, Jun. 2005 (MIT Technical Report CSAIL CSG-TR-843, 2004.
Rivest, et al., “A Method for Obtaining Digital Signatures and Public-Key Cryptosystems”, Communication of the ACM, vol. 21 (2), pp. 120-126. (1978).
Hiltgen, et al., “Secure Internet Banking Authentication”, IEEE Security and Privacy, IEEE Computer Society, New York, NY, US, Mar. 1, 2006 (Mar. 1, 2006), pp. 24-31, XP007908655, ISSN: 1540-7993.
Hegt, “Analysis of Current and Future Phishing Attacks on Internet Banking Services”, Mater Thesis. Techische Universiteit Eindhoven—Department of Mathematics and Computer Science May 31, 2008 (May 31, 2008), pp. 1-149, XP002630374, Retrieved from the Internet: URL:http://alexandria.tue.nl/extral/afstversl/wsk-i/hgt2008.pdf [retrieved on Mar. 29, 2011] *pp. 127-134, paragraph 6.2*.
Gassend, et al., “Controlled Physical Random Functions”, In Proceedings of the 18th Annual Computer Security Conference, Las Vegas, Nevada, Dec. 12, 2002.
Wikipedia (Mar. 2003). “Integrated Circuit,” http://en.wikipedia.org/wiki/integrated—circuit. Revision as of Mar. 23, 2003.
Wikipedia (Dec. 2006). “Integrated circuit” Revision as of Dec. 10, 2006. http://en.widipedia.org/wiki/Integrated—circuit.
bellagiodesigns.com (Internet Archive Wayback Machine, www.bellagiodesigns.com date: Oct. 29, 2005).
Closed Loop Systems, The Free Dictionary, http://www.thefreedictionary.com/closed-loop+system (downloaded Dec. 1, 2011).
Feedback: Electronic Engineering, Wikipedia, p. 5 http://en.wikipedia.org/wiki/Feedback#Electronic—engineering (downloaded Dec. 1, 2011).
Galy et al. (Jul. 2007) “A full fingerprint verification system for a single-line sweep sensor.” IEEE Sensors J., vol. 7 No. 7, pp. 1054-1065.
Drahansky, Liveness Detection in Biometrics, Advanced Biometrics Technologies, 179-199 (Aug. 9, 2011).
“Bona fide, adv. and adj.” OED Online. Oxford University Press, Jun. 2014. Web. Jul. 17, 2014.
Blackberry. 2002. Research in Motion Limited, Waterloo, Ontario, Canada.
Droid. 2009. Lucasfilm Ltd., San Francisco, CA, USA.
Related Publications (1)
Number Date Country
20150379323 A1 Dec 2015 US
Provisional Applications (1)
Number Date Country
61616112 Mar 2012 US
Continuations (1)
Number Date Country
Parent 13763377 Feb 2013 US
Child 14849315 US