Embodiments of the present invention relate to authentication, and more particularly relate to techniques for authenticating a user based on personalized (i.e., user-provided) authentication data.
User authentication is the process of verifying the identity of a principal (i.e., a user) for a security-related purpose. Generally speaking, this verification is performed by testing one or more of the following “authentication factors:” (1) something the principal knows (referred to as a Type I authentication factor), (2) something the principal has (referred to as a Type II authentication factor), and (3) something the principal is (referred to as a Type III authentication factor).
In the field of computer security, a significant number of existing authentication systems rely on textual passwords or personal identification numbers (i.e., Type I authentication factors) in order to verify the identities of users. However, authentication based on textual authentication data is well-known to have several weaknesses. For example, relatively secure (i.e., complex) passwords are generally difficult to remember. Conversely, short or easily memorized passwords are relatively insecure and may be attacked through a variety of methods, such as a dictionary attack. In some cases, Type I authentication factors are combined with Type II authentication factors to increase the security of a system. For example, an ATM machine typically requires both a PIN and a physical ATM card to carry out transactions. However, Type II authentication factors may also be compromised, such as through loss or theft.
To address some of the above weaknesses, a growing number of computer-based authentication systems are incorporating authentication via Type III, or biometric, authentication factors. Exemplary biometric characteristics include fingerprints, voice, face geometry, retinal pattern, and the like. Since biometric characteristics are inherent to a particular individual, there is no need to memorize authentication information such as a textual password. In addition, many believe that a biometric characteristic cannot be stolen as easily as a physical (i.e., Type II) authentication device/token. To support these biometric systems, non-textual authentication interfaces have been developed. For example, acoustic-based interfaces have been developed to support authentication via voice recognition. Similarly, image and video-based interfaces have been developed to support authentication via facial recognition, fingerprint recognition, and the like.
Unfortunately, despite their advantages, biometric authentication systems still possess a number of weaknesses. Voice-based authentication, in particular, is problematic for several reasons. First, the enrollment/registration process necessary to register a user's voice into a voice-based authentication system can be both time-consuming and cumbersome. For example, this process typically requires the user to recite several sentences or paragraphs-worth of training data into the system to generate a voice model for the user. In addition, the training data must generally be recited in a quiet environment to capture the user's voice with adequate clarity and precision. As a result, users may be unwilling to spend the time and effort demanded by this process.
Second, authentication based on voice recognition is not particularly reliable. For example, it is possible for an attacker to impersonate the voice of a legitimate user and thus fool the system into providing access. This is known as a “false positive.” It is also possible for a user's voice to change temporarily (e.g., due to a cold or other illness), and thus cause the user to be denied access even though the user is legitimate. This is known as a “false negative.” In either situation, the authentication mechanism fails to operate as intended.
Third, voice recognition is generally a very complex process and may require significant resources to be performed with a reasonable degree of accuracy. For example, a large amount of storage resources may be necessary to store registered training data and voice models, and a large amount of computational resources may be necessary to verify a given voice sample against a registered model. Accordingly, the costs associated with deploying and maintaining a voice-based authentication system can be prohibitively high.
Fourth, since voice characteristics (and other types of biometric characteristics) are inherent to a user, they cannot be reset or reissued if the user's voice is somehow compromised (e.g., through impersonation or a voice recording). This is a significant concern because the legitimate user will have no way to regain secure access to the system. In addition, the legitimate user will no longer be able to user his or her voice as an authentication credential for other voice-based systems.
Embodiments of the present invention address the foregoing and other such problems by providing techniques for authenticating users based on personalized (i.e., user-provided) authentication data. In one set of embodiments, the personalized authentication data includes acoustic, image, and/or video data, but is exclusive of biometric data of the user. In this manner, existing acoustic, image, and video-based authentication interfaces/mechanisms may be leveraged, without the problems associated with verifying biometric data. In some embodiments, a user may enroll or register multiple pieces of personalized authentication data into an authentication system. The user may then select one piece of personalized authentication data for use by the system at a time of authentication.
According to one embodiment of the present invention, a method for facilitating user authentication comprises associating first non-textual authentication data with a user, where the first non-textual authentication data is provided or selected by the user, and where the first non-textual authentication data does not include biometric data of the user. The user is then authenticated based on the first non-textual authentication data. In some embodiments, the first non-textual authentication data includes acoustic data, such as a user-provided tone sequence or a user provided audio clip. In other embodiments, the first non-textual authentication data includes graphical data, such as a user-provided still image or a user-provided video clip. As described in further detail below, the authentication may be performed using existing acoustic-based or image/video-based authentication interfaces/mechanisms.
In one set of embodiments, the method above further comprises storing the first non-textual authentication data on a device usable by the user, such as a cellular phone, personal computer, personal digital assistant (PDA), smart card, security token, and the like. The first non-textual authentication data is then retrieved from the device at a time of authentication. In these embodiments, authenticating the user may be further based on one or more attributes of the device.
In another set of embodiments, the method above further comprises associating second non-textual authentication data with the user, where the second non-textual authentication data is different from the first non-textual authentication data, and where the second non-textual authentication data is provided by the user. In various embodiments, either the first non-textual authentication data or the second non-textual authentication data may be selected for use in authenticating the user.
In one set of embodiments, associating the first non-textual authentication data with the user comprises receiving the first non-textual authentication data from the user, and registering the first non-textual authentication data with an identity of the user in a registration database. For example, the first non-textual authentication data may be received via a computer (e.g., intranet, Internet) or voice-based (e.g., telephony, cellular) network.
In another set of embodiments, authenticating the user based on the first non-textual authentication data comprises receiving input non-textual authentication data from the user, and determining whether the input non-textual authentication data matches the first non-textual authentication data.
In another set of embodiments, authenticating the user based on the first non-textual authentication data comprises receiving a one-time password from the user, where the one-time password is derived from the first non-textual authentication data. The user is then authenticated based on the one-time password.
According to another embodiment of the present invention, an authentication system is provided. The system comprises a registration unit for registering non-textual authentication data with an identity of a user in a registration database, where the non-textual authentication data is provided by the user, and where the non-textual authentication data does not correspond to biometric data of the user. The system further comprises an authentication unit for authenticating the user based on the non-textual authentication data.
According to yet another embodiment of the present invention, a machine-readable medium is disclosed, the machine-readable medium having stored thereon a series of instructions which, when executed by a processing component, cause the processing component to facilitate user authentication. In various embodiments, the series of instructions cause the processing component to associate non-textual authentication data with a user, where the non-textual authentication data is provided by the user, and where the non-textual authentication data does not correspond to biometric data of the user. The series of instructions further cause the processing component to authenticate the user based on the non-textual authentication data.
A further understanding of the nature and advantages of the embodiments disclosed herein may be realized by reference to the remaining portions of the specification and the attached drawings.
Various embodiments in accordance with the present invention will be described with reference to the drawings, in which:
In the following description, for the purposes of explanation, numerous specific details are set forth in order to provide an understanding of the present invention. It will be apparent, however, to one skilled in the art that the present invention may be practiced without some of these specific details.
Embodiments of the present invention provide techniques for authenticating users based on personalized authentication data. A general approach to user authentication includes two steps—(1) enrollment/registration of authentication data, and (2) verification of input authentication data against the registered authentication data. Embodiments of the present invention allow a user to provide the authentication data they wish to enroll into a system. The authentication data may be provided, for example, by being uploaded or played back from a storage device of the user. The authentication data may also be provided by being selected by the user from a data library of a third party; in this embodiment, the selection of a particular piece of authentication data is provided to the system for enrollment. This user-provided, or personalized, authentication data is then used to authenticate the user at a time of verification. In various embodiments, the personalized authentication data may include acoustic data, image data, video data, or any other type of non-textual data. However, the personalized authentication data will generally not include biometric data of the user. Thus, the problems associated with registering and verifying biometric data (such as voice data) may be minimized or avoided. As described herein, the step of verification/authentication may be performed using existing interfaces and algorithms for verifying acoustic data, image data, and/or video data.
Embodiments of the present invention may be used in a variety of different domains and contexts. For example, one embodiment may be used as part of a voice-based authentication system that enables users to register and verify personalized authentication data using a standard telephone or cellular phone. Another embodiment may be used as part of a network-based authentication system that enables users to register and verify personalized authentication data via a computer network such as the Internet. In the present disclosure, the term “personalized authentication data” is used interchangeably with the term “personalized security signature,” and generically refers to any type of user-provided or user-selected data that facilitates user authentication.
At step 102, non-textual authentication data is associated with a user, where the non-textual authentication data is provided by the user (i.e., personalized by the user), and where the non-textual authentication data is exclusive of biometric data of the user. In a particular embodiment, this step corresponds to registering or enrolling the non-textual authentication data with an identity of the user. The non-textual authentication data may include acoustic data, such as a user-provided tone sequence or a user-provided audio clip. For example, the user may use the alphanumeric keys on a telephone to select a series of notes, or the user may provide a portion of his or her favorite song. Alternatively, the non-textual authentication data may include graphical data, such as a user-provided still image or a user-provided video clip.
In one set of embodiments, the non-textual authentication data described in step 102 may be stored on a device of the user such as a cellular phone, personal computer, smart card, security token, or the like. In this manner, the user may have convenient access to his or her personalized authentication data. If for example, the user needs to login into a system using the stored authentication data, the authentication data may be retrieved from the device at a time of verification/authentication. In addition, the device may be configured to provide a number of additional functions. For example, the device may be configured to generate randomized authentication data on behalf of the user. In another embodiment, attributes of the device may be used, in conjunction with the non-textual authentication data, to verify the identity of the user at a time of authentication. Examples of these devices are discussed in greater detail with respect to
At step 104, the user is authenticated/verified based on the non-textual authentication data associated/registered with the user. In various embodiments, authentication/verification comprises comparing a piece of input authentication data provided by the user with the registered authentication data. Existing mechanisms/algorithms for verifying acoustic, image, and/or video-based data (such as pattern matching algorithms) may be leveraged to perform the authentication.
By utilizing personalized authentication data rather than biometric authentication data to facilitate user authentication, a number of advantages are achieved. For example, there is no need for the user to undergo a lengthy and cumbersome enrollment process to generate a biometric model. Instead, the user need only playback (in the case of registering over a telephone) or upload (in the case of registering over a computer network) his or her personalized authentication data to the authentication system, where it is simply captured and stored. Further, since embodiments of the present invention can make use of existing and well-known techniques for verifying non-textual data, the reliability and performance of authentication is improved. Yet further, the personalized authentication data may be easily reset with a different piece of data if it is compromised.
It should be appreciated that the specific steps illustrated in
In alternative embodiments, the non-textual authentication data may be received from a third party (such as a service provider or an authentication system) rather than the user. For example, a service provider may wish to pre-enroll users of its service by assigning an initial piece of authentication data to each user.
Once the non-textual authentication data is received, it is registered with an identity of the user in a registration database (step 204). The identity of the user may include, for example, a login or account name, as well as other identity information such as real name, address, and the like. In one set of embodiments, this identity information may already be stored in the registration database at the time of registering the non-textual authentication data. In other embodiments, the identity information and the non-textual authentication data may be provided concurrently for registration in the registration database.
At step 304, a determination is made whether the input authentication data matches the non-textual authentication data associated with the user. If the input authentication data matches the non-textual authentication data, authentication succeeds and access is provided to the desired resource/service. If the input authentication data does not match the non-textual authentication data, authentication fails and access is denied. Generally speaking, this determination may be performed using any of a number of well-known algorithms for verifying non-textual data, such as techniques for acoustic, image, or video-based pattern matching. One of ordinary skill in the art would recognize many variations, modifications, and alternatives.
At step 402, a one-time password (OTP) is received from the user, where the one-time password is derived from the non-textual authentication data. In one set of embodiments, the OTP is generated by the device storing the non-textual authentication data. In other embodiments, the OTP is generated by another device. The OTP may correspond to a mathematical algorithm type OTP, a time-synchronized type OTP, a challenge type OTP, or the like. One of ordinary skill in the art would recognize many variations, modifications, and alternatives.
Once the OTP has been received, the user is authenticated based on the OTP (step 404). This may include, for example, extracting/decoding authentication data from the OTP, and then comparing the extracted authentication data to the non-textual authentication data associated with the user. It should be appreciated that the exact mechanism for authenticating the user will depend on the type of the OTP implemented.
At steps 502 and 504, non-textual authentication data is received from, and associated (i.e., registered), with a user. As described above, the non-textual authentication data may include acoustic data, image data, video data, and the like. At step 506, a determination is made whether the user has additional pieces of non-textual authentication data to be registered. If so, steps 502 and 5034 are repeated for each additional piece of non-textual authentication data. In various embodiments, each piece of non-textual authentication data associated with the user at step 504 may be used to verify/authenticate the user. In one set of embodiments, the various pieces of non-textual authentication data may include the same type of data. For example, they may all include acoustic data. In another set of embodiments, the various pieces of non-textual authentication data may include different types of data. For example, first and second pieces of non-textual authentication data may include acoustic data and image data respectively. In most cases, the various pieces of authentication data will not be identical.
At step 506, selection of a specific piece of non-textual authentication data associated with the user is enabled, such that the selected authentication data is used to authenticate the user. This selection may be performed by either the user or a third party, such as a service provider or the authentication system. In addition, this selection may be performed for any of a number of reasons. For example, the user or the authentication system may determine that the first authentication data has been compromised, and therefore switch to using the second authentication data for verification. In one set of embodiments, the selection of a specific piece of authentication data may be performed prior to a time of authentication, such as during the association step 102 of
It should be appreciated that the specific steps illustrated in
At step 602, the type of registration is determined. In various embodiments, the type of registration will depend on the type of personalized authentication data that is provided by the user for registration. It should be appreciated that although only three types of authentication data are illustrated (acoustic, image, and video) any other type of non-textual authentication data may be supported.
Once a particular type of registration as been determined, control proceeds to an appropriate registration block (604, 622, 640). Each registration block includes a registration preprocessing block 606, 624, 642 configured to preprocess the user-provided authentication data. In the case of acoustic authentication data, registration preprocessing block 604 may include steps such as analog-to-digital conversion, data compression, and the like. One of ordinary skill in the art will recognize that the preprocessing may depend on the medium through which the authentication data is transmitted from the user. For example, if an audio clip is transmitted over a computer network such as the Internet, the clip will already be in the digital domain and thus would not require analog-to-digital conversion. In the case of image or video-based authentication data, registration preprocessing blocks 624, 642 may include steps such as noise reduction, data compression, and the like. In some embodiments, the preprocessing performed by registration preprocessing blocks 606, 622, 640 may be omitted.
The preprocessed authentication data is then analyzed in an appropriate registration analysis block 608, 628, 644. Registration analysis blocks 608, 626, 644 may include steps for feature extraction, feature transformation, and the like. Registration analysis blocks 608, 626, 644 may also be configured to store the analyzed authentication data in a storage area, such as registration database, for later retrieval at a time of verification.
At verification blocks 610, 628, 646, input authentication data is received and compared to the authentication data registered in registration blocks 604, 622, 640. Each verification block includes a verification preprocessing block 612, 630, 648 that is similar in function to registration preprocessing blocks 606, 624, 642. In some embodiments, the preprocessing performed by verification preprocessing block 612, 630, 648 may be omitted. Once the input authentication data is preprocessed, it is analyzed against registered authentication data in an appropriate verification analysis block 614, 632, 650. In various embodiments, the steps performed in verification analysis block 614, 632, 650 depend on the steps performed in the corresponding registration analysis block 608, 626, 644. For example, if acoustic registration analysis block 608 simply involves extracting time samples, acoustic verification analysis block 614 will compare the register time samples with time samples provided from acoustic verification preprocessing block 612. Other types of algorithms for verification/analysis of acoustic, image, and video-based data are well-known and are contemplated to be within the scope of the present invention.
Verification analysis block 614, 632, 650 outputs a measure that is passed to decision block 616, 634, 652. If the measure indicates that the input authentication data matches the registered authentication data, the identity of the user is verified (618, 636, 654). On the other hand, if the measure indicates that the input authentication data does not match the registered authentication data, the verification attempt fails (620, 638, 656). The determination of whether a match has occurred may be based on any of a number of user-defined criteria.
In one set of embodiments, user devices 704, 706, 708, 710 may be general purpose personal computers (including, merely by way of example, personal computers and/or laptop computers running various versions of Microsoft Windows and/or Apple Macintosh operating systems), cellular phones or PDAs (running software such as Microsoft Windows Mobile and being Internet, e-mail, SMS, Blackberry, or other communication protocol enabled), and/or workstation computers running any of a variety of commercially-available UNIX or UNIX-like operating systems (including without limitation the variety of GNU/Linux operating systems). Alternatively, user devices 704, 706, 708, 710 may be any other device, such as a smart card, security token, flash memory card, etc. that is capable of storing data in a nonvolatile memory. Although exemplary system environment 700 is shown with four user devices, any number of user devices may be supported.
In some embodiments, system environment 700 includes a network 712. Network 712 may be any type of network familiar to those skilled in the art that can support data communications using any of a variety of commercially-available protocols, including without limitation TCP/IP, SNA, IPX, AppleTalk, and the like. Merely by way of example, network 712 can be a local area network (LAN), such as an Ethernet network, a Token-Ring network and/or the like; a wide-area network; a virtual network, including without limitation a virtual private network (VPN); the Internet; an intranet; an extranet; a public switched telephone network (PSTN); a cellular telephony network; an infra-red network; a wireless network (e.g., a network operating under any of the IEEE 802.11 suite of protocols, the Bluetooth protocol known in the art, and/or any other wireless protocol); and/or any combination of these and/or other networks.
System environment 700 also includes one or more server computers 702 which may be general purpose computers, specialized server computers (including, merely by way of example, PC servers, UNIX servers, mid-range servers, mainframe computers rack-mounted servers, etc.), server farms, server clusters, or any other appropriate arrangement and/or combination. In various embodiments, server 702 may be adapted to run one or more services or software applications described in the foregoing disclosure. For example, server 702 may act as an authentication server that is configured to received authentication data from user devices 704, 706, 708, 710 via network 712 and authenticate users of said devices based on the received authentication data.
Server 702 may run an operating system including any of those discussed above, as well as any commercially available server operating system. Server 702 may also run any of a variety of additional server applications and/or mid-tier applications, including HTTP servers, FTP servers, CGI servers, Java servers, database servers, and the like. Exemplary database servers include without limitation those commercially available from Oracle, Microsoft, Sybase, IBM and the like.
System environment 700 may also include one or more databases 714. For instance, databases 714 may include a registration database configured to store associations between personalized authentication data and identities of users. Databases 714 may reside in a variety of locations. By way of example, one or more of databases 714 may reside on a storage medium local to (and/or resident in) one or more of the computers/devices 702, 704, 706, 708, 710. Alternatively, databases 714 may be remote from any or all of the computers/devices 702, 704, 706, 708, 710, and/or in communication (e.g., via network 712) with one or more of these. In one set of embodiments, databases 714 may reside in a storage-area network (SAN) familiar to those skilled in the art. Similarly, any necessary files for performing the functions attributed to the computers/devices 702, 704, 706, 708, 710 may be stored locally on the respective device and/or remotely, as appropriate. In one set of embodiments, databases 714 may include relational databases, such as Oracle 10g, that are adapted to store, update, and retrieve data in response to SQL-formatted commands.
According to one embodiment of the present invention, a user may register personalized authentication data by transmitting the personalized authentication data to authentication server 702 via network 712. For example, the personalized authentication data may be stored on a device of the user (e.g., one or more of user devices 704, 706, 708, 710), and transmitted to server 702 by being retrieved from the device and uploaded or played back via network 712. In various embodiments, authentication server 702 is configured to perform the processing of registration blocks 604, 622, 640 of
Once the personalized authentication data has been registered, the user may be authenticated by transmitting input authentication data to authentication server 702 via network 712. In various embodiments, the input authentication data corresponds to the personalized authentication data previously registered by the user. As in the case of registration, the input authentication data may be stored on a user device 704, 706, 708, 710 and transmitted to authentication server 702 by being retrieved from the device and uploaded or played back via network 712. In various embodiments, authentication server 702 is configured to perform the processing of verification blocks 610, 628, 646 and decision blocks 616, 634, 652 of
Computer system 800 may additionally include a computer-readable storage media reader 812, a communications subsystem 814 (e.g., a modem, a network card (wireless or wired), an infra-red communication device, etc.), and working memory 818, which may include RAM and ROM devices as described above. In some embodiments, computer system 800 may also include a processing acceleration unit 816, which can include a digital signal processor (DSP), a special-purpose processor, and/or the like.
Computer-readable storage media reader 812 can further be connected to a computer-readable storage medium 810, together (and, optionally, in combination with storage device(s) 808) comprehensively representing remote, local, fixed, and/or removable storage devices plus storage media for temporarily and/or more permanently containing computer-readable information. Communications system 814 may permit data to be exchanged with network 712 and/or any other device described above with respect to system environment 700.
Computer system 800 may also comprise software elements, shown as being currently located within working memory 818, including an operating system 820 and/or other code 822, such as an application program (which may be a client application, Web browser, mid-tier application, RDBMS, etc.). It should be appreciated that alternative embodiments of computer system 800 may have numerous variations from that described above. For example, customized hardware might also be used and/or particular elements might be implemented in hardware, software (including portable software, such as applets), or both. Further, connection to other computing devices such as network input/output devices may be employed.
Storage media and computer readable media for containing code, or portions of code, can include any appropriate media known or used in the art, including storage media and communication media, such as but not limited to volatile and non-volatile, removable and non-removable media implemented in any method or technology for storage and/or transmission of information such as computer readable instructions, data structures, program modules, or other data, including RAM, ROM, EEPROM, flash memory or other memory technology, CD-ROM, digital versatile disk (DVD) or other optical storage, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices, data signals, data transmissions, or any other medium which can be used to store or transmit the desired information and which can be accessed by a computer.
Further, while the present invention has been described using a particular combination of hardware and software, it should be recognized that other combinations of hardware and software are also within the scope of the present invention. The present invention may be implemented only in hardware, or only in software, or using combinations thereof.
The specification and drawings are, accordingly, to be regarded in an illustrative rather than a restrictive sense. Many variations of the invention will become apparent to those skilled in the art upon review of the disclosure. The scope of the invention should, therefore, be determined not with reference to the above description, but instead should be determined with reference to the pending claims along with their full scope or equivalents.
This application claims the benefit of U.S. Provisional Application No. 60/869,385, filed Nov. 11, 2006 by Varghese et al. and entitled “System and Method for Personalized Security Signature,” the disclosure of which is incorporated herein by reference in its entirety for all purposes.
Number | Name | Date | Kind |
---|---|---|---|
D95205 | Harrison | Apr 1935 | S |
D298837 | Thomas | Dec 1988 | S |
5416895 | Anderson et al. | May 1995 | A |
5428349 | Baker | Jun 1995 | A |
5452413 | Blades | Sep 1995 | A |
5555365 | Selby et al. | Sep 1996 | A |
5559961 | Blonder | Sep 1996 | A |
5572644 | Liaw et al. | Nov 1996 | A |
5577125 | Salahshour et al. | Nov 1996 | A |
5604854 | Glassey | Feb 1997 | A |
5623591 | Cseri | Apr 1997 | A |
5664099 | Ozzie et al. | Sep 1997 | A |
5798760 | Vayda et al. | Aug 1998 | A |
D400195 | Utesch | Oct 1998 | S |
5821933 | Keller et al. | Oct 1998 | A |
5875296 | Shi et al. | Feb 1999 | A |
5928364 | Yamamoto | Jul 1999 | A |
5949348 | Kapp et al. | Sep 1999 | A |
5966127 | Yajima | Oct 1999 | A |
D419907 | Vogelbruch | Feb 2000 | S |
6023506 | Ote et al. | Feb 2000 | A |
6064972 | Jankowitz et al. | May 2000 | A |
6111984 | Fukasawa | Aug 2000 | A |
6209102 | Hoover | Mar 2001 | B1 |
6209104 | Jalili | Mar 2001 | B1 |
6240183 | Marchant | May 2001 | B1 |
6253326 | Lincke et al. | Jun 2001 | B1 |
6263447 | French et al. | Jul 2001 | B1 |
6282551 | Anderson et al. | Aug 2001 | B1 |
6343361 | Nendell et al. | Jan 2002 | B1 |
6369839 | Peterson | Apr 2002 | B1 |
6448987 | Easty et al. | Sep 2002 | B1 |
6658574 | Anvekar | Dec 2003 | B1 |
6718471 | Kashima | Apr 2004 | B1 |
6720860 | Narayanaswami | Apr 2004 | B1 |
6725422 | Bauchot et al. | Apr 2004 | B1 |
6741268 | Hayakawa | May 2004 | B1 |
D492691 | Kortis | Jul 2004 | S |
D493471 | McIntosh | Jul 2004 | S |
6853973 | Mathews et al. | Feb 2005 | B2 |
D505135 | Sapp et al. | May 2005 | S |
6895502 | Fraser | May 2005 | B1 |
6895558 | Loveland | May 2005 | B1 |
6934860 | Goldstein | Aug 2005 | B1 |
6972363 | Georges et al. | Dec 2005 | B2 |
7007298 | Shinzaki et al. | Feb 2006 | B1 |
7036091 | Nguyen | Apr 2006 | B1 |
7054819 | Loveland | May 2006 | B1 |
7082227 | Baum et al. | Jul 2006 | B1 |
7100049 | Gasparini et al. | Aug 2006 | B2 |
7137008 | Hamid et al. | Nov 2006 | B1 |
D539809 | Totten et al. | Apr 2007 | S |
7200747 | Riedel et al. | Apr 2007 | B2 |
7219368 | Juels et al. | May 2007 | B2 |
7237264 | Graham et al. | Jun 2007 | B1 |
7240367 | Park | Jul 2007 | B2 |
7437024 | Baum et al. | Oct 2008 | B2 |
7523067 | Nakajima | Apr 2009 | B1 |
7586635 | Maeda et al. | Sep 2009 | B2 |
7596701 | Varghese et al. | Sep 2009 | B2 |
7616764 | Varghese et al. | Nov 2009 | B2 |
7822990 | Varghese et al. | Oct 2010 | B2 |
20010027529 | Sasabe et al. | Oct 2001 | A1 |
20020013905 | Hamada | Jan 2002 | A1 |
20020029341 | Juels et al. | Mar 2002 | A1 |
20020049614 | Rice et al. | Apr 2002 | A1 |
20020087894 | Foley et al. | Jul 2002 | A1 |
20020122031 | Maglio et al. | Sep 2002 | A1 |
20020188872 | Willeby | Dec 2002 | A1 |
20030005329 | Ikonen | Jan 2003 | A1 |
20030018558 | Heffner et al. | Jan 2003 | A1 |
20030097593 | Sawa et al. | May 2003 | A1 |
20030159050 | Gantman et al. | Aug 2003 | A1 |
20030182558 | Lazzaro et al. | Sep 2003 | A1 |
20030210127 | Anderson | Nov 2003 | A1 |
20040010721 | Kirovski et al. | Jan 2004 | A1 |
20040030933 | Park | Feb 2004 | A1 |
20040030934 | Mizoguchi et al. | Feb 2004 | A1 |
20040034801 | Jaeger | Feb 2004 | A1 |
20040059951 | Pinkas et al. | Mar 2004 | A1 |
20040073809 | Wing Keong | Apr 2004 | A1 |
20040083389 | Yoshida | Apr 2004 | A1 |
20040117320 | Morioka et al. | Jun 2004 | A1 |
20040128534 | Walker | Jul 2004 | A1 |
20040139353 | Forcade | Jul 2004 | A1 |
20040153650 | Hillmer et al. | Aug 2004 | A1 |
20040153660 | Gaither et al. | Aug 2004 | A1 |
20040168083 | Gasparini et al. | Aug 2004 | A1 |
20040215980 | Hamid | Oct 2004 | A1 |
20040221163 | Jorgensen | Nov 2004 | A1 |
20040230843 | Jansen | Nov 2004 | A1 |
20040250138 | Schneider | Dec 2004 | A1 |
20050010768 | Light et al. | Jan 2005 | A1 |
20050015601 | Tabi | Jan 2005 | A1 |
20050044425 | Hypponen | Feb 2005 | A1 |
20050097320 | Golan et al. | May 2005 | A1 |
20050097339 | Wiley et al. | May 2005 | A1 |
20050144451 | Voice et al. | Jun 2005 | A1 |
20050154676 | Ronning et al. | Jul 2005 | A1 |
20050193208 | Charrette, III et al. | Sep 2005 | A1 |
20050204131 | Kovarik, Jr. | Sep 2005 | A1 |
20050204145 | Makishima | Sep 2005 | A1 |
20050251752 | Tan et al. | Nov 2005 | A1 |
20050278542 | Pierson et al. | Dec 2005 | A1 |
20050278647 | Leavitt et al. | Dec 2005 | A1 |
20060011045 | Yamashita et al. | Jan 2006 | A1 |
20060212829 | Yahiro et al. | Sep 2006 | A1 |
20060282660 | Varghese et al. | Dec 2006 | A1 |
20060288225 | Jung et al. | Dec 2006 | A1 |
20070028111 | Covely | Feb 2007 | A1 |
20070036314 | Kloberdans et al. | Feb 2007 | A1 |
20070097351 | York et al. | May 2007 | A1 |
20070165849 | Varghese et al. | Jul 2007 | A1 |
20070169190 | Kolton et al. | Jul 2007 | A1 |
20070211925 | Aoki et al. | Sep 2007 | A1 |
20090089869 | Varghese et al. | Apr 2009 | A1 |
Number | Date | Country |
---|---|---|
1615181 | Nov 2006 | EP |
2313460 | Nov 1997 | GB |
2002-222172 | Aug 2002 | JP |
2004-258845 | Sep 2004 | JP |
WO 9618139 | Jun 1996 | WO |
WO0118636 | Mar 2001 | WO |
WO0165375 | Sep 2001 | WO |
WO 2004053674 | Jun 2004 | WO |
WO2005025292 | Mar 2005 | WO |
WO 2006010058 | Jan 2006 | WO |
WO 2006118968 | Nov 2006 | WO |
WO 2007087352 | Aug 2007 | WO |
Entry |
---|
Birget et al., “The Graphical Passwords Project,” printed from website http://clam.rutgers.edu/˜birget/grPssw/, on Nov. 26, 2007 (2 pages). |
“Echo System,” Identita: The Powered Card™ Identity Provider, printed from website http://222.identita.com/sol—eco—sys.html, on Nov. 26, 2007 (2 pages). |
Identita Technologies Inc., “Audio Authentication Devices,” (2 pages). |
Frost & Sullivan, “Leading European Financial Institution Implements Voice Verification Biometrics to Enrich Customer Experience,” Jul. 2006 (13 pages). |
“VoiceVault,” VoiceVault™ World Leaders in Voice Verification, printed from website http://www.voicevault.com/ca.aspx on Nov. 13, 2007 (2 pages). |
“VoiceVault,” VoiceVault™ World Leaders in Voice Verification, “What is ‘two factor authentication’?,” printed from website http://www.voicevault.com/twofactor.aspx on Nov. 13, 2007 (2 pages). |
“VoiceVault,” VoiceVault™ World Leaders in Voice Verification, “How VoiceVault Works,” printed from website http://www.voicevault.com/works.aspx on Nov. 13, 2007 (1 page). |
U.S. Appl. No. 12/882,096, filed Sep. 14, 2010, Varghese et al. |
Notice of Allowance for U.S. Appl. No. 11/169,564 mailed on Jun. 1, 2009; 9 pages. |
Non-Final Office Action for U.S. Appl. No. 11/169,564 mailed on Nov. 13, 2008; 22 pages. |
Supplemental Non-Final Office Action for U.S. Appl. No. 11/318,414 mailed on Jan. 27, 2009; 10 pages. |
Non-Final Office Action for U.S. Appl. No. 11/318,414 mailed on Nov. 13, 2008; 19 pages. |
Notice of Allowance for U.S. Appl. No. 11/318,424 mailed on Jun. 16, 2010; 7 pages. |
Advisory Action for U.S. Appl. No. 11/318,424 mailed on May 25, 2010; 3 pages. |
Final Office Action for U.S. Appl. No. 11/318,424 mailed on Mar. 10, 2010; 11 pages. |
Non-Final Office Action for U.S. Appl. No. 11/318,424 mailed on Sep. 1, 2009; 14 pages. |
Advisory Action for U.S. Appl. No. 11/318,424 mailed on Jul. 30, 2009; 3 pages. |
Final Office Action for U.S. Appl. No. 11/318,424 mailed on Jun. 2, 2009; 16 pages. |
Non-Final Office Action for U.S. Appl. No. 11/318,424 mailed on Dec. 1, 2008; 12 pages. |
Notice of Allowance for U.S. Appl. No. 11/340,376 mailed on Jun. 12, 2009; 11 pages. |
Non-Final Office Action for U.S. Appl. No. 11/340,376 mailed on Jan. 15, 2009; 11 pages. |
Non-Final Office Action for U.S. Appl. No. 11/340,376 mailed on Aug. 4, 2008; 5 pages. |
Advisory Action for U.S. Appl. No. 11/340,376 mailed on May 27, 2008; 3 pages. |
Final Office Action for U.S. Appl. No. 11/340,376 mailed on Mar. 19, 2008; 10 pages. |
Non-Final Office Action for U.S. Appl. No. 11/340,376 mailed on Jul. 27, 2007; 10 pages. |
Notice of Allowance for U.S. Appl. No. 11/412,997 mailed on Dec. 14, 2010; 6 pages. |
Non-Final Office Action for U.S. Appl. No. 11/412,997 mailed on Sep. 13, 2010; 18 pages. |
Final Office Action for U.S. Appl. No. 11/412,997 mailed on May 11, 2010; 24 pages. |
Non-Final Office Action for U.S. Appl. No. 11/412,997 mailed on Nov. 16, 2009; 19 pages. |
Non-Final Office Action for U.S. Appl. No. 11/412,997 mailed on Apr. 10, 2009; 37 pages. |
Non-Final Office Action for U.S. Appl. No. 11/412,997 mailed on Sep. 30, 2008; 27 pages. |
International Search Report for PCT Application No. PCT/US2005/024376 mailed on Apr. 7, 2006; 4 pages. |
International Preliminary Report on Patentability for PCT Application No. PCT/US2005/024376 mailed on Apr. 13, 2007; 4 pages. |
Written Opinion for PCT Application No. PCT/US2005/024376 mailed on Apr. 7, 2006; 4 pages. |
International Search Report for PCT Application No. PCT/US06/16085 mailed on Jul. 7, 2008; 2 pages. |
Written Opinion for PCT Application No. PCT/US06/16085 mailed on Jul. 7, 2008; 9 pages. |
International Preliminary Report on Patentability in International Application No. PCT/US2006/016085 mailed on Aug. 28, 2008; 10 pages. |
International Search Report for PCT Application No. in PCT Application No. PCT/US07/01899 mailed on Feb. 14, 2008; 2 pages. |
Written Opinion for PCT Application No. PCT/US07/01899 mailed on Feb. 14, 2008; 4 pages. |
“How to Set a Wallpaper on the Background of a Folder Window”, Microsoft Corporation, Document No. 879735, http://support.microsoft.com/kb/879735/ja (Dec. 27, 2004). |
Non-Final Office Action for U.S. Appl. No. 12/260,544 (May 22, 2013). 14 pages. |
Government of India Patent Office Examination Report for application IN4248/KOLNP/2007 mailed Jan. 30, 2014. |
Final Office Action for U.S. Appl. No. 12/260,544 (Oct. 31, 2013) 15 pages. |
Final Office Action for U.S. Appl. No. 12/260,544 (Nov. 15, 2012). |
European Patent Office (EPO) Communication for European application EP11169063.2 (Nov. 2, 2012). |
European Examination Report for patent application EP5773133.3 (Apr. 10, 2012 ). |
European Examination Report for patent application EP7762565.5 (Jul. 6, 2012). |
Final Office Action for U.S. Appl. No. 12/882,096, mailed on Oct. 18, 2012. |
European Search Report for application EP06758687.5 (Mar. 14, 2012). |
European Search Report for application EP11169063.2 (Mar. 21, 2012). |
“ssh—OpenSSH SSH client (remote login program)” BSD General Commands Manual, from internet: http://unixhelp.ed.ac.uk/CGI/man-cgi?ssh+1 (Sep. 1999). |
Final Office Action for U.S. Appl. No. 12/882,096, mailed on Jun. 21, 2012. |
Non-Final Office Action for U.S. Appl. No. 12/260,544, mailed on Jun. 7, 2012. |
Office Action from China Intellectual Property Office for application CN200780003321.5 (Jun. 18, 2012). |
Notice of Allowance for U.S. Appl. No. 12/260,544 (Mar. 3, 2014) 5 pages. |
Number | Date | Country | |
---|---|---|---|
20080209526 A1 | Aug 2008 | US |
Number | Date | Country | |
---|---|---|---|
60869385 | Dec 2006 | US |