Adjustment of knowledge-based authentication

Information

  • Patent Grant
  • 10740762
  • Patent Number
    10,740,762
  • Date Filed
    Monday, December 17, 2018
    6 years ago
  • Date Issued
    Tuesday, August 11, 2020
    4 years ago
Abstract
Systems and methods are provided for adjustment of difficulty level, quantity, and/or other parameters of knowledge-based authentication. The questions may be asked of a consumer based on one or more generated risk scores associated with the consumer's behavior, which may be derived from information received from the consumer, a representative responsible for entering information regarding the consumer, and/or from the consumer's computing device.
Description
BACKGROUND

Consumers requesting a new account or enrollment for a new service or product may need to be authenticated. Authentication may involve generating and presenting questions to consumers regarding identity.


SUMMARY

Disclosed herein are systems and methods for setting authentication requirements for a consumer. In one embodiment, a computing system comprises: one or more computer processors and a computer readable storage configured to store instructions executable by the one or more computer processors in order to cause the computing system to: receive a request to set authentication requirements for a consumer in response to a request from the consumer to open a new account, wherein an identity of the consumer is unknown to operators of the computing system; determine device identification information associated with a device from which opening of the new account was requested by the consumer, the device identification information comprising one or more of device ID, device location, browser time zone, browser language settings, proxy settings, stated IP address, or real IP address; derive one or more behaviors of the consumer based on at least the device identification information, the one or more behaviors comprising one or more of a frequency of authentication requests for the consumer, total amount of time taken by the consumer to answer authentication questions, or whether the device location matches a location indicated by a browser executing on the device; calculate a risk of fraud associated with the consumer based on the one or more derived behaviors of the consumer; and determine authentication requirements for the consumer based on the calculated risk of fraud, wherein the authentication requirements indicate one or more of a number of questions the consumer is required to correctly answer, types of questions the consumer is required to answer, or a percentage of questions that must be correctly answered.


In accordance with another aspect of the disclosure, a non-transitory computer-readable storage medium comprises computer-executable instructions that direct a computing system to: receive a request to set authentication requirements for a consumer in response to a request from the consumer to open a new account, wherein an identity of the consumer is unknown to operators of the computing system; determine device identification information associated with a device from which opening of the new account was requested by the consumer, the device identification information comprising one or more of device ID, device location, browser time zone, browser language settings, proxy settings, stated IP address, or real IP address; derive one or more behaviors of the consumer based on at least the device identification information, the one or more behaviors comprising one or more of a frequency of authentication requests for the consumer, total amount of time taken by the consumer to answer authentication questions, or whether the device location matches a location indicated by a browser executing on the device; calculate a risk of fraud associated with the consumer based on the one or more derived behaviors of the consumer; and determine authentication requirements for the consumer based on the calculated risk of fraud, wherein the authentication requirements indicate one or more of a number of questions the consumer is required to correctly answer, types of questions the consumer is required to answer, or a percentage of questions that must be correctly answered.


In accordance with another aspect of the disclosure, a computer-implemented method of setting authentication requirements for a consumer comprises: receiving a request to set authentication requirements for a consumer in response to a request from the consumer to open a new account, wherein an identity of the consumer is unknown to operators of the computing system; determining device identification information associated with a device from which opening of the new account was requested by the consumer, the device identification information comprising one or more of device ID, device location, browser time zone, browser language settings, proxy settings, stated IP address, or real IP address; deriving one or more behaviors of the consumer based on at least the device identification information, the one or more behaviors comprising one or more of a frequency of authentication requests for the consumer, total amount of time taken by the consumer to answer authentication questions, or whether the device location matches a location indicated by a browser executing on the device; calculating a risk of fraud associated with the consumer based on the one or more derived behaviors of the consumer; and determining authentication requirements for the consumer based on the calculated risk of fraud, wherein the authentication requirements indicate one or more of a number of questions the consumer is required to correctly answer, types of questions the consumer is required to answer, or a percentage of questions that must be correctly answered.





BRIEF DESCRIPTION OF THE DRAWINGS

The foregoing aspects and many of the attendant advantages will become more readily appreciated as the same become better understood by reference to the following detailed description, when taken in conjunction with the accompanying drawings, wherein:



FIG. 1A is a flow diagram illustrating one embodiment of data flow in an illustrative operating environment for knowledge-based authentication. FIG. 1B is a block diagram illustrating an embodiment of system architecture in which multiple consumers 102, 104, 106 are in communication with an authentication processing system



FIG. 2 is a block diagram illustrating one embodiment of a system for knowledge-based authentication.



FIG. 3 is a flowchart that illustrates one embodiment of a process of authenticating the identity of a consumer.



FIG. 4 is an illustrative user interface that may be generated and presented to a consumer, which includes questions presented to the consumer for the purpose of opening a new account and also tools that gather browser and/or computing device related information.



FIG. 5 is an illustrative user interface that may be generated and presented to a mobile consumer, which includes questions presented to the consumer for the purpose of opening a new account and also tools that gather mobile device and related information.



FIG. 6 is a flowchart that illustrates one embodiment of determining a risk score and using the risk score to determine whether a consumer's fraud risk is so high that the system will refuse to authenticate the consumer.



FIG. 7 is a flowchart that illustrates one embodiment of tracking behavior of a user with high fraud risk and leading the user to a dead-end authentication process.



FIG. 8A is an illustrative user interface that may be generated and presented to a consumer that is determined to be involved in potentially fraudulent authentication.



FIG. 8B is an illustrative user interface that may be generated and presented to a user who is determined to be involved in potentially fraudulent authentications for purposes of capturing information from the user.



FIG. 8C is an illustrative user interface that may be generated and presented to a user who is determined to be involved in potentially fraudulent authentication.



FIG. 9 is a flowchart that illustrates one embodiment of a method of alerting a consumer that potentially fraudulent authentication was attempted and storing information collected from the authentication session.





DETAILED DESCRIPTION

Authentication may involve detecting risk of fraudulent activity, such as stealing another consumer's identity and trying to open a new account under another person's name. Described herein are systems and methods for adjusting a difficulty level and or quantity of knowledge-based questions that may be asked of a consumer based on one or more generated risk scores associated with the consumer's behavior, which may be derived from information received from the consumer and/or from the consumers' computing device, for example.


Various embodiments of systems, methods, processes, and data structures will now be described with reference to the drawings. Variations to the systems, methods, processes, and data structures which represent other embodiments will also be described. Certain aspects, advantages, and novel features of the systems, methods, processes, and data structures are described herein. It is to be understood that not necessarily all such advantages may be achieved in accordance with any particular embodiment. Accordingly, the systems, methods, processes, and/or data structures may be embodied or carried out in a manner that achieves one advantage or group of advantages as taught herein without necessarily achieving other advantages as may be taught or suggested herein.


Example Knowledge-Based Authentication System Environment and Data Flow



FIG. 1A is a flow diagram illustrating one embodiment of data flow in an illustrative environment for a knowledge-based authentication system. One or more consumers may want to open an account. For example, a consumer 102 (where “consumer” refers to a computing device and/or an individual operating the computing device) may wish to install a new home security monitoring system, apply for a new credit card, join a fitness club, or open a credit monitoring account. The consumer may contact the service provider 103 (e.g., a representative or registration system of the home security system, credit card, fitness club, credit monitoring account, etc.), which may require the consumer 102 to be authenticated before accepting his new account opening request. In order to authenticate the consumer, the service provider 103 may send an authentication request to an authentication system 101, which may be configured to determine additional information required from the consumer 102 in order to authenticate the consumer and/or interface with the consumer either directly (via the network 108) or through the service provider 103 in order to authenticate the consumer (or determine that the consumer is not authenticated).


In some embodiments, consumers may submit information needed for authentication to the authentication system 101 such as at the request of the service provider 103.



FIG. 1B is a block diagram illustrating an example system architecture in which multiple consumers 102, 104, 106 are in communication with an authentication processing system 101, wherein the authentication processing system 101 is configured to authenticate the consumer. Depending on the embodiment, the authentication system 101 may include a device identification module 109, an authentication interface module 110, an authentication requirements module 112, and a risk calculation module 122.


According to some embodiments, the device identification module 109 may collect and process received device information. Device identification information may include device attributes, such as a device ID, device location, browser time zone, browser language settings, proxy settings, stated IP address, or real IP address. For example, a consumer 102 may be directed to a web interface to enroll in the home security monitoring service (that requires authentication). During the authentication process, information including internet usage, GPS location, and other device related data may be gathered from the consumer 102's mobile device 104 (e.g., using a script that runs in the background on the device and gathers the information), and the device identification module 109 may analyze such information to determine a location of the consumer 102's computing device. The device identification module 109 may also gather and analyze information that may reveal the consumer 102's real location instead of a location masked by an IP address. For example, if consumer 102's computing device communicates with the authentication system 101 through a Virtual Private Network (“VPN”) or a proxy, then the real location of the consumer 102 may not be the same as an addressed associated with his IP address. The device identification module 109 may determine the real location of consumer 102 by analyzing current GPS data, historical GPS data, and relevant internet traffic indications such as the whether the device's IP address is associated with a known proxy server address.


In some embodiments, the authentication interface module 110 may generate one or more interfaces so that consumers may be asked questions that may be used to verify their identities. In other embodiments, authentication questions are asked of the consumer orally on a phone, for example. Depending on the embodiment, the authentication interface module 110 may also include device-aware parameter-submitting code (such as device id=“NOKIA_ModelXYZ_OperatingSystem_ver1.1” user_agent=“Mozilla/5.0 (Linux; U; OperatingSystem 1.1-update1; en-us; chn-big5) WebKit/525.10+ (KHTML, like Gecko) Version/3.0.4 Mobile Safari/523.12.2”), in the generated user interfaces so that information about consumers' computing devices may be gathered. In some embodiments, the authentication interface module 110 may determine one or more behaviors of the consumer, such as based on the device identification information. For example, the behavior information may indicate a total number of times and/or frequency of a consumer requesting authentication (e.g., from the authentication system 101) during any given timeframe, such as for multiple services for which the authentication processing system 101 provides authentication services, a total time a consumer spends answering authentication questions presented to him or her, or whether the device location matches a location indicated by a browser executing on the device.


In some embodiments, the risk calculation module 122 may calculate a risk score or determine a risk level based on information received from a consumer and/or the consumer's computing or mobile device. For example, if a consumer states that she lives in Oregon but her device IP address indicates she lives in Siberia, then the risk calculation module 122 may increase the risk level (or risk score) associated with this consumer. In another example, if a consumer has approached the authentication system 101 three times (for enrollment in a same service or for enrollment in two or more different services) before but failed all three authentication attempts, the risk calculation module 122 may also increase the risk score or risk level associated with the consumer. Furthermore, in some embodiments, if a consumer's stated address is also associated with several other consumers (and/or consumers that are about the same age and/or have similar demographic information), the risk calculation module 122 may accordingly increase the risk score or risk level associated with the consumer.


According to some embodiments, the authentication requirements module 112 may generate and present various quantities and/or difficulties of questions to consumers, such as based on risk information generated by the risk calculation module 122. In addition, other criteria may be used for generating questions of various quantities or difficulty. For example, the risk calculation module 122 and/or the authentication requirements module 112 may adjust the authentication requirements for a consumer in response to determining that the consumer has taken a longer than normal time to answer authentication questions. Thus, during a single authentication session (e.g., a set of questions configured to authenticate a consumer), authentication requirements may be adjusted, such as by presenting more difficult (or less difficult) questions and/or additional (or fewer) questions. Additionally, risk score information may be stored and used in a later authentication request for a consumer or device, such as to adjust the authentication requirements.


Depending on the embodiment, the authentication system 101 may also include several data structures that store authentication related data. For example the authentication system 101 may include a risk factors data structure 114, which may store data and/or information related to potential risk factors, risk scores, and/or risk levels associated with respective consumers. The authentication system 101 may also include a device information data structure 116, which may in some embodiments store device information, location information associated with devices, IP addresses, MAC addresses, browser settings, proxy settings, and so forth. Depending on the embodiment, the authentication system 101 may also include an authentication data structure 118. The authentication data structure 118 may store information about consumers, computing devices, and so forth that may be used to generate authentication questions, which may be presented to consumers.


As illustrated in FIG. 1, the network 108 may comprise one or more of a LAN, WAN, and/or the Internet. The network 108 communicates with various modules, computing devices, and/or other electronic devices via wired or wireless communication links, and/or combination of wired and wireless communication links.


Example Computing System



FIG. 2 is a block diagram illustrating example components of the authentication system 101. While the components illustrated in FIG. 2 are discussed with reference to the authentication system 101, any other computing devices discussed and/or illustrated herein may include any or all of these same or similar components.


The authentication system 101 includes, for example, a computer that may be IBM, Macintosh, or Linux/Unix compatible or a server or workstation. In one embodiment, the authentication system 101 comprises a server, desktop computer, a tablet computer, or laptop computer, for example. In one embodiment, the exemplary authentication system 101 includes one or more central processing units (“CPUs”) 215, which may each include a conventional or proprietary microprocessor. The authentication system 101 further includes one or more memory 225, such as random access memory (“RAM”) for temporary storage of information, one or more read only memory (“ROM”) for permanent storage of information, and one or more mass storage device 205, such as a hard drive, diskette, solid state drive, or optical media storage device. Typically, the modules of the authentication system 101 are connected to the computer using a standard based bus system 230. In different embodiments, the standard based bus system could be implemented in Peripheral Component Interconnect (“PCI”), Microchannel, Small Computer System Interface (“SCSI”), Industrial Standard Architecture (“ISA”) and Extended ISA (“EISA”) architectures, for example. In addition, the functionality provided for in the components and modules of authentication system 101 may be combined into fewer components and modules or further separated into additional components and modules.


The authentication system 101 is generally controlled and coordinated by operating system software, such as Windows XP, Windows Vista, Windows 7, Windows 8, Windows Server, Unix, Linux, SunOS, Solaris, or other compatible operating systems. In Macintosh systems, the operating system may be any available operating system, such as MAC OS X. In other embodiments, the authentication system 101 may be controlled by a proprietary operating system. Conventional operating systems control and schedule computer processes for execution, perform memory management, provide file system, networking, I/O services, and provide a user interface, such as a graphical user interface (“GUI”), among other things.


The exemplary authentication system 101 may include one or more commonly available input/output (I/O) devices and interfaces 220, such as a keyboard, mouse, touchpad, and printer. In one embodiment, the I/O devices and interfaces 220 include one or more display devices, such as a monitor, that allows the visual presentation of data to a user. More particularly, a display device provides for the presentation of GUIs, application software data, and multimedia presentations, for example. The authentication system 101 may also include one or more multimedia devices, such as speakers, video cards, graphics accelerators, and microphones, for example.


In the embodiment of FIG. 2, the I/O devices and interfaces 220 provide a communication interface to various external devices. In the embodiment of FIG. 2, for example, the authentication system 101 also includes the authentication interface module 110, the device identification module 109, the authentication requirements module 112, and the risk calculation module 122 (discussed above with reference to FIG. 1B) that may be stored in the mass storage device 205 and/or memory 225 as executable software codes that are executed by the CPU 215. These modules may include, by way of example, components, such as software components, object-oriented software components, class components and task components, processes, functions, attributes, procedures, subroutines, segments of program code, drivers, firmware, microcode, circuitry, data, databases, data structures, tables, arrays, and variables.


In general, the word “module,” as used herein, refers to logic embodied in hardware or firmware, or to a collection of software instructions, possibly having entry and exit points, written in a programming language, such as, for example, Java, Lua, C or C++. A software module may be compiled and linked into an executable program, installed in a dynamic link library, or may be written in an interpreted programming language such as, for example, BASIC, Perl, or Python. It will be appreciated that software modules may be callable from other modules or from themselves, and/or may be invoked in response to detected events or interrupts. Software modules configured for execution on computing devices may be provided on a computer readable medium, such as a compact disc, digital video disc, flash drive, or any other tangible medium. Such software code may be stored, partially or fully, on a memory device of the executing computing device, such as the computing system 103, for execution by the computing device. Software instructions may be embedded in firmware, such as an EPROM. It will be further appreciated that hardware modules may be comprised of connected logic units, such as gates and flip-flops, and/or may be comprised of programmable units, such as programmable gate arrays or processors. The modules described herein are preferably implemented as software modules, but may be represented in hardware or firmware. Generally, the modules described herein refer to logical modules that may be combined with other modules or divided into sub-modules despite their physical organization or storage.


In some embodiments, one or more computing systems, data stores and/or modules described herein may be implemented using one or more open source projects or other existing platforms. For example, one or more computing systems, data stores and/or modules described herein may be implemented in part by leveraging technology associated with one or more of the following: Drools, Hibernate, JBoss, Kettle, Spring Framework, NoSQL (such as the database software implemented by MongoDB) and/or DB2 database software.


Example Method of Adjusting Knowledge-Based Authentication (“KBA”)



FIG. 3 is a flowchart that illustrates one embodiment of a method of determining a difficulty level of authentication for a particular consumer. The method of FIG. 3 may be performed by the authentication system 101, such as the various modules of the authentication system 101. Alternatively, the method may be performed, in whole or part, by another computer device, such as partially by a consumer device and partially by the authentication system 101. For ease of description, the method will be discussed herein as being performed by the authentication system 101, but such description does not preclude performance of the method by multiple devices. Depending on the embodiment, the method may include fewer or additional blocks and/or the blocks may be performed in an order that is different than illustrated.


The method 300 begins at block 305, where a request for opening a new account is received. The new account may be an account of any type. For example, in some embodiments, the new account may be a credit line, satellite TV service, college admission, credit monitoring, or any other. In one embodiment, the particular consumer is anonymous to the authentication system 101 when the authentication request is received. For example, because the consumer is requesting a new account, the service provider likely doesn't already have identification information of the consumer.


In some embodiments, the request for opening a new account may come from a third party, such as a service provider, rather than directly from a consumer. For example, a retail store may initiate authentication of a consumer with the authentication system 100 in order to determine whether a new store credit account for the consumer that purchased some merchandise at the retail store should be opened. Similarly, an employer may send a request to an insurance company to request a new account be set up for one of its employees and, in response to the request, the insurance company may initiate authentication of the employee with the authentication system 100. In some other embodiments, the request for opening a new account may come directly from a consumer or a group of consumers. For example, a consumer may request authentication, which may then be provided to another entity (e.g., via an authentication token) as proof that authentication has been performed.


In some embodiments, a request for opening a new account may be sent via a post card, by mail, via a phone call, or other means besides on-line submission. For example, a consumer may submit paperwork needed to open a new account in a letter. A consumer may also initiate the request by a postcard and/or through a phone call. For example, the consumer may call into a call center and request to open a new account. A representative at the call center may enter information provided by the consumer via a use interface similar to ones disclosed herein to provide details regarding the consumer and the consumer's new account request.


The received request to open a new account may be received directly by a service provider, which may perform authentication itself. In some embodiments, the received request to open a new account may be received by a service provider and then forwarded to the authentication system 101. The authentication system 100 may then be put in direct contact with the consumer to complete the authentication, such as via a phone call, web form, email, and/or any other communication medium, and/or the authentication system 100 communicates with the consumer via the service provider, such as via a Iframe embedded in a website of the service provider.


The method 300 proceeds to block 310, where identification information regarding a consumer's computing device or mobile communication device is received. In some embodiments, the received information may include one or more IP addresses (real IP address, stated IP address, and so forth), device type, device manufacturer, device serial number, and so forth. For example, the information may be obtained by the device identification module 109 by embedding a script sent to a consumer's device and analyzing the information gathered from the script. For example, information received from the embedded script may include: “device name=MyPC, model name=PC123, Operating System=OS1.0, Application=MyBrowser, TouchScreenDevice=No, PhysicalScreenWidth=1024×768, IP address=100.100.100.100.”


The method 300 proceeds to block 315, where one or more behaviors of the consumer may be derived based at least on the device identification information received. Such information may be used by the device identification module 109 to identify attributes of the device and/or consumer that may warrant an increased or decreased risk level associated with the consumer and, thus, an increased or decreased difficulty level associated with authentication of the consumer. For example, any inconsistencies, such as discrepancy between a real IP address and a stated IP address, may indicate an increased risk, which may lead to a greater quantity and/or difficulty level of authentication questions asked of the consumer. In some embodiments, other behaviors of the consumer may also include having already used the same device to attempt authentication unsuccessfully before. Depending on the embodiment, the behaviors of the consumer may be stored in a device information data structure 116.


The method 300 proceeds to block 320, where a risk of fraud associated with the consumer may be calculated based on the one or more behaviors of the consumer. For example, if a consumer's device has a stated IP address that is in Virginia, USA, but the device's unmasked real IP address revealed through software tools indicate that the device is in Thailand, then the risk calculation module 122 may use such information in the calculation of the risk of fraud involved with the consumer.


In some embodiments, if a consumer has already attempted authentication several times unsuccessfully, the risk calculation module 122 may also use such information as a factor in calculating the risk of fraud involved with the consumer.


Depending on the embodiment, the risk of fraud associated with the consumer may be represented as a risk score, which could be a numerical value. The risk of fraud may also be represented as a risk level, which may or may not be a numerical value.


The method 300 then proceeds to block 325, where a quantity and/or difficulty level of authentication questions may be determined and/or adjusted (e.g., from a default level) based on the calculated risk of fraud. In some embodiments, the adjustment is done by the authentication requirements module 112 using information stored in the risk factors data structure 114 and authentication data structure 118.


For example, the authentication system 101 may present questions of low levels of difficulty in response to a determination that the consumer has a low risk of fraud or a low risk score. The questions presented to the consumer may include basic questions that help verify the consumer's identity.


If, however, the consumer has a high risk of fraud or a high risk score, the questions presented to the consumer may include more difficult questions and/or more questions. For example, the authentication requirements module 112 may use information stored in the authentication data structure 118 to generate more difficult questions such as “Please name at least 2 public libraries next to your residence” and/or “Which of the following banks is one that you do not have an account with,” and so forth. In one embodiment, the authentication questions are determined based on credit information of the consumer. For example, information regarding previous addresses, mortgage payments, credit card balance amounts, items purchased, etc., that typically no one other than the individual identified in the credit report (and possibly a spouse and a few other trusted individuals) would know about. Thus, asking questions about this information can help authenticate that the consumer really is who they say they are, in view of them knowing the information that is included on the credit report of the person that they are claiming to be. In other embodiments, various other sources of authentication questions may be used in order to ask questions of the consumer in an attempt to confirm that the consumer really is who they say they are.


In some embodiments, difficulty level of questions presented to a consumer may be dynamically adjusted. For example, if the authentication system 101 detects that the consumer has spent 5 minutes (or some other predetermined time) trying to answer a specific question, which may be considered too long by the authentication system 101, the authentication system 101 may generate an additional and/or a more difficult question for the consumer to answer. For example, if the consumer takes five minutes to provide his/her birthday (or other information that the consumer should be able to readily provide), the risk associated with the consumer may be increased and additional questions provided to that consumer may be of an increased difficulty and/or quantity.


Dealing with High Risk Consumers


In some embodiments, the system may automatically choose and present, based at least in part on difficulty level of questions, trick questions (also referred to as “false questions”) to the consumers. For example, the false question may be: “What is the name of your daughter's first elementary school” when the system, in fact, knows that the identify used in requesting a new account does not have a daughter. In another example, the question may be: “Please indicate which branch of the ABCD Bank is the one that you visit most frequently” when in fact the system is aware that the identity used in requesting a new account does not have an account at the ABCD Bank.


In addition to false questions, the system may also provide false answers. For example, the system may mix real answers and false answers and present the mixed set of answers to both real and false questions to a consumer. In some embodiments, the false questions and false answers are generated randomly and automatically by the system. In some other embodiments, the false questions may also be ranked as to their respective levels of difficulty and presented to consumers by the system based on the chosen level of difficulty. For example, if a consumer is deemed high risk, the system may choose more difficult false questions and present them to the consumer. Examples of relatively easy authentication questions may be “What is your date of birth?” and “What is your mother's maiden name?” Examples of more difficult authentication questions include, for example, “What is your address from 1999-2001?” and “Did you have a car loan for approximately $40,000 in 2005 with CarBuyingCo?”


Depending on the embodiment, in response to the authentication system determining that a consumer is a high risk for fraud, the system may provide automatically generated error messages to the user. For example, instead of telling a user that he provided a wrong password but his email address was correct, which may risk giving a high-risk consumer too much information, the system may instead provide a message such as “Your information does not match our records,” “Your mailing address is wrong,” “Your age does not match our records,” or even “the system has experienced serious problems, please try again later,” etc. These intentionally false random error messages provide as little information as possible to a potentially fraudulent situation and do not aid the person who might be committing fraud in correcting any information and resubmitting a request.


In some embodiments, a combination of false questions and automatic random error messages may be used and no real question is presented to a consumer who has a particularly high risk of fraud. In some other embodiments, real questions, false questions, and automatic random error messages may all be used at the same time and presented to a user who was considered as a high risk of fraud. In some other embodiments, real questions, false questions, and automatic random error messages may be configured to have difficulty levels and presented to a consumer based on the consumer's level of risk for fraud. Further examples of dealing with high risk consumers are discussed below with reference to FIG. 7-9.


In some embodiments, besides adjusting the difficulty levels of questions, the authentication system 101 may also generate more questions that a consumer may have to answer correctly before the consumer can be authenticated.


Example User Interfaces



FIG. 4 is an illustrative user interface 400 that may be generated and presented to a consumer or a person who is tasked with entering information related to a consumer, which includes questions presented to the consumer for the purpose of opening a new account and also tools that gather browser and/or computing device related information. The authentication system 101 and/or the authentication interface module 110 may be used to implement systems and methods described in this disclosure.


In this embodiment, the user interface 400 gathers information from a consumer or a person responsible for entering information related to a consumer so that the consumer may start the process of authentication. The information gathered may include: user name 405, first name 410, last name 415, street address 410, city 425, state 430, zip code 435, and/or account type 430, which may be personal or business.


In some embodiments, the person responsible for entering information related to a consumer may be a representative of a call center who helps enter information for opening credit accounts. In some other embodiments, the person responsible for entering information related to a consumer may be a representative of a credit bureau, a retailer, a hospital, an insurance company, a fitness club, a potential employer, etc., who helps enter information for opening a new account.


In this example, the user interface 400 is presented in a browser. The authentication system 101, via the device identification module 109, may receive additional metadata and/or device related information. For example, information may be gathered when the user types in the text fields and/or when the user submits the information. The information may include, for example, browser type, geographic location, reported location, proxy settings, and so forth.


In some embodiments, if the browser type is a mobile phone browser, but the device information associated with the computing device indicates that the consumer uses a desktop computer, the risk calculation module 122 may use such information to increase the level of risk or increase a risk score associate with the consumer.


In addition, as discussed previously, if the computing device of the consumer has a real IP address discovered by the device identification module 109, but the consumer's stated IP address does not match the real IP address, then the risk calculation module 122 may also use such information to increase the level of risk or increase a risk score associate with the consumer.



FIG. 5 is an illustrative mobile user interface 500 that may be generated and presented to a mobile consumer, which includes questions presented to the consumer for the purpose of opening a new account and also tools that gather mobile device and related information. The authentication system 101 and/or the authentication interface module 110 may be used to implement systems and methods described in this disclosure. The term “mobile user interface” here may include user interface generated and presented to a user of a mobile computing device, such as laptop, mobile phone, and/or tablets, and so forth.


The user interface 500 may include questions for a consumer so that the consumer may start the process of authentication. The information gathered may include: user name, first name, last name, street address, city, state, zip code, and/or account type, which may be personal or business.


As shown in FIG. 5, information regarding the mobile computing device's device type, device ID, location, reported location, and so forth, may be gathered by an authentication system 101 in order to generate authentication questions. Any discrepancies between a location of the mobile computing device and its reported location may trigger a higher risk score or risk level, so that in some embodiments, more difficult authentication questions may be presented to the consumer.


Example Method of Generating a Risk Score



FIG. 6 is a flowchart that illustrates one embodiment of determining a risk score and using the risk score to determine whether a consumer's fraud risk is so high that the system will refuse to authenticate the consumer. The method of FIG. 6 may be performed by the authentication system 101 and/or via any other suitable computing system. Depending on the embodiment, the method may include fewer or additional blocks and/or the blocks may be performed in order different than is illustrated.


The method 600 begins at decision block 605, wherein the system determines whether the consumer's IP address is masked. If the consumer has a masked IP address and/or the masked IP address's geolocation does not match the consumer's real geolocation, the method 600 proceeds to block 610 and increase the risk score of the consumer. For example, if a person uses a proxy service that is available for free or for a fee (so that the person can connect to a proxy server first and direct all internet traffic through the proxy server), his or her real IP address would be hidden. Instead, the proxy server's IP address will be show. The device identification module 109 may determine that the device's purported address is really just a proxy server's IP address by looking up the purported IP address in a data store of known proxy server addresses. If so, the device identification module 109 may be configured to automatically notify the risk calculation module 122 so the risk level associated with the device can be adjusted. Moreover, if the proxy server's IP address is associated with a known malicious group for identity theft or other related crimes or violations, the risk calculation module 122 may adjust the risk level accordingly. However, if the proxy server's IP address is associated with a known cloud-service provider that provides virtual computing capacities to companies or a known employer which masks all of their employee's IP addresses, the device identification module 109 may determine that the device is affiliated with a known organization with low risk. The risk calculation module 122 may then adjust the risk level accordingly.


If the consumer does not have a masked IP address, the method 600 may proceed to block 615 and determine how many times the consumer has attempted authentication before. In some embodiments, the system may derive the information based on information in a data store. In some other embodiments, the system may derive the information based on real-time and/or number of attempts in a recent session. Alternatively, the system may also count the number of times the consumer has attempted authentication in a given period of time, e.g., in the last two months, the last week, the last year, etc., such as based on authentication requests from various other service providers.


The method 600 proceeds to block 620, wherein the consumer's risk score is increased based on the number of attempts. For example, the system may increase the risk score of the consumer by one for each unsuccessful attempt. In some embodiments, the system may assign different weight based on how recent the failed authentication attempt was. For example, a very recent attempt would have a weight higher than a more distant failed attempt to authenticate. Further, in some embodiments, the system may use other formulae to increase the risk score based on the number of past attempts to authenticate. The formulae used may be, for example, based on a risk of fraud analysis modeled using normal distribution, binomial distribution, etc.


In some embodiments, the method 600 proceeds to decision block 630 wherein it is determined whether the risk score is over a pre-determined high risk threshold. For example, if the risk score of the consumer is higher than the pre-determined high risk threshold, the risk that the consumer is committing authentication fraud may be deemed too high to be acceptable by the system and/or an entity requesting authentication (e.g., each entity requesting authentication may require different levels of authentication). In such situations, the method 600 proceeds to block 635, wherein the risk information is communicated to a requesting entity, such as to deny opening of an account.


However, if the risk score of the consumer is not so high as to be more than the high risk threshold as determined in the block 630, the method 600 proceed to block 640, wherein the difficulty level of the authentication questions are determine/adjusted based on the determined risk score. Once the authentication question difficulty is determined (e.g., block 640), the authentication questions may be asked of the consumer, via direct communication between the authentication system and the consumer and/or communication between the authentication system and the consumer via a service provider. In some embodiments, the method of FIG. 6, or certain blocks of FIG. 6, may be performed as authentication questions are being provided to the consumer, such that at block 648 previously determined authentication question difficulty level or quantity may be adjusted based on previous responses, device identification information, and/or determine behaviors of the consumer.


Example Method of Dead-End Authentication and Collecting Information from High Fraud Risk Consumers



FIG. 7 is a flowchart that illustrates one embodiment of tracking behavior of a user with high fraud risk and leading the user to a dead-end authentication process. The method of FIG. 7 may be performed by the authentication system 101 and/or via any other suitable computing system. Depending on the embodiment, the method may include fewer or additional blocks and/or the blocks may be performed in order different than is illustrated.


The method 700 begins at block 705, wherein the authentication system 101, through a user interface or an operator at a call center, requests a consumer who is deemed to be involved with a high risk of fraud, to do one or more of the following: create a password, provide an address, birth date, Social Security Number (SSN), and so forth. The method 700 then proceeds to block 715, wherein the authentication system 101 searches one or more databases using the information provided by the consumer, including searching the database for the same password, address, birthday, SSN, and so forth. The method 700 may store the information provided by the user and in particular, record if any hits or results were returned based on the results of the search. Such information may be used by the authentication system 101 to identify the same high risk user again in the future.


The method 700 proceeds to block 720, wherein the authentication system 101 requests the consumer to answer a combination of easy and difficult questions for purposes of collecting more data from the consumer. The questions asked by the authentication system 101 may be similar to typical authentication questions used by the system so that the consumer does not suspect that these questions were asked for purposes of collecting information from him/her. An embodiment of an easy question may be “What is your mother's maiden name?” An embodiment of a more difficult question may include “What is the name of the city you live in when you were 21 years old?”


The method 700 then proceeds to block 730, wherein the authentication system 101 lets the consumer continue on the dead-end authentication process, regardless of correctness of the user's answers. The authentication system 101 collects the consumer's answers, information, and information about the consumer's device. The authentication system 101 may do this in order to find out more about the potentially high risk consumer and better identify characteristics of such high risk consumers in the future.


The method 700 then proceeds to block 740, wherein the authentication system 101 determines whether the user should be kept in the dead-end authentication process any longer. Depending on the embodiment, the authentication system 101 may use a variety of criteria to make this decision. For example, the authentication system 101 may determine that enough information about the consumer and the consumer's device has been collected so that the consumer should not be kept in the dead-end authentication process anymore.


In some other embodiments, the authentication system 101 may determine that the consumer has not answered enough questions about his or her physical location or past credit history. Therefore, the authentication system 101 may decide to keep the consumer in the dead-end authentication process until the authentication system 101 has collected enough physical location or past credit history information.


If the answer to the question in decision block 740 is yes (that is, the consumer should be kept in the dead-end authentication process), the method 700 proceeds back to block 720, wherein the system again prompts the user to answer questions for data mining and data collection purposes.


However, if the answer to the question in decision block 740 is no (that is, the consumer does not need to be kept in the dead-end authentication process), the method 700 proceeds to block 750, wherein the authentication system 101 stores the collected information and usage patterns of the consumer in order to analyze the consumer's behavior. For example, if the password provided by this consumer is a non-trivial combination of characters, numbers, and symbols, such passwords may be used to detect whether this unusual password is reported as used in other instances. If a match is found, the authentication system 101 may submit such information for further analysis to determine whether the other instances involve a high risk of fraud as well.



FIG. 8A is an illustrative user interface 800 that may be generated and presented to a consumer that is determined to be involved in potentially fraudulent authentication. In this example, the user interface 800 is presented in a browser on a consumer's mobile computing device. However, the user interface may also be presented in an application or other types of user interfaces that can be used on a consumer's laptop, desktop, tablet, and other mobile computing devices. The authentication system 101, via the authentication interface module 112, may generate and present this user interface to a consumer who is may be involved with a high fraud risk. The user interface 800 may generate a message that communicates to the consumer that the two passwords entered by the consumer during the authentication process do not match, when in fact the passwords matched each other. The user interface 800 may request the consumer to enter the password, re-confirm it, and provide an alternative email. A consumer who is involved with a high risk of fraud may use another password that is different from the ones previously entered. The authentication system 101 may collect the additional passwords in order to maximize the amount of information that the system can receive from this potentially high risk consumer for data mining and analysis purposes.


Similarly, FIG. 8B is an illustrative user interface 850 that may be generated and presented to a user who is determined to be involved in potentially fraudulent authentications for purposes of capturing information from the user. As in FIG. 8B, the authentication system 101 may request the user to provide his or her address, regardless of whether the information previously provided by the user was correct or not. Such address information collected from the high fraud risk consumer may be used for risk analysis purposes.



FIG. 8C is an illustrative user interface 870 that may be generated and presented to a user who is determined to be involved in potentially fraudulent authentication so that at the end of authentication, the user is not provided with any hints regarding how to succeed in future fraudulent attempt to authenticate or any hints that the user has stepped into a honeypot. At the end of the authentication process, the authentication system 101 may provide a random error message so that a consumer with a potentially high risk of fraud may receive the random error message. The consumer, receiving a message such as “server error,” may not suspect that he or she has been the target of a fraud analysis.


Another benefit of providing such a message using the illustrative user interface 870 is that the authentication system 101 will not let the consumer know what part of his or her answer triggered the system to detect his or her potential fraud. Minimizing such identifying information to the consumer may help minimize risk of providing useful information to a consumer with high risk of fraud so that the consumer is likely to find a way to get around the authentication system 101 in the future.


The illustrative user interface 870 may also provide a link back to a special version of the user interface so that if the consumer wants to be authenticated again, the consumer will be kept in the special version of the user interface for data collection and analysis process. The special version of the user interface may help ensure that the consumer or a device associated with the consumer's IP address can be quarantined in the dead-end authentication process should the consumer attempts authentication again. This may help minimize the risk that the consumer may get around the authentication system 101 in the future as well.



FIG. 9 is a flowchart that illustrates one embodiment of alerting a consumer that potentially fraudulent authentication was attempted and storing information collected from the authentication session. The method 900 starts at block 905, wherein the authentication system 101 determines how many times the consumer has attempted authentication. In some embodiments, instead of only determining a total amount of times attempted, the system may also determine the frequency of such attempts within a period of time.


The method 900 proceeds to decision block 910, wherein the authentication system 101 determines whether the number of times authentication has been attempted by the consumer exceeds a first threshold. If the answer to the question in decision block 910 is no, then the method 900 proceeds to block 945 and ends without alerting a potential victim of fraud because the number of attempts have not arisen to a level that is above the first threshold.


If the answer to the question in decision block 910 is yes, that is, the number of attempted authentications has exceeded the first threshold, then the method 900 proceeds to block 920, wherein the authentication system 101 sends an alert to the victim (e.g., the consumer whose identity is included in the attempted authentication) and let the victim know that someone has tried to open a new account using the victim's identity.


The method 900 then proceeds to decision block 930, wherein the authentication system 101 determines whether the number of times authentication has been attempted by the consumer exceeds a second threshold, which is higher than the first threshold. If the answer to the question in decision block 930 is no, then the method 900 proceeds to block 945 and ends without sending a second alert to a potential victim of fraud because the number of attempts have not arisen to a level that is above the second threshold.


However, if the answer to the question in decision block 910 is yes, that is, the number of attempted authentications has exceeded the second threshold, then the method 900 proceeds to block 940, wherein the authentication system 101 locks new account opening opportunities involved with identity of the potential fraud victim. Depending on the risk involved and the embodiment, the period of time that the victim's account is locked may vary. Once the victim's account is locked, no new account may be opened using the identity of the victim until the lock is released.


The method 900 then proceeds to block 950, wherein the authentication system 101 stores the statistics related to authentication attempts. The statistics may help adjust the first and second threshold used by the authentication system 101 in the future.


OTHER EMBODIMENTS

Although the foregoing systems and methods have been described in terms of certain embodiments, other embodiments will be apparent to those of ordinary skill in the art from the disclosure herein. Additionally, other combinations, omissions, substitutions and modifications will be apparent to the skilled artisan in view of the disclosure herein. While some embodiments of the inventions have been described, these embodiments have been presented by way of example only, and are not intended to limit the scope of the inventions. Indeed, the novel methods and systems described herein may be embodied in a variety of other forms without departing from the spirit thereof. Further, the disclosure herein of any particular feature, aspect, method, property, characteristic, quality, attribute, element, or the like in connection with an embodiment can be used in all other embodiments set forth herein.


All of the processes described herein may be embodied in, and fully automated via, software code modules executed by one or more general purpose computers or processors. The code modules may be stored in any type of computer-readable medium or other computer storage device. Some or all the methods may alternatively be embodied in specialized computer hardware. In addition, the components referred to herein may be implemented in hardware, software, firmware or a combination thereof.


Conditional language such as, among others, “can,” “could,” “might” or “may,” unless specifically stated otherwise, are otherwise understood within the context as used in general to convey that certain embodiments include, while other embodiments do not include, certain features, elements and/or steps. Thus, such conditional language is not generally intended to imply that features, elements and/or steps are in any way required for one or more embodiments or that one or more embodiments necessarily include logic for deciding, with or without user input or prompting, whether these features, elements and/or steps are included or are to be performed in any particular embodiment.


Any process descriptions, elements or blocks in the flow diagrams described herein and/or depicted in the attached figures should be understood as potentially representing modules, segments, or portions of code which include one or more executable instructions for implementing specific logical functions or elements in the process. Alternate implementations are included within the scope of the embodiments described herein in which elements or functions may be deleted, executed out of order from that shown, or discussed, including substantially concurrently or in reverse order, depending on the functionality involved as would be understood by those skilled in the art.

Claims
  • 1. A system comprising: one or more processors; anda non-transitory computer-readable storage medium storing computer-executable instructions that cause the one or more processors to: receive an indication of a request initiated by a computing device for a financial service provided by a financial service provider;receive device information associated with the computing device gathered through device information gathering code executed by the computing device, wherein the device information includes at least one or more of: a device identifier for the computing device,device location,browser type,browser time zone,browser language settings,proxy settings,stated IP address,real IP address,current GPS location data, orhistorical GPS location data;determine a fraud risk for a consumer, wherein an identity of the consumer is unknown to the financial service provider, wherein the fraud risk is based on one or more potential fraud behaviors derived from at least the device information; andin response to the fraud risk meeting a threshold, transmit to the computing device an authentication question set for display in one or more user interfaces by the computing device, wherein the authentication question set comprises one or more authentication questions, and wherein the one or more authentication questions are selected based at least in part on the fraud risk.
  • 2. The system of claim 1, wherein the one or more potential fraud behaviors includes an identified anomaly between the device information and predetermined information in a profile associated with the consumer.
  • 3. The system of claim 1, wherein the one or more potential fraud behaviors comprises one or more of: frequent attempts to open a new account, requesting a new account from more than one computing device, having a device IP address that matches a known IP addresses associated with a high risk of fraud, or submitting a password that matches a unique password known to be related to a high risk of fraud.
  • 4. The system of claim 1, wherein the request for a financial service includes a request to open a new financial account.
  • 5. The system of claim 1, wherein the one or more potential fraud behaviors of the consumer includes at least one of: a total number of times authentication was requested for the consumer or a frequency at which authentication was requested for the consumer.
  • 6. The system of claim 1, wherein the device information gathering code includes embedded instructions sent to the computing device to collect the device information, and to process information collected from the embedded instructions.
  • 7. The system of claim 1, wherein the computer-executable instructions further cause the one or more processors to, based at least in part on the fraud risk of the consumer, send a request to the computing device for further device information.
  • 8. The system of claim 1, wherein the computer-executable instructions further cause the one or more processors to: in response to a determination that the consumer is not authenticated, transmit an alert to a victim of authentication fraud.
  • 9. The system of claim 1, wherein the computer-executable instructions further cause the one or more processors to: in response to a determination that the consumer is not authenticated, perform a lock on an account associated with a victim of authentication fraud.
  • 10. The system of claim 1, wherein the computer-executable instructions further cause the one or more processors to determine whether the consumer is authenticated based in part on a comparison between location information of the consumer with location information from the device information.
  • 11. A computer-implemented authentication method comprising: receiving indication of a request initiated by a computing device for a financial service provided by a financial service provider;receiving, from a computing device, device information associated with the computing device gathered through device information gathering code executed by the computing device, wherein the device information includes at least one or more of: a device identifier for the computing device,device location,browser type,browser time zone,browser language settings,proxy settings,stated IP address,real IP address,current GPS location data, orhistorical GPS location data;determining a fraud risk for a consumer, wherein an identity of the consumer is unknown to the financial service provider, wherein the fraud risk is based on one or more potential fraud behaviors derived from at least the device information; andin response to the fraud risk meeting a threshold, causing the computing device to present one or more user interfaces that include an authentication question set comprising one or more authentication questions based on the fraud risk.
  • 12. The method of claim 11, wherein the authentication question set is generated based on a level of difficulty corresponding to the fraud risk.
  • 13. The method of claim 11, wherein the fraud risk comprises a risk score.
  • 14. The method of claim 11, wherein the one or more user interfaces provide a false indication of a correctness of a consumer's response to the authentication question set.
  • 15. The method of claim 11, wherein the one or more user interfaces does not indicate whether a consumer's response to the authentication question set was correct.
  • 16. The method of claim 11, wherein the authentication question set includes at least one question that includes false information within the question.
  • 17. A non-transitory computer storage medium storing computer-executable instructions that, when executed by a processor, cause the processor to perform operations comprising: receiving, from a computing device, a request for a financial service provided by a financial service provider;receiving, from the computing device, device information associated with the computing device gathered through device information gathering code, wherein the device information includes at least one or more of: a device identifier for the computing device,device location,browser type,browser time zone,browser language settings,proxy settings,stated IP address,real IP address,current GPS location data; orhistorical GPS location data;determining a fraud risk for a consumer, wherein an identity of the consumer is unknown to the financial service provider, wherein the fraud risk is based on one or more potential fraud behaviors derived from at least the device information; andin response to the fraud risk meeting a threshold, transmitting to the computing device an authentication question set for display in one or more user interfaces, wherein the authentication question set comprises one or more authentication questions based on the fraud risk.
  • 18. The non-transitory computer storage medium of claim 17, wherein the operations further include: dynamically adjusting the authentication question set based on a user response to at least a first question.
  • 19. The non-transitory computer storage medium of claim 17, wherein the fraud risk is based on at least two or more of: the device location, the browser type, the browser time zone, the browser language settings, the proxy settings, the stated IP address, the real IP address, the current GPS location data, or the historical GPS location data of the computing device.
CROSS-REFERENCE TO RELATED APPLICATIONS

This application is a continuation of U.S. application Ser. No. 15/459,927, filed on Mar. 15, 2017, which is a continuation of U.S. application Ser. No. 14/208,595, filed on Mar. 13, 2014, which claims the benefit of priority under 35 U.S.C. § 119(e) of U.S. Provisional Application No. 61/790,321, filed Mar. 15, 2013, the disclosures of which are hereby incorporated by reference in their entireties.

US Referenced Citations (1089)
Number Name Date Kind
3752904 Waterbury Aug 1973 A
4795890 Goldman Jan 1989 A
4891503 Jewell Jan 1990 A
4977595 Ohta et al. Dec 1990 A
4989141 Lyons et al. Jan 1991 A
5126936 Champion et al. Jun 1992 A
5351293 Michener et al. Sep 1994 A
5590038 Pitroda Dec 1996 A
5640577 Scharmer Jun 1997 A
5659725 Levy et al. Aug 1997 A
5659731 Gustafson Aug 1997 A
5715314 Payne et al. Feb 1998 A
5719941 Swift et al. Feb 1998 A
5745654 Titan Apr 1998 A
5748098 Grace May 1998 A
5754632 Smith May 1998 A
5819226 Gopinathan et al. Oct 1998 A
5832068 Smith Nov 1998 A
5844218 Kawan et al. Dec 1998 A
5881131 Farris et al. Mar 1999 A
5884289 Anderson et al. Mar 1999 A
5903830 Joao et al. May 1999 A
5913196 Talmor et al. Jun 1999 A
5956693 Geerlings Sep 1999 A
5966695 Melchione et al. Oct 1999 A
5999596 Walker et al. Dec 1999 A
6021397 Jones et al. Feb 2000 A
6021943 Chastain Feb 2000 A
6026440 Shrader et al. Feb 2000 A
6029154 Pettitt Feb 2000 A
6038551 Barlow et al. Mar 2000 A
6069941 Byrd et al. May 2000 A
6072894 Payne Jun 2000 A
6073106 Rozen et al. Jun 2000 A
6073140 Morgan et al. Jun 2000 A
6085242 Chandra Jul 2000 A
6094643 Anderson et al. Jul 2000 A
6119103 Basch et al. Sep 2000 A
6128602 Northington et al. Oct 2000 A
6157707 Baulier et al. Dec 2000 A
6161139 Win et al. Dec 2000 A
6182068 Culliss Jan 2001 B1
6182229 Nielsen Jan 2001 B1
6196460 Shin Mar 2001 B1
6233588 Marchoili et al. May 2001 B1
6247000 Hawkins et al. Jun 2001 B1
6253202 Gilmour Jun 2001 B1
6254000 Degen et al. Jul 2001 B1
6263447 French et al. Jul 2001 B1
6269369 Robertson Jul 2001 B1
6282658 French et al. Aug 2001 B2
6292795 Peters et al. Sep 2001 B1
6311169 Duhon Oct 2001 B2
6321339 French et al. Nov 2001 B1
6327578 Linehan Dec 2001 B1
6330546 Gopinathan et al. Dec 2001 B1
6343279 Bissonette et al. Jan 2002 B1
6356937 Montville et al. Mar 2002 B1
6397212 Biffar May 2002 B1
6418436 Degen et al. Jul 2002 B1
6453353 Win et al. Sep 2002 B1
6457012 Jatkowski Sep 2002 B1
6463533 Calamera et al. Oct 2002 B1
6473740 Cockrill et al. Oct 2002 B2
6496936 French et al. Dec 2002 B1
6510415 Talmor et al. Jan 2003 B1
6513018 Culhane Jan 2003 B1
6523021 Monberg et al. Feb 2003 B1
6523041 Morgan et al. Feb 2003 B1
6539377 Culliss Mar 2003 B1
6564210 Korda et al. May 2003 B1
6574736 Andrews Jun 2003 B1
6581059 Barrett et al. Jun 2003 B1
6597775 Lawyer et al. Jul 2003 B2
6601173 Mohler Jul 2003 B1
6607136 Atsmon et al. Aug 2003 B1
6615193 Kingdon et al. Sep 2003 B1
6622131 Brown et al. Sep 2003 B1
6629245 Stone et al. Sep 2003 B1
6647383 August et al. Nov 2003 B1
6658393 Basch et al. Dec 2003 B1
6679425 Sheppard et al. Jan 2004 B1
6714918 Hillmer et al. Mar 2004 B2
6714944 Shapiro et al. Mar 2004 B1
6725381 Smith et al. Apr 2004 B1
6734886 Hagan et al. May 2004 B1
6750985 Rhoads Jun 2004 B2
6754665 Futagami et al. Jun 2004 B1
6766327 Morgan, Jr. et al. Jul 2004 B2
6766946 Iida et al. Jul 2004 B2
6782379 Lee Aug 2004 B2
6795812 Lent et al. Sep 2004 B1
6796497 Benkert et al. Sep 2004 B2
6804346 Mewhinney Oct 2004 B1
6805287 Bishop et al. Oct 2004 B2
6816850 Culliss Nov 2004 B2
6816871 Lee Nov 2004 B2
6823319 Lynch et al. Nov 2004 B1
6829711 Kwok et al. Dec 2004 B1
6845448 Chaganti et al. Jan 2005 B1
6850606 Lawyer et al. Feb 2005 B2
6857073 French et al. Feb 2005 B2
6871287 Ellingson Mar 2005 B1
6892307 Wood et al. May 2005 B1
6898574 Regan May 2005 B1
6900731 Kreiner et al. May 2005 B2
6907408 Angel Jun 2005 B2
6908030 Rajasekaran et al. Jun 2005 B2
6910624 Natsuno Jun 2005 B1
6920435 Hoffman et al. Jul 2005 B2
6928487 Eggebraaten et al. Aug 2005 B2
6934714 Meinig Aug 2005 B2
6934849 Kramer et al. Aug 2005 B2
6934858 Woodhill Aug 2005 B2
6947989 Gullotta et al. Sep 2005 B2
6950807 Brock Sep 2005 B2
6950858 Ogami Sep 2005 B2
6965881 Brickell et al. Nov 2005 B1
6968319 Remington et al. Nov 2005 B1
6973462 Dattero et al. Dec 2005 B2
6983381 Jerdonek Jan 2006 B2
6985887 Sunstein et al. Jan 2006 B1
6986461 Geoghegan et al. Jan 2006 B1
6988085 Hedy Jan 2006 B2
6993596 Hinton et al. Jan 2006 B2
6999941 Agarwal Feb 2006 B1
7016907 Boreham et al. Mar 2006 B2
7028013 Saeki Apr 2006 B2
7028052 Chapman et al. Apr 2006 B2
7039607 Watarai et al. May 2006 B2
7043476 Robson May 2006 B2
7058817 Ellmore Jun 2006 B1
7059531 Beenau et al. Jun 2006 B2
7062475 Szabo et al. Jun 2006 B1
7076462 Nelson et al. Jul 2006 B1
7085727 VanOrman Aug 2006 B2
7107241 Pinto Sep 2006 B1
7117172 Black Oct 2006 B1
7121471 Beenau et al. Oct 2006 B2
7124144 Christianson et al. Oct 2006 B2
7154375 Beenau et al. Dec 2006 B2
7155739 Bari et al. Dec 2006 B2
7174454 Roskind Feb 2007 B2
7177846 Moenickheim et al. Feb 2007 B2
7194416 Provost et al. Mar 2007 B1
7200602 Jonas Apr 2007 B2
7203653 McIntosh Apr 2007 B1
7209895 Kundtz et al. Apr 2007 B2
7212995 Schulkins May 2007 B2
7219107 Beringer May 2007 B2
7222369 Vering et al. May 2007 B2
7225464 Satyavolu et al. May 2007 B2
7231657 Honarvar et al. Jun 2007 B2
7234156 French et al. Jun 2007 B2
7234160 Vogel et al. Jun 2007 B2
7237267 Rayes et al. Jun 2007 B2
7243369 Bhat et al. Jul 2007 B2
7246067 Austin et al. Jul 2007 B2
7246740 Swift et al. Jul 2007 B2
7249113 Continelli et al. Jul 2007 B1
7263497 Wiser et al. Aug 2007 B1
7263506 Lee et al. Aug 2007 B2
7277875 Serrano-Morales et al. Oct 2007 B2
7289971 O'Neil et al. Oct 2007 B1
7303120 Beenau et al. Dec 2007 B2
7310611 Shibuya et al. Dec 2007 B2
7314167 Kiliccote Jan 2008 B1
7328233 Salim et al. Feb 2008 B2
7330871 Barber Feb 2008 B2
7333635 Tsantes et al. Feb 2008 B2
7340042 Cluff et al. Mar 2008 B2
7340679 Botscheck et al. Mar 2008 B2
7343149 Benco Mar 2008 B2
7343295 Pomerance Mar 2008 B2
7356503 Johnson et al. Apr 2008 B1
7356516 Richey et al. Apr 2008 B2
7370044 Mulhern et al. May 2008 B2
7370351 Ramachandran et al. May 2008 B1
7383988 Slonecker, Jr. Jun 2008 B2
7386448 Poss et al. Jun 2008 B1
7389913 Starrs Jun 2008 B2
7403942 Bayliss Jul 2008 B1
7421732 Costa-Requena et al. Sep 2008 B2
7428509 Klebanoff Sep 2008 B2
7433855 Gavan et al. Oct 2008 B2
7433864 Malik Oct 2008 B2
7437679 Uemura et al. Oct 2008 B2
7438226 Helsper et al. Oct 2008 B2
7444414 Foster et al. Oct 2008 B2
7444518 Dharmarajan et al. Oct 2008 B1
7451113 Kasower Nov 2008 B1
7458508 Shao et al. Dec 2008 B1
7460857 Roach, Jr. Dec 2008 B2
7467401 Cicchitto Dec 2008 B2
7478157 Bohrer et al. Jan 2009 B2
7480631 Merced et al. Jan 2009 B1
7490356 Lieblich et al. Feb 2009 B2
7503489 Heffez Mar 2009 B2
7509117 Yum Mar 2009 B2
7509278 Jones Mar 2009 B2
7512221 Toms Mar 2009 B2
7519558 Ballard et al. Apr 2009 B2
7529698 Joao May 2009 B2
7530097 Casco-Arias et al. May 2009 B2
7542993 Satterfield et al. Jun 2009 B2
7543739 Brown et al. Jun 2009 B2
7546271 Chmielewski et al. Jun 2009 B1
7548886 Kirkland et al. Jun 2009 B2
7552080 Willard et al. Jun 2009 B1
7552123 Wade et al. Jun 2009 B2
7552467 Lindsay Jun 2009 B2
7555459 Dhar et al. Jun 2009 B2
7562184 Henmi et al. Jul 2009 B2
7562814 Shao et al. Jul 2009 B1
7571473 Boydstun et al. Aug 2009 B1
7575157 Barnhardt et al. Aug 2009 B2
7577665 Ramer et al. Aug 2009 B2
7577934 Anonsen et al. Aug 2009 B2
7580884 Cook Aug 2009 B2
7581112 Brown et al. Aug 2009 B2
7584126 White Sep 2009 B1
7584146 Duhon Sep 2009 B1
7587366 Grim, III et al. Sep 2009 B2
7587368 Felsher Sep 2009 B2
7603701 Gaucas Oct 2009 B2
7606401 Hoffman et al. Oct 2009 B2
7606725 Robertson et al. Oct 2009 B2
7610216 May et al. Oct 2009 B1
7613600 Krane Nov 2009 B2
7620596 Knudson et al. Nov 2009 B2
7623844 Herrmann et al. Nov 2009 B2
7630932 Danaher et al. Dec 2009 B2
7634737 Beringer et al. Dec 2009 B2
7636941 Blinn et al. Dec 2009 B2
7647344 Skurtovich, Jr. et al. Jan 2010 B2
7653592 Flaxman et al. Jan 2010 B1
7653600 Gustin Jan 2010 B2
7653688 Bittner Jan 2010 B2
7657431 Hayakawa Feb 2010 B2
7668769 Baker et al. Feb 2010 B2
7672833 Blume et al. Mar 2010 B2
7685096 Margolus et al. Mar 2010 B2
7685209 Norton et al. Mar 2010 B1
7686214 Shao et al. Mar 2010 B1
7689487 Britt et al. Mar 2010 B1
7689505 Kasower Mar 2010 B2
7689563 Jacobson Mar 2010 B1
7690032 Peirce Mar 2010 B1
7698214 Lindgren Apr 2010 B1
7698217 Phillips et al. Apr 2010 B1
7698445 Fitzpatrick et al. Apr 2010 B2
7707271 Rudkin et al. Apr 2010 B2
7708190 Brandt et al. May 2010 B2
7711635 Steele et al. May 2010 B2
7725385 Royer et al. May 2010 B2
7730078 Schwabe et al. Jun 2010 B2
7739139 Robertson et al. Jun 2010 B2
7747494 Kothari et al. Jun 2010 B1
7747520 Livermore et al. Jun 2010 B2
7747521 Serio Jun 2010 B2
7761379 Zoldi et al. Jul 2010 B2
7761384 Madhogarhia Jul 2010 B2
7761568 Levi et al. Jul 2010 B1
7765166 Beringer et al. Jul 2010 B2
7765311 Itabashi et al. Jul 2010 B2
7769696 Yoda Aug 2010 B2
7769697 Fieschi et al. Aug 2010 B2
7769998 Lynch et al. Aug 2010 B2
7774270 MacCloskey Aug 2010 B1
7778885 Semprevivo et al. Aug 2010 B1
7788040 Haskell et al. Aug 2010 B2
7792715 Kasower Sep 2010 B1
7792725 Booraem et al. Sep 2010 B2
7793835 Coggeshall et al. Sep 2010 B1
7797725 Lunt et al. Sep 2010 B2
7801811 Merrell et al. Sep 2010 B1
7801956 Cumberbatch et al. Sep 2010 B1
7802104 Dickinson Sep 2010 B2
7805362 Merrell et al. Sep 2010 B1
7810036 Bales et al. Oct 2010 B2
7818228 Coulter Oct 2010 B1
7827115 Weller et al. Nov 2010 B2
7841004 Balducci et al. Nov 2010 B1
7841008 Cole et al. Nov 2010 B1
7844520 Franklin Nov 2010 B1
7849014 Erikson Dec 2010 B2
7849624 Holt et al. Dec 2010 B2
7853493 DeBie et al. Dec 2010 B2
7853533 Eisen Dec 2010 B2
7853984 Antell et al. Dec 2010 B2
7856494 Kulkarni Dec 2010 B2
7865958 Lieblich et al. Jan 2011 B2
7870078 Clark et al. Jan 2011 B2
7877304 Coulter Jan 2011 B1
7877784 Chow et al. Jan 2011 B2
7880728 de los Reyes et al. Feb 2011 B2
7908242 Achanta Mar 2011 B1
7909246 Hogg et al. Mar 2011 B2
7912865 Akerman et al. Mar 2011 B2
7930285 Abraham et al. Apr 2011 B2
7930411 Hayward Apr 2011 B1
7941324 Sholtis May 2011 B1
7958046 Doerner et al. Jun 2011 B2
7966192 Pagliari et al. Jun 2011 B2
7970679 Kasower Jun 2011 B2
7975299 Balducci et al. Jul 2011 B1
7979908 Millwee Jul 2011 B2
7983932 Kane Jul 2011 B2
7983979 Holland, IV Jul 2011 B2
7991688 Phelan et al. Aug 2011 B2
7991689 Brunzell et al. Aug 2011 B1
8001042 Brunzell et al. Aug 2011 B1
8001153 Skurtovich, Jr. et al. Aug 2011 B2
8001235 Russ et al. Aug 2011 B2
8005155 Lee et al. Aug 2011 B1
8011582 Ghafarzadeh Sep 2011 B2
8032932 Speyer et al. Oct 2011 B2
8037097 Guo et al. Oct 2011 B2
8041956 White et al. Oct 2011 B1
8055904 Cato et al. Nov 2011 B1
8060424 Kasower Nov 2011 B2
8060916 Bajaj et al. Nov 2011 B2
8065233 Lee et al. Nov 2011 B2
8078453 Shaw Dec 2011 B2
8078524 Crawford et al. Dec 2011 B2
8078881 Liu Dec 2011 B1
8099341 Varghese Jan 2012 B2
8104679 Brown Jan 2012 B2
8116731 Buhrmann et al. Feb 2012 B2
8116751 Aaron Feb 2012 B2
8127982 Casey et al. Mar 2012 B1
8127986 Taylor et al. Mar 2012 B1
8131777 McCullouch Mar 2012 B2
8151327 Eisen Apr 2012 B2
8175889 Girulat et al. May 2012 B1
8185747 Wood et al. May 2012 B2
8195549 Kasower Jun 2012 B2
8214262 Semprevivo et al. Jul 2012 B1
8219771 Le Neel Jul 2012 B2
8224723 Bosch et al. Jul 2012 B2
8225395 Atwood et al. Jul 2012 B2
8229810 Butera et al. Jul 2012 B2
8234498 Britti et al. Jul 2012 B2
8239677 Colson Aug 2012 B2
8239929 Kwan et al. Aug 2012 B2
8241369 Stevens Aug 2012 B2
8244848 Narayanan et al. Aug 2012 B1
8255452 Piliouras Aug 2012 B2
8255971 Webb et al. Aug 2012 B1
8260706 Freishtat et al. Sep 2012 B2
8261334 Hazlehurst et al. Sep 2012 B2
8266065 Dilip et al. Sep 2012 B2
8280348 Snyder et al. Oct 2012 B2
8281372 Vidal Oct 2012 B1
8285613 Coulter Oct 2012 B1
8285656 Chang et al. Oct 2012 B1
8291218 Garcia et al. Oct 2012 B2
8291477 Lunt Oct 2012 B2
8295898 Ashfield et al. Oct 2012 B2
8302164 Lunt Oct 2012 B2
8312033 McMillan Nov 2012 B1
8315940 Winbom et al. Nov 2012 B2
8327429 Speyer et al. Dec 2012 B2
8359278 Domenikos et al. Jan 2013 B2
8364588 Celka et al. Jan 2013 B2
8374634 Dankar et al. Feb 2013 B2
8374973 Herbrich et al. Feb 2013 B2
8406736 Das et al. Mar 2013 B2
8423648 Ferguson et al. Apr 2013 B2
8442886 Haggerty et al. May 2013 B1
8443202 White et al. May 2013 B2
8447016 Kugler May 2013 B1
8456293 Trundle et al. Jun 2013 B1
8464939 Taylor et al. Jun 2013 B1
8468090 Lesandro et al. Jun 2013 B2
8478674 Kapczynski et al. Jul 2013 B1
8484186 Kapczynski et al. Jul 2013 B1
8515828 Wolf et al. Aug 2013 B1
8515844 Kasower Aug 2013 B2
8527357 Ganesan Sep 2013 B1
8527417 Telle et al. Sep 2013 B2
8527773 Metzger Sep 2013 B1
8533118 Weller et al. Sep 2013 B2
8560381 Green et al. Oct 2013 B2
8572391 Golan et al. Oct 2013 B2
8578496 Krishnappa Nov 2013 B1
8588748 Buhrman et al. Nov 2013 B2
8600886 Ramavarjula et al. Dec 2013 B2
8601602 Zheng Dec 2013 B1
8606234 Pei et al. Dec 2013 B2
8606694 Campbell et al. Dec 2013 B2
8630938 Cheng et al. Jan 2014 B2
8646051 Paden et al. Feb 2014 B2
8671115 Skurtovich, Jr. et al. Mar 2014 B2
8688543 Dominguez Apr 2014 B2
8701199 Dotan et al. Apr 2014 B1
8705718 Baniak et al. Apr 2014 B2
8706599 Koenig et al. Apr 2014 B1
8725613 Celka et al. May 2014 B1
8744956 DiChiara et al. Jun 2014 B1
8751388 Chapa Jun 2014 B1
8768914 Scriffignano et al. Jul 2014 B2
8769614 Knox et al. Jul 2014 B1
8781882 Arboletti et al. Jul 2014 B1
8781953 Kasower Jul 2014 B2
8781975 Bennett et al. Jul 2014 B2
8782217 Arone et al. Jul 2014 B1
8782753 Lunt Jul 2014 B2
8793166 Mizhen Jul 2014 B2
8793777 Colson Jul 2014 B2
8800005 Lunt Aug 2014 B2
8806584 Lunt Aug 2014 B2
8818888 Kapczynski et al. Aug 2014 B1
8819793 Gottschalk, Jr. Aug 2014 B2
8826371 Webb et al. Sep 2014 B2
8826393 Eisen Sep 2014 B2
8831564 Ferguson et al. Sep 2014 B2
8839394 Dennis et al. Sep 2014 B2
8856894 Dean Oct 2014 B1
8862514 Eisen Oct 2014 B2
D717332 Nies et al. Nov 2014 S
8931058 DiChiara et al. Jan 2015 B2
8954459 McMillan et al. Feb 2015 B1
8972400 Kapczynski et al. Mar 2015 B1
9100400 Lunt Aug 2015 B2
9106691 Burger et al. Aug 2015 B1
9147042 Haller et al. Sep 2015 B1
9185123 Dennis et al. Nov 2015 B2
9195984 Spector et al. Nov 2015 B1
9196004 Eisen Nov 2015 B2
9235728 Gottschalk, Jr. et al. Jan 2016 B2
9246899 Durney et al. Jan 2016 B1
9256624 Skurtovich, Jr. et al. Feb 2016 B2
9269085 Webb et al. Feb 2016 B2
9361597 Britton et al. Jun 2016 B2
9380057 Knauss Jun 2016 B2
9390384 Eisen Jul 2016 B2
9420448 Dankar et al. Aug 2016 B2
9491160 Livesay et al. Nov 2016 B2
9600651 Ryan et al. Mar 2017 B1
9607336 Dean et al. Mar 2017 B1
9626680 Ryan et al. Apr 2017 B1
9633322 Burger Apr 2017 B1
9665854 Burger et al. May 2017 B1
9684905 Haller et al. Jun 2017 B1
9697521 Webb et al. Jul 2017 B2
9710523 Skurtovich, Jr. et al. Jul 2017 B2
9721147 Kapczynski Aug 2017 B1
9734501 Durney et al. Aug 2017 B2
9754256 Britton et al. Sep 2017 B2
9754311 Eisen Sep 2017 B2
9818121 Snyder et al. Nov 2017 B2
10075446 McMillan et al. Sep 2018 B2
10089679 Eisen Oct 2018 B2
10115079 Burger et al. Oct 2018 B1
10169761 Burger Jan 2019 B1
20010001877 French et al. May 2001 A1
20010029482 Tealdi et al. Oct 2001 A1
20010039532 Coleman, Jr. et al. Nov 2001 A1
20010042785 Walker et al. Nov 2001 A1
20010044729 Pomerance Nov 2001 A1
20010044756 Watkins et al. Nov 2001 A1
20010049274 Degraeve Dec 2001 A1
20020004736 Roundtree et al. Jan 2002 A1
20020013827 Edstrom et al. Jan 2002 A1
20020013899 Faul Jan 2002 A1
20020026519 Itabashi et al. Feb 2002 A1
20020032635 Harris et al. Mar 2002 A1
20020033846 Balasubramanian et al. Mar 2002 A1
20020035684 Vogel et al. Mar 2002 A1
20020045154 Wood et al. Apr 2002 A1
20020059201 Work May 2002 A1
20020059521 Tasler May 2002 A1
20020069122 Yun et al. Jun 2002 A1
20020077964 Brody et al. Jun 2002 A1
20020087460 Hornung Jul 2002 A1
20020091544 Middeljans et al. Jul 2002 A1
20020091635 Dilip et al. Jul 2002 A1
20020099635 Guiragosian Jul 2002 A1
20020099649 Lee et al. Jul 2002 A1
20020103933 Garon et al. Aug 2002 A1
20020111816 Lortscher et al. Aug 2002 A1
20020120537 Morea et al. Aug 2002 A1
20020120757 Sutherland et al. Aug 2002 A1
20020120846 Stewart et al. Aug 2002 A1
20020128962 Kasower Sep 2002 A1
20020133365 Grey et al. Sep 2002 A1
20020133462 Shteyn Sep 2002 A1
20020138470 Zhou Sep 2002 A1
20020143943 Lee et al. Oct 2002 A1
20020147801 Gullotta et al. Oct 2002 A1
20020157029 French et al. Oct 2002 A1
20020161711 Sartor et al. Oct 2002 A1
20020169747 Chapman et al. Nov 2002 A1
20020173994 Ferguson, III Nov 2002 A1
20020174048 Dheer et al. Nov 2002 A1
20020184509 Scheidt et al. Dec 2002 A1
20020198800 Shamrakov Dec 2002 A1
20020198806 Blagg et al. Dec 2002 A1
20020198824 Cook Dec 2002 A1
20030002671 Inchalik et al. Jan 2003 A1
20030009418 Green et al. Jan 2003 A1
20030009426 Ruiz-Sanchez Jan 2003 A1
20030023531 Fergusson Jan 2003 A1
20030036995 Lazerson Feb 2003 A1
20030046311 Baidya et al. Mar 2003 A1
20030046554 Leydier et al. Mar 2003 A1
20030048904 Wang et al. Mar 2003 A1
20030050882 Degen et al. Mar 2003 A1
20030061163 Durfield Mar 2003 A1
20030065563 Elliott et al. Apr 2003 A1
20030069839 Whittington et al. Apr 2003 A1
20030069943 Bahrs et al. Apr 2003 A1
20030078877 Beirne et al. Apr 2003 A1
20030097320 Gordon May 2003 A1
20030097342 Whittingtom May 2003 A1
20030097380 Mulhern et al. May 2003 A1
20030105710 Barbara et al. Jun 2003 A1
20030105733 Boreham Jun 2003 A1
20030105742 Boreham et al. Jun 2003 A1
20030115133 Bian Jun 2003 A1
20030131102 Umbreit Jul 2003 A1
20030154162 Danaher et al. Aug 2003 A1
20030158751 Suresh et al. Aug 2003 A1
20030158960 Engberg Aug 2003 A1
20030163513 Schaeck et al. Aug 2003 A1
20030163733 Barriga-Caceres et al. Aug 2003 A1
20030171942 Gaito Sep 2003 A1
20030177028 Cooper et al. Sep 2003 A1
20030182214 Taylor Sep 2003 A1
20030187837 Culliss Oct 2003 A1
20030195859 Lawrence Oct 2003 A1
20030200447 Sjoblom Oct 2003 A1
20030204429 Botscheck et al. Oct 2003 A1
20030204752 Garrison Oct 2003 A1
20030208412 Hillestad et al. Nov 2003 A1
20030220858 Lam et al. Nov 2003 A1
20030225692 Bosch et al. Dec 2003 A1
20040002878 Hinton Jan 2004 A1
20040006488 Fitall et al. Jan 2004 A1
20040010458 Friedman Jan 2004 A1
20040010698 Rolfe Jan 2004 A1
20040015714 Abraham et al. Jan 2004 A1
20040015715 Brown Jan 2004 A1
20040019518 Abraham et al. Jan 2004 A1
20040019549 Gulbrandsen Jan 2004 A1
20040019799 Vering et al. Jan 2004 A1
20040024671 Freund Feb 2004 A1
20040024709 Yu et al. Feb 2004 A1
20040030649 Nelson et al. Feb 2004 A1
20040039586 Garvey et al. Feb 2004 A1
20040044628 Mathew et al. Mar 2004 A1
20040044673 Brady et al. Mar 2004 A1
20040044739 Ziegler Mar 2004 A1
20040054619 Watson et al. Mar 2004 A1
20040064401 Palaghita et al. Apr 2004 A1
20040078324 Lonnberg et al. Apr 2004 A1
20040083159 Crosby et al. Apr 2004 A1
20040088237 Moenickheim et al. May 2004 A1
20040088255 Zielke et al. May 2004 A1
20040107250 Marciano Jun 2004 A1
20040110119 Riconda et al. Jun 2004 A1
20040111305 Gavan et al. Jun 2004 A1
20040111359 Hudock Jun 2004 A1
20040111375 Johnson Jun 2004 A1
20040117235 Shacham Jun 2004 A1
20040117302 Weichert et al. Jun 2004 A1
20040122681 Ruvolo et al. Jun 2004 A1
20040122696 Beringer Jun 2004 A1
20040123162 Antell et al. Jun 2004 A1
20040128150 Lundegren Jul 2004 A1
20040128156 Beringer et al. Jul 2004 A1
20040133440 Carolan et al. Jul 2004 A1
20040133509 McCoy et al. Jul 2004 A1
20040133513 McCoy et al. Jul 2004 A1
20040133515 McCoy et al. Jul 2004 A1
20040138994 DeFrancesco et al. Jul 2004 A1
20040141005 Banatwala et al. Jul 2004 A1
20040143546 Wood et al. Jul 2004 A1
20040143596 Sirkin Jul 2004 A1
20040153521 Kogo Aug 2004 A1
20040158523 Dort Aug 2004 A1
20040158723 Root Aug 2004 A1
20040159700 Khan et al. Aug 2004 A1
20040167793 Masuoka et al. Aug 2004 A1
20040177046 Ogram Sep 2004 A1
20040193891 Ollila Sep 2004 A1
20040199462 Starrs Oct 2004 A1
20040199789 Shaw et al. Oct 2004 A1
20040210661 Thompson Oct 2004 A1
20040220865 Lozowski et al. Nov 2004 A1
20040220918 Scriffignano et al. Nov 2004 A1
20040225643 Alpha et al. Nov 2004 A1
20040230448 Schaich Nov 2004 A1
20040230527 Hansen et al. Nov 2004 A1
20040243514 Wankmueller Dec 2004 A1
20040243518 Clifton et al. Dec 2004 A1
20040243588 Tanner et al. Dec 2004 A1
20040243832 Wilf et al. Dec 2004 A1
20040249811 Shostack Dec 2004 A1
20040250085 Tattan et al. Dec 2004 A1
20040250107 Guo Dec 2004 A1
20040254935 Chagoly et al. Dec 2004 A1
20040255127 Arnouse Dec 2004 A1
20040267714 Frid et al. Dec 2004 A1
20050005168 Dick Jan 2005 A1
20050010513 Duckworth et al. Jan 2005 A1
20050021476 Candella et al. Jan 2005 A1
20050021551 Silva et al. Jan 2005 A1
20050027983 Klawon Feb 2005 A1
20050027995 Menschik et al. Feb 2005 A1
20050055231 Lee Mar 2005 A1
20050058262 Timmins et al. Mar 2005 A1
20050060332 Bernstein et al. Mar 2005 A1
20050065874 Lefner et al. Mar 2005 A1
20050071328 Lawrence Mar 2005 A1
20050075985 Cartmell Apr 2005 A1
20050086126 Patterson Apr 2005 A1
20050091164 Varble Apr 2005 A1
20050097017 Hanratty May 2005 A1
20050097039 Kulcsar et al. May 2005 A1
20050097051 Madill, Jr. et al. May 2005 A1
20050097320 Golan et al. May 2005 A1
20050102180 Gailey et al. May 2005 A1
20050105719 Huda May 2005 A1
20050108396 Bittner May 2005 A1
20050108631 Amorin et al. May 2005 A1
20050114335 Wesinger, Jr. et al. May 2005 A1
20050114344 Wesinger, Jr. et al. May 2005 A1
20050114345 Wesinger, Jr. et al. May 2005 A1
20050119978 Ates Jun 2005 A1
20050125291 Demkiw Grayson et al. Jun 2005 A1
20050125397 Gross et al. Jun 2005 A1
20050125686 Brandt Jun 2005 A1
20050137899 Davies et al. Jun 2005 A1
20050138391 Mandalia et al. Jun 2005 A1
20050144452 Lynch et al. Jun 2005 A1
20050154664 Guy et al. Jul 2005 A1
20050154665 Kerr Jul 2005 A1
20050154769 Eckart et al. Jul 2005 A1
20050166262 Beattie et al. Jul 2005 A1
20050171884 Arnott Aug 2005 A1
20050181765 Mark Aug 2005 A1
20050204041 Blinn et al. Sep 2005 A1
20050208461 Krebs et al. Sep 2005 A1
20050216434 Haveliwala et al. Sep 2005 A1
20050216582 Toomey et al. Sep 2005 A1
20050216953 Ellingson Sep 2005 A1
20050216955 Wilkins et al. Sep 2005 A1
20050226224 Lee et al. Oct 2005 A1
20050240578 Biederman et al. Oct 2005 A1
20050256809 Sadri Nov 2005 A1
20050267840 Holm-Blagg et al. Dec 2005 A1
20050273431 Abel et al. Dec 2005 A1
20050273442 Bennett et al. Dec 2005 A1
20050278542 Pierson et al. Dec 2005 A1
20050288998 Verma et al. Dec 2005 A1
20060004623 Jasti Jan 2006 A1
20060004626 Holmen et al. Jan 2006 A1
20060010072 Eisen Jan 2006 A1
20060010391 Uemura et al. Jan 2006 A1
20060010487 Fierer et al. Jan 2006 A1
20060016107 Davis Jan 2006 A1
20060032909 Seegar Feb 2006 A1
20060036543 Blagg et al. Feb 2006 A1
20060036748 Nusbaum et al. Feb 2006 A1
20060036870 Dasari et al. Feb 2006 A1
20060041464 Powers et al. Feb 2006 A1
20060041670 Musseleck et al. Feb 2006 A1
20060059110 Madhok et al. Mar 2006 A1
20060059362 Paden et al. Mar 2006 A1
20060069635 Ram et al. Mar 2006 A1
20060074986 Mallalieu et al. Apr 2006 A1
20060074991 Lussier et al. Apr 2006 A1
20060079211 Degraeve Apr 2006 A1
20060080230 Freiberg Apr 2006 A1
20060080251 Fried et al. Apr 2006 A1
20060080263 Willis et al. Apr 2006 A1
20060085361 Hoerle et al. Apr 2006 A1
20060101508 Taylor May 2006 A1
20060129419 Flaxer et al. Jun 2006 A1
20060129481 Bhatt et al. Jun 2006 A1
20060129533 Purvis Jun 2006 A1
20060131390 Kim Jun 2006 A1
20060136595 Satyavolu Jun 2006 A1
20060140460 Coutts Jun 2006 A1
20060155573 Hartunian Jul 2006 A1
20060155780 Sakairi et al. Jul 2006 A1
20060161435 Atef et al. Jul 2006 A1
20060161554 Lucovsky et al. Jul 2006 A1
20060173776 Shalley et al. Aug 2006 A1
20060173792 Glass Aug 2006 A1
20060178971 Owen et al. Aug 2006 A1
20060179050 Giang et al. Aug 2006 A1
20060184585 Grear et al. Aug 2006 A1
20060195351 Bayburtian Aug 2006 A1
20060202012 Grano et al. Sep 2006 A1
20060204051 Holland, IV Sep 2006 A1
20060212407 Lyon Sep 2006 A1
20060218407 Toms Sep 2006 A1
20060229943 Mathias et al. Oct 2006 A1
20060229961 Lyftogt et al. Oct 2006 A1
20060235935 Ng Oct 2006 A1
20060239512 Petrillo Oct 2006 A1
20060253358 Delgrosso et al. Nov 2006 A1
20060262929 Vatanen et al. Nov 2006 A1
20060265243 Racho et al. Nov 2006 A1
20060271456 Romain et al. Nov 2006 A1
20060271457 Romain et al. Nov 2006 A1
20060271633 Adler Nov 2006 A1
20060277089 Hubbard et al. Dec 2006 A1
20060282429 Hernandez-Sherrington et al. Dec 2006 A1
20060282660 Varghese et al. Dec 2006 A1
20060282819 Graham et al. Dec 2006 A1
20060287764 Kraft Dec 2006 A1
20060287765 Kraft Dec 2006 A1
20060287766 Kraft Dec 2006 A1
20060287767 Kraft Dec 2006 A1
20060288090 Kraft Dec 2006 A1
20060294199 Bertholf Dec 2006 A1
20070005508 Chiang Jan 2007 A1
20070005984 Florencio et al. Jan 2007 A1
20070016522 Wang Jan 2007 A1
20070022141 Singleton et al. Jan 2007 A1
20070027816 Writer Feb 2007 A1
20070032240 Finnegan et al. Feb 2007 A1
20070038483 Wood Feb 2007 A1
20070038568 Greene et al. Feb 2007 A1
20070043577 Kasower Feb 2007 A1
20070047714 Baniak et al. Mar 2007 A1
20070059442 Sabeta Mar 2007 A1
20070067297 Kublickis Mar 2007 A1
20070072190 Aggarwal Mar 2007 A1
20070073889 Morris Mar 2007 A1
20070078908 Rohatgi et al. Apr 2007 A1
20070078985 Shao et al. Apr 2007 A1
20070083460 Bachenheimer Apr 2007 A1
20070083463 Kraft Apr 2007 A1
20070093234 Willis et al. Apr 2007 A1
20070094137 Phillips et al. Apr 2007 A1
20070094230 Subramaniam et al. Apr 2007 A1
20070094241 Blackwell et al. Apr 2007 A1
20070106517 Cluff et al. May 2007 A1
20070106582 Baker et al. May 2007 A1
20070112667 Rucker May 2007 A1
20070112668 Celano et al. May 2007 A1
20070121843 Atazky et al. May 2007 A1
20070124256 Crooks et al. May 2007 A1
20070143825 Goffin Jun 2007 A1
20070156692 Rosewarne Jul 2007 A1
20070162307 Austin et al. Jul 2007 A1
20070174186 Hokland Jul 2007 A1
20070174448 Ahuja et al. Jul 2007 A1
20070174903 Greff Jul 2007 A1
20070192121 Routson et al. Aug 2007 A1
20070192248 West Aug 2007 A1
20070192853 Shraim et al. Aug 2007 A1
20070198432 Pitroda et al. Aug 2007 A1
20070204338 Aiello et al. Aug 2007 A1
20070205266 Carr et al. Sep 2007 A1
20070226093 Chan et al. Sep 2007 A1
20070226122 Burrell et al. Sep 2007 A1
20070240206 Wu et al. Oct 2007 A1
20070244807 Andringa et al. Oct 2007 A1
20070245245 Blue et al. Oct 2007 A1
20070250441 Paulsen et al. Oct 2007 A1
20070250459 Schwarz et al. Oct 2007 A1
20070261108 Lee et al. Nov 2007 A1
20070261114 Pomerantsev Nov 2007 A1
20070266439 Kraft Nov 2007 A1
20070282743 Lovelett Dec 2007 A1
20070288355 Roland et al. Dec 2007 A1
20070288360 Seeklus Dec 2007 A1
20070294195 Curry et al. Dec 2007 A1
20080010203 Grant Jan 2008 A1
20080010206 Coleman Jan 2008 A1
20080010687 Gonen et al. Jan 2008 A1
20080028446 Burgoyne Jan 2008 A1
20080033742 Bernasconi Feb 2008 A1
20080033956 Saha et al. Feb 2008 A1
20080040610 Fergusson Feb 2008 A1
20080047017 Renaud Feb 2008 A1
20080052182 Marshall Feb 2008 A1
20080052244 Tsuei et al. Feb 2008 A1
20080059364 Tidwell et al. Mar 2008 A1
20080066188 Kwak Mar 2008 A1
20080071682 Dominguez Mar 2008 A1
20080072316 Chang et al. Mar 2008 A1
20080077526 Arumugam Mar 2008 A1
20080082536 Schwabe et al. Apr 2008 A1
20080083021 Doane et al. Apr 2008 A1
20080086431 Robinson et al. Apr 2008 A1
20080086759 Colson Apr 2008 A1
20080091530 Egnatios et al. Apr 2008 A1
20080103800 Domenikos et al. May 2008 A1
20080103972 Lanc May 2008 A1
20080104672 Lunde et al. May 2008 A1
20080109422 Dedhia May 2008 A1
20080109875 Kraft May 2008 A1
20080114670 Friesen May 2008 A1
20080115191 Kim et al. May 2008 A1
20080115226 Welingkar et al. May 2008 A1
20080120569 Mann et al. May 2008 A1
20080120716 Hall et al. May 2008 A1
20080126233 Hogan May 2008 A1
20080141346 Kay et al. Jun 2008 A1
20080147454 Walker et al. Jun 2008 A1
20080148368 Zurko et al. Jun 2008 A1
20080154758 Schattmaier et al. Jun 2008 A1
20080155686 McNair Jun 2008 A1
20080162317 Banaugh et al. Jul 2008 A1
20080162350 Allen-Rouman et al. Jul 2008 A1
20080162383 Kraft Jul 2008 A1
20080167883 Khazaneh Jul 2008 A1
20080175360 Schwarz et al. Jul 2008 A1
20080183480 Carlson et al. Jul 2008 A1
20080183585 Vianello Jul 2008 A1
20080195548 Chu et al. Aug 2008 A1
20080201401 Pugh et al. Aug 2008 A1
20080205655 Wilkins et al. Aug 2008 A1
20080208726 Tsantes et al. Aug 2008 A1
20080208735 Balet et al. Aug 2008 A1
20080208752 Gottlieb et al. Aug 2008 A1
20080208873 Boehmer Aug 2008 A1
20080212845 Lund Sep 2008 A1
20080216156 Kosaka Sep 2008 A1
20080222706 Renaud et al. Sep 2008 A1
20080222722 Navratil et al. Sep 2008 A1
20080229415 Kapoor et al. Sep 2008 A1
20080249869 Angell et al. Oct 2008 A1
20080255992 Lin Oct 2008 A1
20080256613 Grover Oct 2008 A1
20080263058 Peden Oct 2008 A1
20080270295 Lent et al. Oct 2008 A1
20080270299 Peng Oct 2008 A1
20080281737 Fajardo Nov 2008 A1
20080288283 Baldwin, Jr. et al. Nov 2008 A1
20080288299 Schultz Nov 2008 A1
20080288382 Smth et al. Nov 2008 A1
20080294540 Celka et al. Nov 2008 A1
20080301016 Durvasula et al. Dec 2008 A1
20080306750 Wunder et al. Dec 2008 A1
20080319889 Hammad Dec 2008 A1
20090006230 Lyda et al. Jan 2009 A1
20090018986 Alcorn et al. Jan 2009 A1
20090031426 Dal Lago et al. Jan 2009 A1
20090037332 Cheung et al. Feb 2009 A1
20090043691 Kasower Feb 2009 A1
20090044279 Crawford et al. Feb 2009 A1
20090055322 Bykov et al. Feb 2009 A1
20090055894 Lorsch Feb 2009 A1
20090064297 Seigas et al. Mar 2009 A1
20090094237 Churi et al. Apr 2009 A1
20090094674 Schwartz et al. Apr 2009 A1
20090100047 Jones et al. Apr 2009 A1
20090106141 Becker Apr 2009 A1
20090106150 Pelegero et al. Apr 2009 A1
20090106846 Dupray et al. Apr 2009 A1
20090119299 Rhodes May 2009 A1
20090125369 Kloostra et al. May 2009 A1
20090125972 Hinton et al. May 2009 A1
20090126013 Atwood et al. May 2009 A1
20090132347 Anderson et al. May 2009 A1
20090138335 Lieberman May 2009 A1
20090144166 Dickelman Jun 2009 A1
20090150166 Leite et al. Jun 2009 A1
20090150238 Marsh et al. Jun 2009 A1
20090157564 Cross Jun 2009 A1
20090157693 Palahnuk Jun 2009 A1
20090158030 Rasti Jun 2009 A1
20090164232 Chmielewski et al. Jun 2009 A1
20090164380 Brown Jun 2009 A1
20090172788 Veldula et al. Jul 2009 A1
20090172795 Ritari et al. Jul 2009 A1
20090177529 Hadi Jul 2009 A1
20090177562 Peace et al. Jul 2009 A1
20090182653 Zimiles Jul 2009 A1
20090183259 Rinek et al. Jul 2009 A1
20090199264 Lang Aug 2009 A1
20090199294 Schneider Aug 2009 A1
20090204514 Bhogal et al. Aug 2009 A1
20090204599 Morris et al. Aug 2009 A1
20090210241 Calloway Aug 2009 A1
20090210807 Xiao et al. Aug 2009 A1
20090215431 Koraichi Aug 2009 A1
20090216640 Masi Aug 2009 A1
20090222308 Zoldi et al. Sep 2009 A1
20090222449 Hom et al. Sep 2009 A1
20090228918 Rolff et al. Sep 2009 A1
20090234665 Conkel Sep 2009 A1
20090234775 Whitney et al. Sep 2009 A1
20090234876 Schigel et al. Sep 2009 A1
20090240624 James et al. Sep 2009 A1
20090247122 Fitzgerald et al. Oct 2009 A1
20090254375 Martinez et al. Oct 2009 A1
20090254476 Sharma et al. Oct 2009 A1
20090254572 Redlich et al. Oct 2009 A1
20090254656 Vignisson et al. Oct 2009 A1
20090254971 Herz et al. Oct 2009 A1
20090260064 Mcdowell et al. Oct 2009 A1
20090307778 Mardikar Dec 2009 A1
20090313562 Appleyard et al. Dec 2009 A1
20090327270 Teevan et al. Dec 2009 A1
20090328173 Jakobson et al. Dec 2009 A1
20100011428 Atwood et al. Jan 2010 A1
20100030578 Siddique et al. Feb 2010 A1
20100030677 Melik-Asianian et al. Feb 2010 A1
20100042542 Rose et al. Feb 2010 A1
20100043055 Baumgart Feb 2010 A1
20100049803 Ogilvie et al. Feb 2010 A1
20100058404 Rouse Mar 2010 A1
20100063942 Arnott et al. Mar 2010 A1
20100063993 Higgins et al. Mar 2010 A1
20100076836 Giordano et al. Mar 2010 A1
20100077483 Stolfo et al. Mar 2010 A1
20100083371 Bennetts et al. Apr 2010 A1
20100088233 Tattan et al. Apr 2010 A1
20100094664 Bush et al. Apr 2010 A1
20100094768 Miltonberger Apr 2010 A1
20100094910 Bayliss Apr 2010 A1
20100100945 Ozzie et al. Apr 2010 A1
20100114744 Gonen May 2010 A1
20100114776 Weller et al. May 2010 A1
20100121767 Coulter et al. May 2010 A1
20100122324 Welingkar et al. May 2010 A1
20100122333 Noe et al. May 2010 A1
20100130172 Vendrow et al. May 2010 A1
20100136956 Drachev et al. Jun 2010 A1
20100138298 Fitzgerald et al. Jun 2010 A1
20100145836 Baker et al. Jun 2010 A1
20100153278 Farsedakis Jun 2010 A1
20100153290 Duggan Jun 2010 A1
20100161816 Kraft et al. Jun 2010 A1
20100169159 Rose et al. Jul 2010 A1
20100174638 DeBie Jul 2010 A1
20100174813 Hildreth et al. Jul 2010 A1
20100179906 Hawkes Jul 2010 A1
20100185546 Pollard Jul 2010 A1
20100205076 Parson et al. Aug 2010 A1
20100205662 Ibrahim et al. Aug 2010 A1
20100211445 Bodington Aug 2010 A1
20100211636 Starkenburg et al. Aug 2010 A1
20100217837 Ansari et al. Aug 2010 A1
20100223192 Levine et al. Sep 2010 A1
20100229245 Singhal Sep 2010 A1
20100241493 Onischuk Sep 2010 A1
20100241535 Nightengale et al. Sep 2010 A1
20100250338 Banerjee et al. Sep 2010 A1
20100250410 Song et al. Sep 2010 A1
20100250411 Ogrodski Sep 2010 A1
20100250955 Trevithick Sep 2010 A1
20100257102 Perlman Oct 2010 A1
20100258623 Beerner et al. Oct 2010 A1
20100262932 Pan Oct 2010 A1
20100280914 Carlson Nov 2010 A1
20100281020 Drubner Nov 2010 A1
20100293090 Domenikos et al. Nov 2010 A1
20100299262 Handler Nov 2010 A1
20100325442 Petrone et al. Dec 2010 A1
20100325694 Bhagavatula et al. Dec 2010 A1
20100332393 Weller et al. Dec 2010 A1
20110004498 Readshaw Jan 2011 A1
20110016533 Zeigler et al. Jan 2011 A1
20110023115 Wright Jan 2011 A1
20110029388 Kendall et al. Feb 2011 A1
20110035788 White et al. Feb 2011 A1
20110040736 Kalaboukis Feb 2011 A1
20110071950 Ivanovic Mar 2011 A1
20110082768 Eisen Apr 2011 A1
20110083181 Nazarov Apr 2011 A1
20110113084 Ramnani May 2011 A1
20110126024 Beatson et al. May 2011 A1
20110126275 Anderson et al. May 2011 A1
20110131096 Frew et al. Jun 2011 A1
20110131123 Griffin et al. Jun 2011 A1
20110137760 Rudie et al. Jun 2011 A1
20110142213 Baniak et al. Jun 2011 A1
20110145899 Cao et al. Jun 2011 A1
20110148625 Velusamy Jun 2011 A1
20110161218 Swift Jun 2011 A1
20110166988 Coulter Jul 2011 A1
20110167011 Paltenghe et al. Jul 2011 A1
20110173681 Qureshi et al. Jul 2011 A1
20110179139 Starkenburg et al. Jul 2011 A1
20110184780 Alderson et al. Jul 2011 A1
20110184838 Winters et al. Jul 2011 A1
20110196791 Dominguez Aug 2011 A1
20110208601 Ferguson et al. Aug 2011 A1
20110211445 Chen Sep 2011 A1
20110260832 Ross et al. Oct 2011 A1
20110264566 Brown Oct 2011 A1
20110270754 Kelly et al. Nov 2011 A1
20110307397 Benmbarek Dec 2011 A1
20110307957 Barcelo et al. Dec 2011 A1
20120011158 Avner et al. Jan 2012 A1
20120016948 Sinha Jan 2012 A1
20120017266 DiChiara et al. Jan 2012 A1
20120030216 Churi et al. Feb 2012 A1
20120030771 Pierson et al. Feb 2012 A1
20120047219 Feng et al. Feb 2012 A1
20120054095 Lesandro Mar 2012 A1
20120054592 Jaffe et al. Mar 2012 A1
20120072382 Pearson et al. Mar 2012 A1
20120084866 Stolfo Apr 2012 A1
20120089438 Tavares et al. Apr 2012 A1
20120108274 Acebo Ruiz et al. May 2012 A1
20120110467 Blake et al. May 2012 A1
20120110677 Abendroth et al. May 2012 A1
20120124498 Santoro et al. May 2012 A1
20120130898 Snyder et al. May 2012 A1
20120136763 Megdal et al. May 2012 A1
20120151045 Anakata et al. Jun 2012 A1
20120158574 Brunzell et al. Jun 2012 A1
20120173339 Flynt et al. Jul 2012 A1
20120173563 Griffin et al. Jul 2012 A1
20120215682 Lent et al. Aug 2012 A1
20120215719 Verlander Aug 2012 A1
20120216125 Pierce Aug 2012 A1
20120235897 Hirota Sep 2012 A1
20120239497 Nuzzi Sep 2012 A1
20120246060 Conyack, Jr. et al. Sep 2012 A1
20120246730 Raad Sep 2012 A1
20120253852 Pourfallah et al. Oct 2012 A1
20120290660 Rao et al. Nov 2012 A1
20120297484 Srivastava Nov 2012 A1
20120323717 Kirsch Dec 2012 A1
20120331557 Washington Dec 2012 A1
20130004033 Trugenberger et al. Jan 2013 A1
20130006843 Tralvex Jan 2013 A1
20130018811 Britti et al. Jan 2013 A1
20130031109 Roulson et al. Jan 2013 A1
20130031624 Britti et al. Jan 2013 A1
20130041701 Roth Feb 2013 A1
20130066775 Milam Mar 2013 A1
20130080467 Carson et al. Mar 2013 A1
20130085804 Leff et al. Apr 2013 A1
20130085939 Colak et al. Apr 2013 A1
20130110678 Vigier et al. May 2013 A1
20130117087 Coppinger May 2013 A1
20130125010 Strandell May 2013 A1
20130132151 Stibel et al. May 2013 A1
20130139229 Fried et al. May 2013 A1
20130173449 Ng et al. Jul 2013 A1
20130173450 Celka et al. Jul 2013 A1
20130179955 Bekker et al. Jul 2013 A1
20130198525 Spies et al. Aug 2013 A1
20130205135 Lutz Aug 2013 A1
20130246528 Ogura Sep 2013 A1
20130254096 Serio et al. Sep 2013 A1
20130271272 Dhesi et al. Oct 2013 A1
20130279676 Baniak et al. Oct 2013 A1
20130290097 Balestrieri et al. Oct 2013 A1
20130293363 Plymouth Nov 2013 A1
20130298238 Shah et al. Nov 2013 A1
20130332342 Kasower Dec 2013 A1
20130339217 Breslow et al. Dec 2013 A1
20130339249 Weller et al. Dec 2013 A1
20140012733 Vidal Jan 2014 A1
20140025475 Burke Jan 2014 A1
20140032723 Nema Jan 2014 A1
20140046872 Arnott et al. Feb 2014 A1
20140051464 Ryan et al. Feb 2014 A1
20140061302 Hammad Mar 2014 A1
20140089167 Kasower Mar 2014 A1
20140110477 Hammad Apr 2014 A1
20140164112 Kala Jun 2014 A1
20140164398 Smith et al. Jun 2014 A1
20140164519 Shah Jun 2014 A1
20140201100 Rellas et al. Jul 2014 A1
20140258083 Achanta et al. Sep 2014 A1
20140279467 Chapa et al. Sep 2014 A1
20140280945 Lunt Sep 2014 A1
20140283123 Lonstein et al. Sep 2014 A1
20140289812 Wang et al. Sep 2014 A1
20140298485 Gardner Oct 2014 A1
20140317023 Kim Oct 2014 A1
20140331282 Tkachev Nov 2014 A1
20140379600 Chapa et al. Dec 2014 A1
20150067341 Deen et al. Mar 2015 A1
20150249655 Lunt Sep 2015 A1
20150254658 Bondesen et al. Sep 2015 A1
20150326580 McMillan et al. Nov 2015 A1
20160027008 John Jan 2016 A1
20160275476 Artman et al. Sep 2016 A1
20170186012 McNeal Jun 2017 A1
20170200223 Kasower Jul 2017 A1
20170337557 Durney et al. Nov 2017 A1
20180046856 Kapczynski Feb 2018 A1
20180343265 McMillan et al. Nov 2018 A1
Foreign Referenced Citations (40)
Number Date Country
1 028 401 Aug 2000 EP
1 239 378 Sep 2002 EP
1 301 887 Apr 2003 EP
1 850 278 Oct 2007 EP
2 074 513 Feb 2016 EP
2 392 748 Mar 2004 GB
2005-208945 Aug 2005 JP
10-2000-0063313 Nov 2000 KR
10-2002-0039203 May 2002 KR
10-2007-0081504 Aug 2007 KR
256569 Jun 2006 TW
WO 99054803 Oct 1999 WO
WO 99060481 Nov 1999 WO
WO 00030045 May 2000 WO
WO 01009752 Feb 2001 WO
WO 01009792 Feb 2001 WO
WO 01084281 Nov 2001 WO
WO 02027610 Apr 2002 WO
WO 02029636 Apr 2002 WO
WO 03071388 Aug 2003 WO
WO 2004031986 Apr 2004 WO
WO 2004046882 Jun 2004 WO
WO 2005033979 Apr 2005 WO
WO 2006019752 Feb 2006 WO
WO 2006050278 May 2006 WO
WO 2006069199 Jun 2006 WO
WO 2006099081 Sep 2006 WO
WO 2007001394 Jan 2007 WO
WO 2007050156 May 2007 WO
WO 2008042614 Apr 2008 WO
WO 2008147918 Dec 2008 WO
WO 2009064694 May 2009 WO
WO 2009102391 Aug 2009 WO
WO 2009117468 Sep 2009 WO
WO 2010001406 Jan 2010 WO
WO 2010062537 Jun 2010 WO
WO 2010077989 Jul 2010 WO
WO 2010150251 Dec 2010 WO
WO 2011005876 Jan 2011 WO
WO 2015038520 Mar 2015 WO
Non-Patent Literature Citations (126)
Entry
Wang, et al., “User Identification Based on Finger-vein Patterns for Consumer Electronics Devices”, IEEE Transactions on Consumer Electronics, vol. 56, No. 2, May 2010, pp. 799-804 (Year: 2010).
U.S. Appl. No. 14/208,595, U.S. Pat. No. 9,633,322, Adjustment of Knowledge-Based Authentication, filed Mar. 13, 2014.
U.S. Appl. No. 15/459,927, U.S. Pat. No. 10,169,761, Adjustment of Knowledge-Based Authentication, filed Mar. 15, 2017.
U.S. Appl. No. 12/705,489, filed Feb. 12, 2010, Bargoli et al.
U.S. Appl. No. 12/705,511, filed Feb. 12, 2010, Bargoli et al.
Actuate, “Delivering Enterprise Information for Corporate Portals”, White Paper, 2004, pp. 1-7.
“Aggregate and Analyze Social Media Content: Gain Faster and Broader Insight to Market Sentiment,” SAP Partner, Mantis Technology Group, Apr. 2011, pp. 4.
Aharony et al., “Social Area Networks: Data Networking of the People, by the People, for the People,” 2009 International Conference on Computational Science and Engineering, May 2009, pp. 1148-1155.
Aktas et al., “Personalizing PageRank Based on Domain Profiles”, WEBKDD workshop: Webmining and Web Usage Analysis, Aug. 22, 2004, pp. 83-90.
Aktas et al., “Using Hyperlink Features to Personalize Web Search”, WEBKDD workshop: Webmining and Web Usage Analysis, Aug. 2004.
“Arizona Company Has Found Key in Stopping ID Theft,” PR Newswire, New York, Aug. 10, 2005 http://proquest.umi.com/pqdweb?did=880104711&sid=1&Fmt=3&clientId=19649&RQT=309&Vname=PQD.
ABC News Now:Money Matters, as broadcasted Nov. 15, 2005 with guest Todd Davis (CEO of Lifelock), pp. 6.
Anonymous, “Credit-Report Disputes Await Electronic Resolution,” Credit Card News, Chicago, Jan. 15, 1993, vol. 5, No. 19, p. 5.
Anonymous, “MBNA Offers Resolution of Credit Card Disputes,” Hempstead, Feb. 2002, vol. 68, No. 2, p. 47.
Anonymous, “Feedback”, Credit Management, ABI/INFORM Global, Sep. 2006, pp. 6.
Bielski, Lauren, “Will you Spend to Thwart ID Theft?” ABA Banking Journal, Apr. 2005, pp. 54, 56-57, 60.
BlueCava, “What We Do”, http://www.bluecava.com/what-we-do/, printed Nov. 5, 2012 in 3 pages.
Buxfer, http://www.buxfer.com/ printed Feb. 5, 2014 in 1 page.
Check, http://check.me/ printed Feb. 5, 2014 in 3 pages.
Chores & Allowances, “Do Kids Have Credit Reports?” Oct. 15, 2007, http://choresandallowances.blogspot.com/2007/10/do-kids-have-credit-reports.html, pp. 5.
Comlounge.net, “plonesocial.auth.rpx” http://web.archive.org/web/20101026041841/http://comlounge.net/rpx as captured Oct. 26, 2010 in 9 pages.
“Consumers Gain Immediate and Full Access to Credit Score Used by Majority of U.S. Lenders”, PR Newswire, ProQuest, Mar. 19, 2001, p. 1.
Cowie, Norman, “Warning Bells & ‘The Bust-Out’”, Business Credit, Jul. 1, 2000, pp. 5.
“CreditCheck Monitoring Services,” Dec. 11, 2000, pp. 1, lines 21-23.
Cullen, Terri; “The Wall Street Journal Complete Identity Theft Guidebook:How to Protect Yourself from the Most Pervasive Crime in America”; Chapter 3, pp. 59-79; Jul. 10, 2007.
“D&B Corporate Family Linkage”, D&B Internet Access for U.S. Contract Customers, https://www.dnb.com/ecomp/help/linkage.htm as printed Dec. 17, 2009, pp. 1.
Day, Jo and Kevin; “ID-ology: A Planner's Guide to Identity Theft”; Journal of Financial Planning:Tech Talk; pp. 36-38; Sep. 2004.
EFunds Corporation, “Data & Decisioning: Debit Report” printed Apr. 1, 2007, http://www.efunds.com/web/industry-solutions/financial-services/frm-debit-report/htm in 1 page.
Equifax; “Equifax Credit Watch”; https://www.econsumer.equifax.co.uk/consumer/uk/sitepage.ehtml, dated Jun. 27, 2007 on www.archive.org.
Ettorre, “Paul Kahn on Exceptional Marketing,” Management Review, vol. 83, No. 11, Nov. 1994, pp. 48-51.
Facebook, “Facebook helps you connect and share with the people in your life,” www.facebook.com printed Nov. 16, 2010 in 1 page.
“Fair Isaac Introduces Falcon One System to Combat Fraud at Every Customer Interaction”, Business Wire, May 5, 2005, pp. 3.
“Fair Isaac Offers New Fraud Tod”, National Mortgage News & Source Media, Inc., Jun. 13, 2005, pp. 2.
FamilySecure.com, “Frequently Asked Questions”, http://www.familysecure.com/FAQ.aspx as archived Jul. 15, 2007 in 3 pages.
FamilySecure.com; “Identity Theft Protection far the Whole Family | FamilySecure.com” http://www.familysecure.com/, as retrieved on Nov. 5, 2009.
Fenner, Peter, “Mobile Address Management and Billing for Personal Communications”, 1st International Conference on Universal Personal Communications, 1992, ICUPC '92 Proceedings, pp. 253-257.
“Fictitious Business Name Records”, Westlaw Database Directory, http://directory.westlaw.com/scope/default.asp?db=FBN-ALL&RS-W...&VR=2.0 as printed Dec. 17, 2009, pp. 5.
“FinExtra, Basepoint Analytics Introduces Predictive Technology for Mortgage Fraud”, Oct. 5, 2005, pp. 3.
Fisher, Joseph, “Access to Fair Credit Reports: Current Practices and Proposed Legislation,” American Business Law Journal, Fall 1981, vol. 19, No. 3, p. 319.
Gibbs, Adrienne; “Protecting Your Children from Identity Theft,” Nov. 25, 2008, http://www.creditcards.com/credit-card-news/identity-ID-theft-and-kids-children-1282.php, pp. 4.
Gordon et al., “Identity Fraud: A Critical National and Global Threat,” LexisNexis, Oct. 28, 2003, pp. 1-48.
Harrington et al., “iOS 4 In Action”, Chapter 17, Local and Push Notification Services, Manning Publications Co., Jun. 2011, pp. 347-353.
Herzberg, Amir, “Payments and Banking with Mobile Personal Devices,” Communications of the ACM, May 2003, vol. 46, No. 5, pp. 53-58.
Hoofnagle, Chris Jay, “Identity Theft: Making the Known Unknowns Known,” Harvard Journal of Law & Technology, Fall 2007, vol. 21, No. 1, pp. 98-122.
ID Analytics, “ID Analytics® Consumer Notification Service” printed Apr. 16, 2013 in 2 pages.
ID Theft Assist, “Do You Know Where Your Child's Credit Is?”, Nov. 26, 2007, http://www.idtheftassist.com/pages/story14, pp. 3.
“ID Thieves These Days Want Your Number, Not Your Name”, The Columbus Dispatch, Columbus, Ohio, http://www.dispatch.com/content/stories/business/2014/08/03/id-thieves-these-days-want-your-number-not-your-name.html, Aug. 3, 2014 in 2 pages.
Identity Theft Resource Center; Fact Sheet 120 A—To Order a Credit Report for a Child; Fact Sheets, Victim Resources; Apr. 30, 2007.
“Identity Thieves Beware: Lifelock Introduces Nation's First Guaranteed Proactive Solution to Identity Theft Protection,” PR Newswire, New York, Jun. 13, 2005 http://proquest.umi.com/pqdweb?did=852869731&sid=1&Fmt=3&clientId=19649&RQT=309&Vname=PQD.
Ideon, Credit-Card Registry that Bellyflopped this Year, Is Drawing some Bottom-Fishers, The Wall Street Journal, Aug. 21, 1995, pp. C2.
“Industry News, New Technology Identifies Mortgage Fraud: Basepoint Analytics Launches FraudMark”, Inman News, American Land Title Association, Oct. 5, 2005, pp. 1.
Information Brokers of America, “Information Brokers of America Child Identity Theft Protection” http://web.archive.org/web/20080706135451/http://iboainfo.com/child-order.html as archived Jul. 6, 2008 in 1 page.
Information Brokers of America, “Safeguard Your Child's Credit”, http://web.archive.org/web/20071215210406/http://www.iboainfo.com/child-id-protect.html as archived Dec. 15, 2007 in 1 page.
Intelius, “People Search—Updated Daily, Accurate and Fast!” http://www.intelius.com/people-search.html?=&gclid=CJqZIZP7paUCFYK5KgodbCUJJQ printed Nov. 16, 2010 in 1 page.
International Search Report and Written Opinion for Application No. PCT/US2007/06070, dated Nov. 10, 2008.
Iovation, Device Identification & Device Fingerprinting, http://www.iovation.com/risk-management/device-identification printed Nov. 5, 2012 in 6 pages.
Jacob et al., A Case Study of Checking Account Inquiries and Closures in Chicago, The Center for Financial Services Innovation, Nov. 2006.
Khan, Muhammad Khurram, PhD., “An Efficient and Secure Remote Mutual Authentication Scheme with Smart Cards” IEEE International Symposium on Biometrics & Security Technologies (ISBAST), Apr. 23-24, 2008, pp. 1-6.
Lanubile, et al., “Evaluating Empirical Models for the Detection of High-Risk Components: Some Lessons Learned”, 20th Annual Software Engineering Workshop, Nov. 29-30, 1995, Greenbelt, Maryland, pp. 1-6.
Lee, W.A.; “Experian, on Deal Hunt, Nets Identity Theft Insurer”, American Banker: The Financial Services Daily, Jun. 4, 2003, New York, NY, 1 page.
Lefebvre et al., “A Robust Soft Hash Algorithm for Digital Image Signature”, International Conference on Image Processing 2:11 (ICIP), vol. 3, Oct. 2003, pp. 495-498.
Leskovec, Jure, “Social Media Analytics: Tracking, Modeling and Predicting the Flow of Information through Networks”, WWW 2011—Tutorial, Mar. 28-Apr. 1, 2011, Hyderabad, India, pp. 277-278.
Letter to Donald A. Robert from Carolyn B. Maloney, dated Oct. 31, 2007, pp. 2.
Letter to Donald A. Robert from Senator Charles E. Schumer, dated Oct. 11, 2007, pp. 2.
Letter to Harry C. Gambill from Carolyn B. Maloney, dated Oct. 31, 2007, pp. 2.
Letter to Harry C. Gambill from Senator Charles E. Schumer, dated Oct. 11, 2007, pp. 2.
Letter to Richard F. Smith from Carolyn B. Maloney, dated Oct. 31, 2007, pp. 2.
Letter to Richard F. Smith from Senator Charles E. Schumer, dated Oct. 11, 2007, pp. 2.
Li et al., “Automatic Verbal Information Verification for User Authentication”, IEEE Transactions on Speech and Audio Processing, vol. 8, No. 5, Sep. 2000, pp. 585-596.
LifeLock, “How LifeLock Works,” http://www.lifelock.com/lifelock-for-people printed Mar. 14, 2008 in 1 page.
LifeLock, “LifeLock Launches First ID Theft Prevention Program for the Protection of Children,” Press Release, Oct. 14, 2005, http://www.lifelock.com/about-us/press-room/2005-press-releases/lifelock-protection-for-children.
LifeLock; “How Can LifeLock Protect My Kids and Family?” http://www.lifelock.com/lifelock-for-people/how-we-do-it/how-can-lifelock-protect-my-kids-and-family printed Mar. 14, 2008 in 1 page.
LifeLock, Various Pages, www.lifelock.com/, 2007.
Lobo, Jude, “MySAP.com Enterprise Portal Cookbook,” SAP Technical Delivery, Feb. 2002, vol. 1, pp. 1-13.
Magid, Lawrence, J., Business Tools: When Selecting an ASP Ensure Data Mobility, Los Angeles Times, Los Angeles, CA, Feb. 26, 2001, vol. C, Issue 4, pp. 3.
Manilla, http://www.manilla.com/how-it-works/ printed Feb. 5, 2014 in 1 page.
Meyers et al., “Using Your Social Networking Accounts To Log Into NPR.org,” NPR.org, Jun. 24, 2010, http://web.archive.org/web/20100627034054/http://www.npr.org/blogs/inside/2010/06/24/128079309/using-your-social-networking-accounts-to-log-into-npr-org in 3 pages.
Micarelli et al., “Personalized Search on the World Wide Web,” The Adaptive Web, LNCS 4321, 2007, pp. 195-230.
Microsoft, “Expand the Reach of Your Business,” Microsoft Business Solutions, 2004, in 16 pages.
Mint.com, http://www.mint.com/how-it-works/ printed Feb. 5, 2013 in 2 pages.
Mvelopes, http://www.mvelopes.com/ printed Feb. 5, 2014 in 2 pages.
My Call Credit http://www.mycallcredit.com/products.asp?product=ALR dated Dec. 10, 2005 on www.archive.org.
My Call Credit http://www.mycallcredit.com/rewrite.asp?display=faq dated Dec. 10, 2005 on www.archive.org.
My ID Alerts, “Why ID Alerts” http://www.myidalerts.com/why-id-alerts.jsps printed Apr. 3, 2012 in 2 pages.
My ID Alerts, “How it Works” http://www.myidalerts.com/how-it-works.jsps printed Apr. 3, 2012 in 3 pages.
“Name Availability Records”, Westlaw Database Directory, http://directory.westlaw.com/scope/default.asp?db=NA-ALL&RS=W..&VR=2.0 as printed Dec. 17, 2009, pp. 5.
National Alert Registry Launches RegisteredOffendersList.org to Provide Information on Registered Sex Offenders, May 16, 2005, pp. 2, http://www.prweb.com/printer/240437.htm accessed on Oct. 18, 2011.
National Alert Registry Offers Free Child Safety “Safe From Harm” DVD and Child Identification Kit, Oct. 24, 2006. pp. 2, http://www.prleap.com/pr/53170 accessed on Oct. 18, 2011.
National Alert Registry website titled, “Does a Sexual offender live in your neighborhood”, Oct. 22, 2006, pp. 2, http://web.archive.org/wb/20061022204835/http://www.nationallertregistry.com/ accessed on Oct. 13, 2011.
Next Card: About Us, http://web.cba.neu.edu/˜awatson/NextCardCase/NextCardAboutUs.htm printed Oct. 23, 2009 in 10 pages.
Ogg, Erica, “Apple Cracks Down on UDID Use”, http://gigaom.com/apple/apple-cracks-down-on-udid-use/ printed Nov. 5, 2012 in 5 Pages.
Pagano, et al., “Information Sharing in Credit Markets,” Dec. 1993, The Journal of Finance, vol. 48, No. 5, pp. 1693-1718.
Partnoy, Frank, Rethinking Regulation of Credit Rating Agencies: An Institutional Investor Perspective, Council of Institutional Investors, Apr. 2009, pp. 21.
Paustian, Chuck, “Every Cardholder a King Customers get the Full Treatment at Issuers' Web Sites,” Card Marketing, New York, Mar. 2001, vol. 5, No. 3, pp. 4.
People Finders, http://www.peoplefinders.com/?CMP=Google&utm_source=google&utm_medium=cpc printed Nov. 16, 2010 in 1 page.
People Lookup, “Your Source for Locating Anyone!” www.peoplelookup.com/people-search.html printed Nov. 16, 2010 in 1 page.
People Search, “The Leading Premium People Search Site on the Web,” http://www.peoplesearch.com printed Nov. 16, 2010 in 2 pages.
PersonalCapital.com, http://www.personalcapital.com/how-it-works printed Feb. 5, 2014 in 5 pages.
Press Release—“Helping Families Protect Against Identity Theft—Experian Announces FamilySecure.com; Parents and guardians are alerted for signs of potential identity theft for them and their children; product features an industry-leading $2 million guarantee”; PR Newswire; Irvine, CA; Oct. 1, 2007.
Privacy Rights Clearinghouse, “Identity Theft: What to do if it Happens to You,” http://web.archive.org/web/19990218180542/http://privacyrights.org/fs/fs17a.htm printed Feb. 18, 1999.
Ramaswamy, Vinita M., Identity-Theft Toolkit, The CPA Journal, Oct. 1, 2006, vol. 76, Issue 10, pp. 66-70.
Rawe, Julie; “Identity Thieves”, Time Bonus Section, Inside Business, Feb. 2002, pp. 2.
Roth, Andrew, “CheckFree to Introduce E-Mail Billing Serving,” American Banker, New York, Mar. 13, 2001, vol. 166, No. 49, pp. 3.
SAS, “SAS® Information Delivery Portal”, Fact Sheet, 2008, in 4 pages.
Scholastic Inc.:Parent's Request for Information http://web.archive.org/web/20070210091055/http://www.scholastic.com/inforequest/index.htm as archived Feb. 10, 2007 in 1 page.
Scholastic Inc.:Privacy Policy http://web.archive.org/web/20070127214753/http://www.scholastic.com/privacy.htm as archived Jan. 27, 2007 in 3 pages.
Securities and Futures Commission, “Guideline on Anti-Money Laundering and Counter-Terrorist Financing”, Jul. 2012, pp. 135.
Singletary, Michelle, “The Littlest Victims of ID Theft”, The Washington Post, The Color of Money, Oct. 4, 2007.
Sumner, Anthony, “Tackling The Issue of Bust-Out Fraud”, Retail Banker International, Jul. 24, 2007, pp. 4.
Sun, Hung-Min, “An Efficient Remote Use Authentication Scheme Using Smart Cards”, IEEE Transactions on Consumer Electronics, Nov. 2000, vol. 46, No. 4, pp. 958-961.
Target, “Free Credit Monitoring and Identity Theft Protection with Experian's ProtectMyID Now Available”, Jan. 13, 2014, pp. 2. http://corporate.target.com.
“TransUnion—Child Identity Theft Inquiry”, TransUnion, http://www.transunion.com/corporate/personal/fraudIdentityTheft/fraudPrevention/childIDInquiry.page as printed Nov. 5, 2009 in 4 pages.
Truston, “Checking if your Child is an ID Theft Victim can be Stressful,” as posted by Michelle Pastor on Jan. 22, 2007 at http://www.mytruston.com/blog/credit/checking_if_your_child_is_an_id_theft_vi.html.
US Legal, Description, http://www.uslegalforms.com/us/US-00708-LTR.htm printed Sep. 4, 2007 in 2 pages.
Vamosi, Robert, “How to Handle ID Fraud's Youngest Victims,” Nov. 21, 2008, http://news.cnet.com/8301-10789_3-10105303-57.html.
Waggoner, Darren J., “Having a Global Identity Crisis,” Collections & Credit Risk, Aug. 2001, vol. vol. 6, No. 8, pp. 6.
Wang et al., “User Identification Based on Finger-vein Patterns for Consumer Electronics Devices”, IEEE Transactions on Consumer Electronics, May 2010, vol. 56, No. 2, pp. 799-804.
WhatIs.com, “Risk-Based Authentication (RBA)”, https://web.archive.org/web/20121025033106/http://whatis.techtarget.com/definition/risk-based-authentication-RBA, Oct. 23, 2012, pp. 1.
Yahoo! Search, “People Search,” http://people.yahoo/com printed Nov. 16, 2010 in 1 page.
YODLEE | Money Center, https://yodleemoneycenter.com/ printed Feb. 5, 2014 in 2 pages.
You Need A Budget, http://www.youneedabudget.com/features printed Feb. 5, 2014 in 3 pages.
Cheng, Fred, “Security Attack Safe Mobile and Cloud-based One-time Password Tokens Using Rubbing Encryption Algorithm”, MONET, 2011, vol. 16, pp. 304-336.
“Fraud Alert | Learn How”. Fight Identity Theft. http://www.fightidentitytheft.com/flag.html, accessed on Nov. 5, 2009.
Gordon et al., “Using Identity Authentication and Eligibility Assessment to Mitigate the Risk of Improper Payments”, LexisNexis, Jan. 28, 2008, pp. 18. https://risk.lexisnexis.com/-/media/files/government/white-paper/identity_authentication-pdf.pdf.
LifeLock, “Personal Identity Theft Protection & Identity Theft Products,” http://www.lifelock.com/lifelock-for-people, accessed Nov. 5, 2007.
TheMorningCall.Com, “Cheap Ways to Foil Identity Theft,” www.mcall.com/business/columnists/all-karp.5920748jul01,0..., published Jul. 1, 2007.
Related Publications (1)
Number Date Country
20190259030 A1 Aug 2019 US
Provisional Applications (1)
Number Date Country
61790321 Mar 2013 US
Continuations (2)
Number Date Country
Parent 15459927 Mar 2017 US
Child 16222016 US
Parent 14208595 Mar 2014 US
Child 15459927 US