USER IDENTIFICATION WITH INPUT PROFILE RECORD

Information

  • Patent Application
  • 20220103548
  • Publication Number
    20220103548
  • Date Filed
    September 30, 2021
    3 years ago
  • Date Published
    March 31, 2022
    2 years ago
Abstract
User identification with an input profile record (IPR). In one embodiment, a server includes a memory and an electronic processor. The electronic processor is configured to receive a plurality of input profile records (IPRs) associated with a first user, the plurality of IPRs each based on a plurality of user inputs and indicative of identity of the first user, control the memory to store the plurality of IPRs in the input profile record repository, receive a current IPR associated with a second user, determine whether the second user is the first user by comparing a first one or more biometric features based on the plurality of IPRs and a second one or more biometric features based on the current IPR, and responsive to determining that the second user is the first user, output an identity confirmation that the second user is the first user.
Description
FIELD

The present disclosure relates generally to user identification. More specifically, the present disclosure relates to user identification with an input profile record.


BACKGROUND

Conventionally, user identification occurs in a variety of different ways. For example, a user may be identified with individual or combinations of distinctive biometrics that are associated with the user. In a different example, a user may be identified after receiving a one-time password to a registered user device associated with the user.


SUMMARY

However, several problems exist with conventional user identification. One problem is that conventional identification only occurs at certain points in time (e.g., turning on a smartphone). Another problem is that conventional biometric identification is fixed to the initial biometric used to set up the user identification.


The present disclosure improves upon the conventional user identification and solves the aforementioned problems by performing user identification with an input profile record (IPR). The input profile record is based on a plurality of user inputs of a user and the input profile record changes over time. The input profile record may then be continuously used to identify the user's use of any device over time. Further, the addition of IPR events like key-up, and mobile sensors (i.e. acceleration, orientation and rotation etc.), derivation of biometric features from the generated IPRs, and identifying the “right” balance between IPR size, sampling frequency resolution and effectiveness of data capture are all improvements over the conventional user identification.


One example of the present disclosure includes a server for user identification. The server includes a memory and an electronic processor in communication with the memory. The memory including an input profile record repository. The electronic processor is configured to receive a plurality of input profile records (IPRs) associated with a first user, the plurality of input profile records each based on a plurality of user inputs and indicative of identity of the first user, control the memory to store the plurality of IPRs in the input profile record repository, receive a current IPR associated with a second user, determine whether the second user is the first user by comparing a first one or more biometric features based on the plurality of IPRs and a second one or more biometric features based on the current IPR, and responsive to determining that the second user is the first user, output an identity confirmation that the second user is the first user.


Another example of the present disclosure includes a method for user identification. The method includes receiving, with the electronic processor, a plurality of input profile records (IPRs) associated with a first user, the plurality of input profile records each based on a plurality of user inputs and indicative of identity of the first user. The method includes controlling, with the electronic processor, a memory to store the plurality of IPRs in an input profile record repository. The method includes receiving, with the electronic processor, a current IPR associated with a second user. The method includes determining, with the electronic processor, whether the second user is the first user by comparing a first one or more biometric features based on the plurality of IPRs and a second one or more biometric features based on the current IPR. The method also includes responsive to determining that the second user is the first user, outputting, with the electronic processor, an identity confirmation that the second user is the first user.


Yet another example of the present disclosure includes a system. The system includes a user interface device and a server. The user interface device is configured to output a plurality of input profile records (IPRs) associated with a first user, the plurality of input profile records each based on a plurality of user inputs and indicative of identity of the first user. The server includes a memory including an input profile record repository and an electronic processor in communication with the memory. The electronic processor is configured to receive the plurality of IPRs, control the memory to store the plurality of IPRs in the input profile record repository, receive a current IPR associated with a second user, determine whether the second user is the first user by comparing a first one or more biometric features based on the plurality of IPRs and a second one or more biometric features based on the current IPR, and responsive to determining that the second user is the first user, output an identity confirmation that the second user is the first user.





BRIEF DESCRIPTION OF THE DRAWINGS


FIG. 1 is a block diagram illustrating a system with user identification based on an input profile record, in accordance with various aspects of the present disclosure.



FIG. 2 is a block diagram illustrating a second system with user identification based on an input profile record, in accordance with various aspects of the present disclosure.



FIG. 3 is a flowchart illustrating a method for identifying a user, in accordance with various aspects of the present disclosure.



FIG. 4 is a diagram illustrating an example of an input profile record (IPR), in accordance with various aspects of the present disclosure.



FIG. 5 is a diagram illustrating a second example of the IPR, in accordance with various aspects of the present disclosure.



FIG. 6 is a diagram illustrating a first example of a dwell time feature, in accordance with various aspects of the present disclosure.



FIG. 7 is a diagram illustrating four different latency times, in accordance with various aspects of the present disclosure.



FIG. 8 is a diagram illustrating different latency times for a portion of an example OTP “356024,” in accordance with various aspects of the present disclosure.



FIG. 9 is a block diagram illustrating a standard number pad layout, in accordance with various aspects of the present disclosure.



FIG. 10 is block diagram illustrating a standard number row layout, in accordance with various aspects of the present disclosure.



FIG. 11 is diagram illustrating different categories of distances between number positions in the standard number pad layout of FIG. 9, in accordance with various aspects of the present disclosure.





DETAILED DESCRIPTION

Before any embodiments of the present disclosure are explained in detail, it is to be understood that the present disclosure is not limited in its application to the details of construction and the arrangement of components set forth in the following description or illustrated in the following drawings. The present disclosure is capable of other embodiments and of being practiced or of being carried out in various ways.



FIG. 1 is a block diagram illustrating a system 10 with user identification based on an input profile record, in accordance with various aspects of the present disclosure. It should be understood that, in some embodiments, there are different configurations from the configuration illustrated in FIG. 1. The functionality described herein may be extended to any number of servers providing distributed processing.


In the example of FIG. 1, the system 10 includes a server 100, a user interface device 120, and a network 180. The server 100 includes an electronic processor 102 (for example, a microprocessor or another suitable processing device), a memory 104 (for example, a non-transitory computer-readable storage medium), and a communication interface 112. It should be understood that, in some embodiments, the server 100 may include fewer or additional components in configurations different from that illustrated in FIG. 1. Also, the server 100 may perform additional functionality than the functionality described herein. In addition, the functionality of the server 100 may be incorporated into other servers. As illustrated in FIG. 1, the electronic processor 102, the memory 104, and the communication interface 112 are electrically coupled by one or more control or data buses enabling communication between the components.


The electronic processor 102 executes machine-readable instructions stored in the memory 104. For example, the electronic processor 102 may execute instructions stored in the memory 104 to perform the functionality described herein.


The memory 104 may include a program storage area (for example, read only memory (ROM)) and a data storage area (for example, random access memory (RAM), and other non-transitory, machine-readable medium). In some examples, the program storage area may store machine-executable instructions regarding an input profile record (IPR) program 106. In some examples, the data storage area may store data regarding an input profile record repository 108.


The IPR program 106 causes the electronic processor 102 to collect and store input profile records in the input profile record repository 108. Specifically, the IPR program 106 causes the electronic processor 102 to parse the IPR content received from a user interface device, determine biometric features based on the current IPR and historical/older IPRs associated with the user, and perform user identification using a biometric identification algorithm that compares current biometrics features based on a current IPR to the historical biometric features based on a set of historical IPRs. In some examples, a successful user identification may require ten historical IPRs associated with the user to establish a “user profile.”


The IPR program 106 also causes the electronic processor 102 to update an input profile record stored in the input profile record repository 108. Additionally, the user identification with the IPRs is a “passive” identification that does not need to query a user for additional information.


In examples, the input profile record repository 108 is a central repository including a plurality of input profile records. Each input profile record is associated with a specific user (e.g., a user account) and/or a specific user interface device. An input profile record stored in the input profile record repository 108 is updated periodically with the IPR program 106 as described above. The input profile record associated with the user interface device 120 is indicative of an identity of a user over a specific period of time. In other words, the input profile record as described herein solves the aforementioned problems with user identification because the input profile record is a dynamic identification of a user over a specific period of time rather than occurring at certain points in time and fixed to an initial biometric used to set up the user identification.


For example, the biometric algorithm of the IPR program 106 includes a number of typing and sensor behavioral features as set forth in Tables 4-7 (also referred to as “biometric features”) from the user inputs set forth in Tables 1-3 and 8 (i.e., events included in the IPR data construct). The maximum available sample rate (or data delay) is 16 milliseconds (ms), which means sensor data is recorded every 16 ms. However, as with a sample rate of 16 ms from the sensors, a size of the IPR exceeds an upload size threshold set forth in Appendix D (e.g., an upload size threshold of 20,000 bytes). Additionally, as described in Appendix D, the size of the IPR may be reduced below the upload size threshold by increasing the sample rate of some or all of the sensors (e.g., an increase to every 100 ms and/or an increase to every 50 ms), which means a balance between a lower size of recorded data (e.g., the IPR) with lower frequency, less accuracy, and a lower number of samples from some or all of the sensors.


The communication interface 112 receives data from and provides data to devices external to the server 100, such as an input profile record (IPR) from the user interface device 120 via the network 180. For example, the communication interface 112 may include a port or connection for receiving a wired connection (for example, an Ethernet cable, fiber optic cable, a telephone cable, or the like), a wireless transceiver, or a combination thereof. In some examples, the network 180 is the Internet.


In the example of FIG. 1, the user interface device 120 includes an electronic processor 122 (for example, a microprocessor or another suitable processing device), a memory 124 (for example, a non-transitory computer-readable storage medium), a communication interface 132, a camera 134, and a presence-sensitive display 136. In some examples, the user interface device may be a smartphone, tablet, laptop, or other suitable user interface device with a presence-sensitive display. As illustrated in FIG. 1, the electronic processor 122, the memory 124, the communication interface 132, the camera 134, and the presence-sensitive display 136 are electrically coupled by one or more control or data buses enabling communication between the components.


The electronic processor 122 executes machine-readable instructions stored in the memory 124. For example, the electronic processor 122 may execute instructions stored in the memory 124 to perform the functionality described herein.


The memory 124 may include a program storage area (for example, read only memory (ROM)) and a data storage area (for example, random access memory (RAM), and other non-transitory, machine-readable medium). The program storage area includes a user input collection and input profile record (IPR) application 126. In some examples, the user input collection and IPR application 126 may be a standalone application. In other examples, the user input collection and IPR application 126 is a feature that is part of a separate application (e.g., the user input collection and IPR application 126 may be included as part of a camera application, a banking application, or other suitable application).


The user input collection and IPR application 126 causes the electronic processor 122 to collect user inputs, i.e., user interactions, from a user relative to a mobile application (e.g., time to fill data field entries, use of specific autofill, or other suitable user inputs) of the user interface device 120 and generate an input profile record (IPR) based on the user inputs (also referred to as a “a mobile platform”). The user input collection and IPR program 106 may also cause the electronic processor 122 to collect user inputs at a particular website (e.g., time to fill data field entries, use of specific autofill, or other suitable user inputs) and generate (or update) the input profile record based on these user inputs (also referred to as a “web platform”).


In some examples, the user input collection and IPR application 126 causes the electronic processor 122 to collect user inputs with respect to the presence-sensitive display 136 (e.g., type of keyboard, typing speed, use of patterns, or other suitable user inputs (see Tables 1-3)). In these examples, the user input collection and IPR application 126 may also cause the electronic processor 122 to output the generated IPR to the server 100 via the communication interface 132 and the network 180. Additionally, in some examples, the user input collection and IPR application 126 may cause electronic processor 122 to control the memory 124 to store the user inputs that are collected and/or the IPR that is generated for a period of time or until the generated IPR is output to the server 100.


In other examples, the user input collection and IPR application 126 causes the electronic processor 122 to collect user inputs with respect to the camera 134 (e.g., facial recognition, user gestures, or other suitable user inputs, which may be part of the mobile platform. In these examples, the user input collection and IPR application 126 may also cause the electronic processor 122 to generate (or update) an IPR based on the aforementioned user inputs and output the IPR to the server 100 via the communication interface 132 and the network 180. Additionally, in some examples, the user input collection and IPR application 126 may cause electronic processor 122 to control the memory 124 to store the user inputs that are collected and/or the IPR that is generated for a period of time or until the generated IPR is output to the server 100.


The communication interface 132 receives data from and provides data (e.g., generated IPR(s)) to devices external to the user interface device 120, i.e., the server 100. For example, the communication interface 132 may include a port or connection for receiving a wired connection (for example, an Ethernet cable, fiber optic cable, a telephone cable, or the like), a wireless transceiver, or a combination thereof.


The camera 134 includes an image sensor that generates and outputs image data of a subject. In some examples, the camera 134 includes a semiconductor charge-coupled device (CCD) image sensor, a complementary metal-oxide-semiconductor (CMOS) image sensor, or other suitable image sensor. The electronic processor 122 receives the image data of the subject that is output by the camera 134.


The presence-sensitive display 136 includes a display screen with an array of pixels that generate and output images. In some examples, the display screen is one of a liquid crystal display (LCD) screen, a light-emitting diode (LED) and liquid crystal display (LCD) screen, a quantum dot light-emitting diode (QLED) display screen, an interferometric modulator display (IMOD) screen, a micro light-emitting diode display screen (mLED), a virtual retinal display screen, or other suitable display screen. The presence-sensitive display 136 also includes circuitry that is configured to detect the presence of the user. In some examples, the circuitry is a resistive or capacitive panel that detects the presence of an object (e.g., a user's finger).


It should be understood that, in some embodiments, the server 100 may include fewer or additional components in configurations different from that illustrated in FIG. 1. Also, the server 100 may perform additional functionality than the functionality described herein. In addition, some of the functionality of the user interface device 120 (for example, the IPR generation) may be incorporated into other servers (e.g., incorporated into the server 100). Likewise, some of the functionality of the server 100 may be incorporated into the user interface device 120 (for example, the user identification).


To summarize, the user interface device 120 collects IPR data for each transaction at a mobile application or at a web page. From the raw IPR data, the server 100 may parse out a set of meaningful biometric features that differentiates same users from different users.


A passive biometric identification algorithm included in the IPR program 106 compares biometric feature values (from current IPR) to biometric feature values seen in the past (from historical IPRs), and when the current biometric feature values fall within a “reasonable” range of what is seen in the past, the server 100 may identify the user to be the same as a previous user. The passive biometric identification algorithm is an anomaly detection type of algorithm.


For the set of biometric feature values seen in the past IPRs, the set may be considered as a “training profile.” In general, a minimum of two to ten and a maximum of ten to fifteen (i.e. rolling window of last X transactions) IPRs may be required to build profiles for comparison with the biometric identification algorithm. Each biometric feature may also contribute a different weight to the overall model prediction, where a biometric feature with higher predictability power would have a higher weight.


To return the “different user” identification confirmation, the server 100 may determine whether a biometric score is less than a lower threshold. To return the “same user” identification confirmation, the server 100 may determine whether a biometric score is greater than an upper threshold and the lower threshold. To return the “undetermined” identification confirmation, the server 100 may determine whether a biometric score is greater than the lower threshold and less than the upper threshold.


In some examples, the biometric identification algorithm returns a biometric score between 0 to 1, where closer to 1 means more likely a match. Additionally, in some examples, the upper and lower thresholds are set based on feedback data (i.e. confirmed fraudulent identifications) from clients such that the biometric identification algorithm accurately classifies all different users as no-matches to reduce or eliminate false positives.



FIG. 2 is a block diagram illustrating a second system 200 with user identification based on an input profile record, in accordance with various aspects of the present disclosure. It should be understood that, in some embodiments, there are different configurations from the configuration illustrated in FIG. 2. The functionality described herein may be extended to any number of servers providing distributed processing.


In the example of FIG. 2, the system 200 includes the server 100 as described above in FIG. 1 and a user interface device 220. Consequently, the description of the server 100 is not repeated below to avoid redundant descriptions. Additionally, the user interface device 220 is any electronic device that user may use to interface with the server 100. For example, the user interface device 200 may be a mouse, a keyboard, a desktop computer, or other suitable user interface device.


In the example of FIG. 2, the user interface device 220 includes an electronic processor 222 (for example, a microprocessor or another suitable processing device), a memory 224 (for example, a non-transitory computer-readable storage medium), and a communication interface 232.


It should be understood that, in some embodiments, the user interface device 220 may include fewer or additional components in configurations different from that illustrated in FIG. 2. Also, the user interface device 220 may perform additional functionality than the functionality described herein. As illustrated in FIG. 2, the electronic processor 222, the memory 224, and the communication interface 232 are electrically coupled by one or more control or data buses enabling communication between the components.


The electronic processor 222 executes machine-readable instructions stored in the memory 224. For example, the electronic processor 222 may execute instructions stored in the memory 224 to perform the functionality described herein.


The memory 224 may include a program storage area (for example, read only memory (ROM)) and a data storage area (for example, random access memory (RAM), and other non-transitory, machine-readable medium). For example, when the user interface device 220 is a desktop computer, the program storage area may include a user input collection and input profile record (IPR) application 226 that is similar to the user input collection and IPR application 126 as described above.


The communication interface 232 receives data from (e.g., IPR generation signal) and provides data (e.g., generated IPR(s)) to devices external to the user interface device 220, i.e., the server 100. For example, the communication interface 232 may include a port or connection for receiving a wired connection (for example, an Ethernet cable, fiber optic cable, a telephone cable, a universal serial bus (USB) cable, or other suitable wired connection), a wireless transceiver, or a combination thereof.


In the example of FIG. 2, the server 100 may send a command (e.g., the IPR generation signal) to the user interface device 220 to collect user input(s) from a user's interaction with the user interface device 220 for a specific period of time. For example, when the user interface device 220 is a computer mouse, the server 100 may send a command to the computer mouse to collect user input(s) from the user's interaction with the computer mouse for a specific period of time.


In the example of FIG. 2, the user input collection and IPR application 226 may also cause the electronic processor 222 to generate (or update) an IPR based on the aforementioned user input(s) and output the IPR to the server 100 via the communication interface 232 and the network 180. Additionally, in some examples, the user input collection and IPR application 226 may cause electronic processor 222 to control the memory 224 to store the user input(s) that are collected and/or the IPR that is generated for a period of time or until the generated IPR is output to the server 100.



FIG. 3 is a flowchart illustrating a method 300 for identifying a user, in accordance with various aspects of the present disclosure. FIG. 3 is described with respect to the server 100 and the user interface device 120 of FIG. 1. However, FIG. 3 is equally applicable to the server 100 and the user interface device 220 of FIG. 2, although the server 100 controls the user interface device 220 to collect user inputs for a specific period of time.


The method 300 includes receiving, with an electronic processor, a plurality of input profile records (IPRs) associated with a first user, the plurality of IPRs are each based on a plurality of user inputs and are each indicative of an identity of the first user (at block 302). For example, the electronic processor 102 receives a plurality of input profile records associated with a first user, the plurality of input profile records are each based on the plurality of user inputs provided by the first user, and are each indicative of an identity of the first user of the user interface device 120.


The method 300 includes controlling, with the electronic processor, a memory to store the plurality of input profile records (IPRs) in an input profile record repository (at block 304). For example, the electronic processor 102 controls the memory 104 to store the IPRs that are received in the input profile record repository 108.


The method 300 includes receiving, with the electronic processor, a current input profile record (IPR) associated with a second user (at block 306). For example, the electronic processor 102 receives a current IPR associated with a current user of the user interface device 120 from the user interface device 120.


The method 300 includes determining, with the electronic processor and a biometric identification algorithm, whether the second user is the first user by comparing a first one or more biometric features based on the plurality of input profile records and a second one or more biometric features based on the current IPR (at block 308). For example, the electronic processor 102 determines whether the current user of the user interface device 120 is the first user of the user interface device 120 by comparing a first one or more biometric features based on the plurality of input profile records associated with the first user and a second one or more biometric features based on the current IPR associated with the second user.


The method 300 includes responsive to determining that the second user is the first user, outputting, with the electronic processor, an identity confirmation that the second user is the first user (at block 310). For example, the electronic processor 102 controls the communication interface 112 to output an identity confirmation that the current user of the user interface device 120 is the first user of the user interface device 120 to the user interface device 120 via the network 180 in response to the electronic processor 102 determining that the current user is the first user.


Alternatively, in some examples, the electronic processor 102 controls the communication interface 112 to output an identity confirmation that the current user of the user interface device 120 is the first user of the user interface device 120 to a second server or other computing device via the network 180 in response to the electronic processor 102 determining that the current user is the first user. In these examples, the second server or other computing device may have initiated the identification of the second user by requesting the server 100 to identify whether the first user is the second user.


In some examples, the current IPR may be from a second user interface device that is different from the user interface device. In these examples, the identity confirmation confirms the second user of the second user interface is the same as the first user of the user interface device.


Additionally, in some examples, in determining whether the second user is the first user by comparing the first one or more biometric features based on the plurality of IPRs and the second one or more biometric features based on the current IPR, the method 300 may further include generating, with a biometric identification algorithm, the first one or more biometric features from the plurality of IPRs, generating, with the biometric identification algorithm, the second one or more biometric features from the current IPR, generating, with the biometric identification algorithm, a biometric score based on difference between the second one or more biometric features and the first one or more biometric features, determining whether the biometric score is less than a lower threshold, determining whether the biometric score greater than the lower threshold and less than an upper threshold, and determining whether the biometric score is greater than the lower threshold and the upper threshold. In these examples, the second user is the first user when the biometric score is greater than the lower threshold and the upper threshold, the second user is not the first user when the biometric score is lower than the lower threshold and the upper threshold, and the second user is undetermined relative to the first user when the biometric score is higher than the lower threshold and lower than the upper threshold.


Additionally, in these examples, in generating, with the biometric identification algorithm, the first one or more biometric features from the plurality of IPRs and generating, with the biometric identification algorithm, the second one or more biometric features from the current IPR, the method 300 may further include determining a first one or more latencies of a first dwell time based on the plurality of IPRs, and determining a second one or more latencies of a second dwell time based on the current IPR.


In some examples, the plurality of IPRs and the current IPR may each include an IPR header and a plurality of IPR events. The plurality of IPR events includes a key down event and a key up event. The plurality of user inputs is a one-time-password (OTP) and each user input of the plurality of user inputs includes the key down event and the key up event associated with each key in the OTP.



FIG. 4 is a diagram illustrating an example of an input profile record 400, in accordance with various aspects of the present disclosure. The input profile record (IPR) 400 is a transport mechanism that collects and verifies an end user's device interactions and behaviors. Interactions related to the users are captured and the use of their keyboard, mouse, motion and other interaction behaviors that can be extracted from the end user's device. In a typical integration, the IPR 400 is sent to the platform for processing, profiling, analysis and verification.


The device interaction events may be captured, for example, using a JavaScript Widget or Native Mobile SDKs, by hooking into application and/or platform based event callbacks that are available and compiles them into a text based data structure as illustrated in the IPR 400 of FIG. 4. The text based data structure is composed mainly of individual events separated by a token and concatenated into a string. Each event type may have a variable number of parameters to capture the details of that event type. Each parameter within an event is also separated by another token or sentinel value.


The server-side parsers are built to support any combination of input events as long as the header event, described below, is present. This enables IPR parsing to be forward compatible such that the parser will not cause any errors when the parser sees event types that it does not support. These events will be logged as “Unknown Events” and execute no special parsing rules.


When split on the event separator token (semi-colon character), the IPR 400 expands into an IPR 500. FIG. 5 is a diagram illustrating a second example of the IPR 500, in accordance with various aspects of the present disclosure.









TABLE 1







IPR Header Event Details


The first event in the IPR 500 contains


header information in the format of










Index
Title
Description
Example Value













0
Encoding Type
Set to ncip for all 2.2
ncip




IPRs.


1
Reserved
Reserved field,
0




always zero


2
Unix Timestamp
The unix time in
538eb08a



(base 16)
seconds as recorded




at the initialization of




the JavaScript.




Represented as base




16.


3
Encoding Version
Encoding version.
3



(base 16)
Current versions are




1 and 2 and 3




(current)


4
Time Resolution
The number of
a



(base 16)
milliseconds in each




time interval. Default




is 10 (or ‘a’ in base




16).
















TABLE 2







IPR Common Event Details


All other events, other than the header event, follow the


base/common structure described in this table:










Index
Title
Type
Description





0
Event Type
string
An ID indicating the





event type (reference





the ID column in the





next table)


1
Time Since Last
string
The number of time



Event

intervals since the



(base 16)

last event. The Time





Resolution parameter





provided in the





header defines the





number of





milliseconds in each





time interval.


2 . . . N
Event Type
Mixed
Numbers are



Parameter 1 . . . N

represented as





base 16, otherwise





string.





1 . . . N denotes a





variable range of





possible Event Type





Parameters, where N





is the total number of





Event Type specific





parameters.










The following table describes each of the events and the associated data parameters they contain. These event specific parameters start after the “Time Since Last Event,” as mentioned above in Table 2.









TABLE 3







IPR Events










Identi-





fier
Event
Event Parameters
Description





st
Form
N pairs of:
Logged each time the IPR



State
1. DOM ID -
widget initializes in the end




Element ID/Name
user's browser.




of the target field




2. Length - The




current length of




the input element




when the state was




logged. These pairs




continue for each




input field that are




bound to and




recording IPR data




from


ff
Form
1. ID - Element
Sent when a user focuses



Field
ID/Name of the
an input field on the form.



Focus
target field


fb
Form
1. ID - Element
Sent when a user blurs



Field
ID/Name of the
(leaves focus) any type of



Blur
target field
HTML input field on the





form.


kd
Key down

Sent whenever a key down





occurs.


ku
Key up

Sent whenever a key up





occurs.


mm
Mouse
1. X - Horizontal
Sent at configurable



Move
position of the
frequency, providing




mouse
mouse position, in pixels,




2. Y - Vertical
relative to the top left of the




position of the
document area.




mouse
Default sample rate is





every 5 seconds.


mc
Mouse
1. X - Horizontal
Sent whenever the mouse is



Click
position of the
clicked on the page.




mouse




2. Y - Vertical




position of the




mouse




3. ID - Element




ID/Name that was




clicked


te
Touch
1. X - Horizontal
Sent whenever a touch start



Event
coordinate of the
event occurs on the page.




touch
When available, the X and




2. Y - Vertical
Y coordinate are the touch




coordinate of the
point relative to the




touch
viewport,




3. ID - Element
including any scroll offset.




ID/Name that was
These will be −1 if the




touched
touches page X and page Y





properties are unavailable.


ac
Acceler-

For devices with



ometer

accelerometer data.


fs
Form
1. X - Horizontal
A special event that is



Submit
coordinate of the
called before the post back




mouse
occurs. Called for both




2. Y - Vertical
‘Enter’ pressed and button




coordinate of the
click. Passes in the mouse




mouse
position at the time of





event.


kk
Total
1. Length - The
Triggers along with the



Keys
current length of
FormFieldFocus(ff) event.




the value of the
This field is not currently




element when it
used internally, but may be




was focused.
useful in the future so it has




2. ID -
been restored in this




ElementID/Name
encoding format document.




that was focused
This event type is still





currently active in the





JavaScript widget IPR.


sp
Scroll

Determine if the page was



Position

scrolled or not and if so





what position it's at on a





configurable frequency.


nccl
Control



List


ts
Time Sync
1. Now - Current
Logs a time sync every 15




time in MS
seconds




2. Delta - Time




since the IPR Init




in MS


mms
Mouse
1. Time Since Last
Mouse Movement Data is



Movement
MMS
cached any time the mouse



Sample
2. Number of sub
is moved and samples of




samples taken
the movement are taken on




3. NOP or
configurable frequencies.




minVelocityX
Samples are made up of a




minVelocityY **
number of sub-samples that




If the event is not a
collect and aggregate




“NOP” event:
movement data to keep the




4. maxVelocityX
payload small. The third




maxVelocityY **
(and last) parameter will




5. Average
hold the value “NOP” if no




Magnitude of
mouse activity was




Velocity
detected among the




6. Total Distance
collection of sub-samples




Moved
when a full sample is




7. Min
recorded.




Acceleration
If mouse movement




8. Max
activity is detected for at




Acceleration
least one sub-sample the




9. Average
full sample will be




Acceleration
populated and provided to





the IPR.





** Min and Max velocity





are expressed as vectors





separated by a space, since





the comma is used for





parameters.





All numbers are





represented as base 16 with





the decimal shifted 4 places





to the





right to preserve accuracy





of values below 0.





All numerical units are





expressed as Screens/





Second. Where a screen is





the size of the





window.screen.(width|height)





properties.


dms
Device
Same format as
The Device Motion Sample



Motion
mms, but with 3
uses the same format and



Sample
dimensions.
sampling implementation




Uses
as mms.




alpha/beta/gamma
Devices return Alpha as 0




instead of x/y.
to 360




3.
Alpha = DeviceAlpha




minVelocityAlpha
Devices return Beta




minVelocityBeta
as −180 to 180




minVelocityGamma
Beta = DeviceBeta + 180




4.
Devices return Gamma




maxVelocityAlpha
as −90 to 90




maxVelocityBeta
Gamma = DeviceGamma +




maxVelocityGamma
90





Note in iOS pitch, roll, and





yaw terminology is used by





CoreMotion. They correlate





as such:





alpha - yaw





beta - pitch





gamma - roll


dm
Device
1. DeviceAlpha
Sent at configurable



Motion
2. DeviceBeta +
frequency, providing




180
device motion, in




3. DeviceGamma +
alpha/beta/gamma notation.




90
Default sample rate is





every 5 seconds.


so
Stop

Indicates that IPR





recording was turned off





using the widget “stop”





function.


tr
Truncate
1. Length of
Indicates that a truncation




original IPR before
event has occurred. The




truncation
truncated IPR is appended





with a truncate event and





data contains the original





length of the IPR before it





was truncated.









The details above show no field identifier on key down events. The lack of a field identifier reduces the size of the IPR payload. When a form field focus event occurs, the following key down events are assumed to belong to that form field focus event. The parsing code however is set up to parse key down event entries that also contain an element name, for example, key down, bf (i.e., number of milliseconds since the last event in base 16), password. The key down events will contain the form identifier and so the behavior described above must be preserved if IPR parsing is changed.


Since key up event capture and enhanced key down profiling were added for both desktop and mobile IPRs, additional features could generally apply for both physical and touch keyboards, although there would be feature implementation differences based on differences in desktop IPR data and mobile IPR data. For example, FIG. 6 is a diagram illustrating a first example of a dwell time feature 600, in accordance with various aspects of the present disclosure.


The dwell time feature 600 is an amount of time during which a key (physical or software) remains in contact (down/up for physical and press/release for software) with a user. As illustrated in FIG. 6, the dwell time feature 600 includes a first down press 600, a first dwell time 602, a first up release 604, a second down press 606, a second dwell time 608, and a second up release 610. In the IPRs 400 and 500 described above, time elements are in time deltas (time since last event), rather than timestamps.









TABLE 4







Characteristics of Dwell Time Feature











Feature
Feature
Feature


#
Name
Description
Type













1
Each of these are
Dwell time for each
Numeric



individual features:
position (i.e., 1, 2, 3, 4, 5,



otp_otp_dwell_1
6) in the one-time



otp_otp_dwell_2
password (OTP) sequence.



otp_otp_dwell_3
A difference in



otp_otp_dwell_4
performance is noted when



otp_otp_dwell_5
the position of a digit



otp_otp_dwell_6
within an OTP sequence is




included for key press




durations and latencies.


2
total_dwell
Total dwell time for the
Numeric




OTP sequence (sum of




dwell times for all keys




pressed when inputting the




OTP)




SUM(otp_position1_dwell,




otp_position2_dwell,




otp_position3_dwell,




otp_position4_dwell,




otp_position5_dwell,




otp_position6_dwell)


3
total_dwell_to_dd
Proportion of total dwell
Numeric




time relative to total down-




down latency, (total_dwell −




otp_position6_dwell)/




total_dd_time




Since the last key does not




have an associated down-




down time, the last key is




excluded from the




calculation.


4
total_dwell_to_uu
Proportion of total dwell
Numeric




time relative to total up-up




latency. (total_dwell −




otp_position1_dwell)/




total_uu_time




Since the first key does not




have an associated up-up




time, the first key is




excluded from the




calculation.


5
mean_dwell
Average dwell time (of a
Numeric




single key) for the OTP




sequence


6
max_dwell
Longest dwell time in the
Numeric




OTP sequence


7
min_dwell
Shortest dwell time in the
Numeric




OTP sequence


8
std_dwell
Standard deviation of
Numeric




dwell times for the OTP




sequence


9
Each of these are
Proportion of dwell time
Numeric



individual features:
relative to down-down



dwell_to_dd1
latency (see “Latency”



dwell_to_dd2
section below



dwell_to_dd3
for definition) for each key



dwell_to_dd4
pressed.* For example:



dwell_to_dd5
otp_position1_dwell/dd1




*Since the last key pressed




does not have an




associated down-down




latency, there are only 5




down-down latencies in a




6-digit OTP, so the last




key pressed would not




have this feature.


10
mean_dwell_to_dd
Average proportion of
Numeric




dwell time relative to




down-down latency across




all keys




pressed.




AVG(dwell_to_dd1,




dwell_to_dd2,




dwell_to_dd3,




dwell_to_dd4,




dwell_to_dd5)


11
std_dwell_to_dd
Standard deviation of
Numeric




proportion of dwell time




relative to down-down




latency across




all keys pressed.




STD(dwell_to_dd1,




dwell_to_dd2,




dwell_to_dd3,




dwell_to_dd4,




dwell_to_dd5)


12
Each of these are
Proportion of dwell time
Numeric



individual features:
relative to up-up latency



dwell_to_uu1
(see “Latency” section



dwell_to_uu2
below for definition) for



dwell_to_uu3
each key pressed.* For



dwell_to_uu4
example:



dwell_to_uu5
otp_position2_dwell/uul




*Since the first key




pressed does not have an




associated up-up latency,




there are only 5 up-up




latencies in a 6-digit OTP,




so the first key pressed




would not have this




feature.


13
mean_dwell_to_uu
Average proportion of
Numeric




dwell time relative to up-




up latency across all keys




pressed.




AVG(dwell_to_uu1,




dwell_to_uu2,




dwell_to_uu3,




dwell_to_uu4,




dwell_to_uu5)


14
std_dwell_to_uu
Standard deviation of
Numeric




proportion of dwell time




relative to up-up latency




across all keys pressed.




STD(dwell_to_uu1,




dwell_to_uu2,




dwell_to_uu3,




dwell_to_uu4,




dwell_to_uu5)









Another aspect of the dwell time feature 600 is latency, which is an amount of time between consecutive keystrokes, where keystroke is a pair of key events involving a press and release of a single key. Latency may be broken into four different types: 1) Down-Down, 2) Up-Down, 3) Up-Up, and 4) Down-Up. FIG. 7 is a diagram illustrating four different latency times 700-706, in accordance with various aspects of the present disclosure. The first latency 700 is the down-down latency that is the amount of time between pressing a key and pressing the next key. The second latency 702 is the down-up latency that is the amount of time between pressing a key and releasing the next key. The third latency 704 is the up-down latency (also known as “Flight Time”) that is the amount of time between releasing a key and pressing the next key. The fourth latency 706 is the up-up latency that is the amount of time between releasing a key and releasing the next key.


Generally, dwell time is positive because keystrokes follow a down-up-down-up pattern. However, in some instances, dwell time may be negative when the sequence of keystrokes does not follow the down-up-down-up pattern (for example, due to fast typing or use of shift keys).


For the example OTP “356024,” the server 100 may determine each type of latency time for all diagraphs (a diagraph being two consecutive keystrokes). FIG. 8 is a diagram illustrating different latency times 800-806 for a portion of an example OTP “356024,” in accordance with various aspects of the present disclosure. As illustrated in FIG. 8, the portion of the example OTP “356024” is “3560” and includes down-down latencies 800A-800D, down-up latencies 802A-802C, up-down latencies 804A-804C, and up-up latencies 806A-806C.









TABLE 5







Dwell Time Latency Features












Feature
Feature
Feature



#
Name
Description
Type
Consideration














1
Down-Down (dd)
Amount of time between
Numeric




latency.
pressing a key and pressing the



Each of these are
next key for each digraph in the



individual features:
OTP sequence.



dd1
With the example OTP 356024,



dd2
there are 5 total digraphs where



dd3
each digraph corresponds to a



dd4
transition between the following



dd5
pairs of keys:




Digraph 1: (3, 5)




Digraph 2: (5, 6)




Digraph 3: (6, 0)




Digraph 4: (0, 2)




Digraph 5: (2, 4)


2
Up-Up (uu) latency. Each
Amount of time between
Numeric



of these are individual
releasing a key and releasing the



features:
next key for each digraph in the



uu1
OTP sequence



uu2



uu3



uu4



uu5


3
Up-Down (ud) latency.
Amount of time between
Numeric



Each of these are
releasing a key and pressing the



individual features:
next key for each digraph in the



ud1
OTP sequence



ud2



ud3



ud4



ud5


4
Down-Up (du) latency
Amount of time between
Numeric



Each of these are
pressing a key and releasing the



individual features:
next key for each digraph in the



du1
OTP sequence



du2



du3



du4



du5


5
total_x_time
Total dd, uu, ud, or du time in
Numeric
Error



where x in [dd, uu, ud,
the OTP sequence

corrections



du]


will make this






a larger






number - for






all features, it






would be






simpler to only






use samples






where OTP






was inputted






without any






error


6
mean_x_time
Average dd, uu, ud or du time in
Numeric



where x in [dd, uu, ud,
the OTP sequence



du]


7
min_x_time where x in
Shortest dd, uu, ud or du time in
Numeric



[dd, uu, ud, du]
the OTP sequence


8
max_x_time
Longest dd, uu, ud or du time in
Numeric



where x in [dd, uu, ud,
the OTP sequence



du]


9
std_x_time
Standard deviation of dd, uu, ud
Numeric



where x in [dd, uu, ud,
or du times in the OTP sequence



du]


10
otp_position_max_ud_time
The position in the OTP
Numeric -
Error




sequence of the key which
discrete
corrections (if




precedes the longest up-down
(range
made) might




(flight) time (e.g. longest pause
1-6)
make it trickier




comes after the 2nd digit is

to determine




typed) and may be further

the position




extended to max 1, max 2, max




3 . . . i.e. position of key preceding




longest flight time, position of




key preceding second longest




flight time, etc. ← This may




help characterize how users have




different rhythms when




inputting an OTP (e.g. 3 + 3 =




type first 3 digits, small pause,




then types next 3 digits - other




patterns like 2 + 2 + 2 are also




possible)









In some examples, the actual OTP assigned may be known in advance and whether the OTP typed was correct/accepted. In these examples, the location/layout structure of the keyboard gives rise to three additional latency features: 1) latency for specific pairs of keys, 2) latencies for distance categories based on a standard number pad layout, and 3) latencies for distance categories based on a standard number row layout. FIG. 9 is a block diagram illustrating a standard number pad layout 900, in accordance with various aspects of the present disclosure. FIG. 10 is block diagram illustrating a standard number row layout 1000, in accordance with various aspects of the present disclosure. FIG. 11 is diagram illustrating different categories 1100-1114 of distances between number positions in the standard number pad layout 900 of FIG. 9, in accordance with various aspects of the present disclosure.









TABLE 6







Latency Time Features










Feature
Feature


#
Name
Description





1
Latencies for
Each of these are



specific pairs
individual features:



of keys
pair_00_x




pair_01_x




pair_02_x




. . .




pair_99_x




(100 total)




where x in [dd, uu, ud, du]




Given 10 possible digits, there are 10 × 10 = 100 possible




combinations that exist in an OTP:




(0, 0)




(0, 1)




(0, 2)




. . .




(9, 9)




For the example OTP 356024, the 5 pairs would be: (3, 5), (5, 6),




(6, 0), (0, 2), (2, 4)




The server may then determine, for example, the up-down time for




each of these pairs and fill those 100, leaving the pairs which are




not applicable in this OTP entry.


2
Latencies for
Each of these are



distance
individual features:



categories
numpad_pair_cat1_x



based on a
numpad_pair_cat2_x



standard
. . .



number pad
numpad_pair_cat_8_x



layout.
(8 total)




where x in [dd, uu, ud, du]




Assuming, for example, that the latency between 7 and 9 is




comparable to the latency between 7 and 1 since they are equally as




far apart on the number pad. Each of the 100 pairs of digits




categorized into 8 different categories of distances (see FIG. 11),




and latencies are calculated only within multiple latencies between




pairs exist within the same category for an OTP. Assuming for,




example, that the latency between 7 and 9 is comparable to the




latency between 4 and 6 or between 7 and 1 since they are equally




as far apart on the number pad. Each of the 100 pairs of digits




above are therefore categorized into 8 different categories of




distances, and latencies are calculated only within these 8




categories. Where multiple latencies between pairs exist within the




same category for an OTP, the latencies are averaged to produce




one latency value for the category.


3
Latencies for
Each of these are individual features:



distance
numrow_pair_cat1_x



categories
numrow_pair_cat2_x



based on a
. . .



standard
numrow_pair_cat_10_x



number row
(10 total)



layout.
where x in [dd, uu, ud, du]




Assuming, for example, that the latency between 1 and 3 is




comparable to the latency be and 7 since they are equally as far




apart on the number pad. Each of the 100 pairs of digits categorized




into 10 different categories of distances, and latencies are




calculated only with where multiple latencies between pairs exist




within the same category for an OTP. Assuming for, example, that




the latency between 1 and 3 is comparable to the latency between 2




and 4 or between 5 and 7 since they are equally as far apart on the




number pad. Each of the 100 pairs of digits above are therefore




categorized into 10 different categories of distances, and latencies




are calculated only within these 10 categories. Where multiple




latencies between pairs exist within the same category for an OTP,




the latencies are averaged to produce one latency value for the




category.
















TABLE 7







Miscellaneous Keystroke Dynamics












Feature
Feature
Feature



#
Name
Description
Type
Considerations





1
total_kd
Total number of
Numeric -
Considering our




key presses (key
discrete
expected OTP codes




down events) in the

are 6 digits in length,




OTP sequence

this should be at






least 6. More kd's






may indicate errors






or error correction,






and fewer kd's may






indicate use of






keyboard shortcuts






(e.g. copy and paste,






which should be






flagged/disqualified).






Error corrections






will make this a






larger number - for






all features, it would






be simpler to only






use samples where






OTP was inputted






without any error.


2
numeric_kd_to_total_kd
Proportion of total
Numeric
If error correction




number of key

cases are excluded




down events where

and there is no shift




a numeric key was

use, this ratio should




pressed

be 1 most of the time






for desktop






keyboards. For






mobile touch






keyboards however,






there could be a lot






more touch events






not for inputting






numeric values (e.g.






scrolling/flicking up






and down).


3
total_edit_kd
Total number of
Numeric
Uses new keyboard




times an editing

location profiling of




key was used in the

kd events in IPR




OTP sequence (i.e.




number of kd's on




backspace, delete,




insert regardless of




keyboard location)


4
edit_kd_to_total_kd
Proportion of total
Numeric
This would be 0




number of key

most of the time if




down events where

excluded in error




an editing key was

correction cases




pressed


5
numpad_ind
Indicates whether a
Binary




full keyboard




containing a




numpad was used




(at least one key




was pressed where




numpad location




was indicated)









The mobile sensor data may be collected, for example, using a JavaScript widget or Native Mobile SDKs from four sensors that capture orientation, rotation, and acceleration data (both with and without the effect of gravity) in three dimensions. In some examples, the sensor events are not aggregated and may be driven at a sixteen millisecond (ms) rate.









TABLE 7







Mobile Sensor Features











Feature
Feature
Feature


#
Name
Description
Type













1
avg_sensor_value_x
Average sensor value for each axis
Numeric



avg_sensor_value_y



avg_sensor_value_z


2
med_sensor_value_x
median sensor value for each axis
Numeric



med_sensor_value_y



med_sensor_value_z


3
mean_med_ratio_sensor_value_x
mean to median sensor value ratio for
Numeric



mean_med_ratio_sensor_value_y
each axis



mean_med_ratio_sensor_value_z


4
std_sensor_value_x
Std deviation of sensor values for each
Numeric



std_sensor_value_y
axis



std_sensor_value_z


5
coefvar_sensor_value_x
coef. of variation of sensor values for
Numeric



coefvar_sensor_value_y
each axis



coefvar_sensor_value_z


6
avg_abs_diff_x
Average absolute difference between
Numeric



avg_abs_diff_y
each of the sensor readings and their



avg_abs_diff_z
mean for each axis


7
iqr_sensor_value_x
Interquartile range sensor value for
Numeric



iqr_sensor_value_y
each axis



iqr_sensor_value_z


8
avg_result_acceleration
the average of the square root of the
Numeric




sum of the square of the x, y, z axis




values


9
binned_distrib_x_i
for i from 1 to n determine the range of
Numeric



binned_distrib_y_i
values for each axis (max − min),



binned_distrib_z_i
divide this range into n equal sized




bins, and then record what fraction of




the sensor values fell within each of the




bins. Note: here n is a parameter.




Usually n = 10


10
n_peaks_norm_x
Number of the peaks for each axis
Numeric -



n_peaks_norm_y
normalized by the total session time
discrete



n_peaks_norm_z
(usually the sensor time series,




similarly to other signal data, looks like




repetitive ways on the graph, e.g.




sinusoid. In this case, the server




computes the number of those waves




for each axis). The server may also




define a threshold value that defines a




peak, e.g. discard small peaks.


11
range_peak_x
The difference between max and min
Numeric



range_peak_y
peak values for each axis. Note: The



range_peak_z
server may also use ratio.


12
avg_peak_x
Average peak value for each axis
Numeric



avg_peak_y



avg_peak_z


13
avg_time_bw_peaks_x
Average time between peaks for each
Numeric



avg_time_bw_peaks_y
axis.



avg_time_bw_peaks_z
Note: this feature assumes that there's




more than one peak. If there's no




distinguishable peaks based on the




threshold in 9, then the server may




lower the threshold or compute the




average time between first n maximum




values for each axis, where n is a




parameter
















TABLE 8







Mobile Sensor Events










ID
Event
Custom Data Parameters
Description





ac
devicemotion.acceleration
1. Represents the acceleration upon the x
Acceleration



IncludingGravity
axis which is the west to east axis
of the device




2. Represents the acceleration upon the y
on the three




axis which is the south to north axis
axis X, Y and




3. Represents the acceleration upon the z
Z with the




axis which is the down to up axis
effect of




NOP is device is stationary
gravity.





Acceleration is





expressed in





m/s2


gy
devicemotion.rotationRate
1. The rate at which the device is rotating
Rate of change




about its Z axis; that is, being twisted
of the device's




about a line perpendicular to the screen.
orientation on




2. The rate at which the device is rotating
the three




about its X axis; that is, front to back.
orientation




3. The rate at which the device is rotating
axis alpha,




about its Y axis; that is, side to side.
beta and




NOP is device is stationary
gamma.





Rotation rate is





expressed in





degrees per





seconds.


lac
devicemotion.acceleration
1. Represents the acceleration upon the x
Acceleration




axis which is the west to east axis
of the device




2. Represents the acceleration upon the y
on the three




axis which is the south to north axis
axis X, Y and




3. Represents the acceleration upon the z
Z.




axis which is the down to up axis
Acceleration is




NOP is device is stationary
expressed in





m/s2


or
deviceorientationevent
1. a number representing the motion of the
Information




device around the z axis, express in
from the




degrees with values ranging from 0 to
physical




360.
orientation of




2. a number representing the motion of the
the device




device around the x axis, express in
running the




degrees with values ranging from −180 to
web page or




180. This represents a front to back
mobile




motion of the device
application




3. a number representing the motion of the




device around the y axis, express in




degrees with values ranging from −90 to




90. This represents a left to right motion




of the device




4. a boolean that indicates whether or not




the device is providing orientation data




absolutely - this value is optional, true if




provided.




NOP is device is stationary









One example sampling frequency used for the data collection described above is 62.5 hertz (Hz) (i.e., sensor events driven every sixteen milliseconds). However, the sensor events are stored in the IPR (e.g., the IPR 400 or the IPR 500) and the resulting size of the IPR may exceed a desired threshold (e.g., 20,000 bytes maximum, more preferably, 5 kB).


After collecting data every 16 ms, the mobile sample resulted in 167,017 observations and a mean of 29,000 bytes. In order the meet the recommended production IPR size of approximately 5 kB, then the IPR must be approximately six times smaller. With the sensor data consuming more than 90% of the IPR size, then the sensor sampling rate must be at least six times slower than the current 16 ms or roughly 100 ms (10 events per second). With the sensor sampling rate set to 100 ms, more than 99% of IPRs will not require truncation and the average IPR would be approximately 5,000 bytes.


Alternatively, in some examples, instead of setting the sensor sampling rate to 100 ms, the number of sensors collecting data may be reduced (e.g., remove gravity accelerator) and the sensor sampling rate may be set at a more accurate 50 ms sampling rate. In these examples, the data collection is most accurate when using higher sampling rates for sensors that do not have much short time variation (e.g., gyroscope and orientation).


Thus, the present disclosure provides, among other things, user identification based on an input profile record. Various features and advantages of the invention are set forth in the following claims.

Claims
  • 1. A server comprising: a memory including an input profile record repository; andan electronic processor in communication with the memory, the electronic processor configured to receive a plurality of input profile records (IPRs) associated with a first user, the plurality of input profile records each based on a plurality of user inputs and indicative of identity of the first user,control the memory to store the plurality of IPRs in the input profile record repository,receive a current IPR associated with a second user,determine whether the second user is the first user by comparing a first one or more biometric features based on the plurality of IPRs and a second one or more biometric features based on the current IPR, andresponsive to determining that the second user is the first user, output an identity confirmation that the second user is the first user.
  • 2. The server of claim 1, wherein, to determine whether the second user is the first user by comparing the first one or more biometric features based on the plurality of IPRs and the second one or more biometric features based on the current IPR, the electronic processor is further configured to generate, with a biometric identification algorithm, the first one or more biometric features from the plurality of IPRs,generate, with the biometric identification algorithm, the second one or more biometric features from the current IPR,generate, with the biometric identification algorithm, a biometric score based on difference between the second one or more biometric features and the first one or more biometric features,determine whether the biometric score is less than a lower threshold,determine whether the biometric score greater than the lower threshold and less than an upper threshold, anddetermine whether the biometric score is greater than the lower threshold and the upper threshold, andwherein the second user is the first user when the biometric score is greater than the lower threshold and the upper threshold.
  • 3. The server of claim 2, wherein, to generate, with a biometric identification algorithm, the first one or more biometric features from the plurality of IPRs and generate, with the biometric identification algorithm, the second one or more biometric features from the current IPR, the electronic processor is further configured to determine a first one or more latencies of a first dwell time based on the plurality of IPRs, anddetermine a second one or more latencies of a second dwell time based on the current IPR.
  • 4. The server of claim 2, wherein the second user is not the first user when the biometric score is lower than the lower threshold and the upper threshold, and wherein the second user is undetermined relative to the first user when the biometric score is higher than the lower threshold and lower than the upper threshold.
  • 5. The server claim 1, wherein the plurality of IPRs and the current IPR each include an IPR header and a plurality of IPR events, wherein the plurality of IPR events includes a key down event and a key up event, wherein the plurality of user inputs is a one-time-password (OTP), and wherein each user input of the plurality of user inputs includes the key down event and the key up event associated with each key in the OTP.
  • 6. The server of claim 1, wherein the plurality of IPRs and the current IPR each include an IPR header and a plurality of IPR events, and wherein the plurality of IPR events includes two or more of: a form state event,a form field focus event,a form field blur event,a key down event,a key up event,a mouse move event,a mouse click event,a touch event,an accelerometer event,a form submit event,a total keys event,a scroll position event,a control list event,a time sync event,a mouse movement sample event,a device motion sample event,a device motion event,a stop event, anda truncate event.
  • 7. The server of claim 1, wherein the first one or more biometric features and the second one or more biometric features each include a plurality of sensor features including two or more of: average sensor value for each axis,median sensor value for the each axis,mean to median sensor value ratio for the each axis,standard deviation of sensor values for the each axis,coefficient of variation of sensor values for the each axis,average absolute difference between sensor readings and the mean for the each axis,interquartile range sensor value for the each axis,an average of a square root of a sum of a square of x, y, z axis values,binned distribution for the each axis,number of peaks for the each axis normalized by total session time,difference between maximum and minimum peak values for the each axis,average peak value for the each axis, andaverage time between the peaks for the each axis.
  • 8. A method for user identification, the method comprising: receiving, with the electronic processor, a plurality of input profile records (IPRs) associated with a first user, the plurality of input profile records each based on a plurality of user inputs and indicative of identity of the first user;controlling, with the electronic processor, a memory to store the plurality of IPRs in an input profile record repository;receiving, with the electronic processor, a current IPR associated with a second user;determining, with the electronic processor, whether the second user is the first user by comparing a first one or more biometric features based on the plurality of IPRs and a second one or more biometric features based on the current IPR; andresponsive to determining that the second user is the first user, outputting, with the electronic processor, an identity confirmation that the second user is the first user.
  • 9. The method of claim 8, wherein determining whether the second user is the first user by comparing the first one or more biometric features based on the plurality of IPRs and the second one or more biometric features based on the current IPR further includes generating, with a biometric identification algorithm, the first one or more biometric features from the plurality of IPRs,generating, with the biometric identification algorithm, the second one or more biometric features from the current IPR,generating, with the biometric identification algorithm, a biometric score based on difference between the second one or more biometric features and the first one or more biometric features,determining whether the biometric score is less than a lower threshold,determining whether the biometric score greater than the lower threshold and less than an upper threshold, anddetermining whether the biometric score is greater than the lower threshold and the upper threshold, andwherein the second user is the first user when the biometric score is greater than the lower threshold and the upper threshold.
  • 10. The method of claim 9, wherein, generating, with the biometric identification algorithm, the first one or more biometric features from the plurality of IPRs and generating, with the biometric identification algorithm, the second one or more biometric features from the current IPR further includes determining a first one or more latencies of a first dwell time based on the plurality of IPRs, anddetermining a second one or more latencies of a second dwell time based on the current IPR.
  • 11. The method of claim 9, wherein the second user is not the first user when the biometric score is lower than the lower threshold and the upper threshold, and wherein the second user is undetermined relative to the first user when the biometric score is higher than the lower threshold and lower than the upper threshold.
  • 12. The method claim 8, wherein the plurality of IPRs and the current IPR each include an IPR header and a plurality of IPR events, wherein the plurality of IPR events includes a key down event and a key up event, wherein the plurality of user inputs is a one-time-password (OTP), and wherein each user input of the plurality of user inputs includes the key down event and the key up event associated with each key in the OTP.
  • 13. The method of claim 8, wherein the plurality of IPRs and the current IPR each include an IPR header and a plurality of IPR events, and wherein the plurality of IPR events includes two or more of: a form state event,a form field focus event,a form field blur event,a key down event,a key up event,a mouse move event,a mouse click event,a touch event,an accelerometer event,a form submit event,a total keys event,a scroll position event,a control list event,a time sync event,a mouse movement sample event,a device motion sample event,a device motion event,a stop event, anda truncate event.
  • 14. The method of 8, wherein the first one or more biometric features and the second one or more biometric features each include a plurality of sensor features including two or more of: average sensor value for each axis,median sensor value for the each axis,mean to median sensor value ratio for the each axis,standard deviation of sensor values for the each axis,coefficient of variation of sensor values for the each axis,average absolute difference between sensor readings and the mean for the each axis,interquartile range sensor value for the each axis,an average of a square root of a sum of a square of x, y, z axis values,binned distribution for the each axis,number of peaks for the each axis normalized by total session time,difference between maximum and minimum peak values for the each axis,average peak value for the each axis, andaverage time between the peaks for the each axis.
  • 15. A system comprising: a user interface device configured to output a plurality of input profile records (IPRs) associated with a first user, the plurality of input profile records each based on a plurality of user inputs and indicative of identity of the first user; anda server including a memory including an input profile record repository; andan electronic processor in communication with the memory, the electronic processor configured to receive the plurality of IPRs,control the memory to store the plurality of IPRs in the input profile record repository,receive a current IPR associated with a second user,determine whether the second user is the first user by comparing a first one or more biometric features based on the plurality of IPRs and a second one or more biometric features based on the current IPR, andresponsive to determining that the second user is the first user, output an identity confirmation that the second user is the first user.
  • 16. The system of claim 15, wherein, to determine whether the second user is the first user by comparing the first one or more biometric features based on the plurality of IPRs and the second one or more biometric features based on the current IPR, the electronic processor is further configured to generate, with a biometric identification algorithm, the first one or more biometric features from the plurality of IPRs,generate, with the biometric identification algorithm, the second one or more biometric features from the current IPR,generate, with the biometric identification algorithm, a biometric score based on difference between the second one or more biometric features and the first one or more biometric features,determine whether the biometric score is less than a lower threshold,determine whether the biometric score greater than the lower threshold and less than an upper threshold, anddetermine whether the biometric score is greater than the lower threshold and the upper threshold, andwherein the second user is the first user when the biometric score is greater than the lower threshold and the upper threshold.
  • 17. The system claim 16, wherein, to generate, with a biometric identification algorithm, the first one or more biometric features from the plurality of IPRs and generate, with the biometric identification algorithm, the second one or more biometric features from the current IPR, the electronic processor is further configured to determine a first one or more latencies of a first dwell time based on the plurality of IPRs, anddetermine a second one or more latencies of a second dwell time based on the current IPR.
  • 18. The system of claim 16, wherein the second user is not the first user when the biometric score is lower than the lower threshold and the upper threshold, and wherein the second user is undetermined relative to the first user when the biometric score is higher than the lower threshold and lower than the upper threshold.
  • 19. The system of claim 15, wherein the plurality of IPRs and the current IPR each include an IPR header and a plurality of IPR events, wherein the plurality of IPR events includes a key down event and a key up event, wherein the plurality of user inputs is a one-time-password (OTP), and wherein each user input of the plurality of user inputs includes the key down event and the key up event associated with each key in the OTP.
  • 20. The system of claim 15, wherein the plurality of IPRs and the current IPR each include an IPR header and a plurality of IPR events, and wherein the plurality of IPR events includes two or more of: a form state event,a form field focus event,a form field blur event,a key down event,a key up event,a mouse move event,a mouse click event,a touch event,an accelerometer event,a form submit event,a total keys event,a scroll position event,a control list event,a time sync event,a mouse movement sample event,a device motion sample event,a device motion event,a stop event, anda truncate event.
CROSS-REFERENCE TO RELATED APPLICATION

This application claims the benefit of U.S. Provisional Application No. 63/085,598, filed on Sep. 30, 2020, the entire contents of which are hereby incorporated by reference.

Provisional Applications (1)
Number Date Country
63085598 Sep 2020 US