The utilization of touchscreens in the construction of modern computers, tablets devices, and mobile phones is rapidly increasing. At the same time, the use of computers, tablet devices, and mobile phones for storing sensitive information or accessing personal online accounts is also growing. Methods for confirming and securing personal devices and private online information of the users are highly needed. Using additional hardware with the touchscreen such as iris-scanners or biometric fingertip readers is not a practical solution. The need for a solution that enables touchscreens to recognize the identity of its user without utilizing a special biometric hardware is apparent. This solution will make our information more protected in a society that counts mainly on touchscreens in accessing information.
The present invention discloses an innovative touchscreen that has the ability to recognize the identity of its user with one touch of two fingers. In one embodiment of the present invention, the touchscreen utilizes force sensors to detect the 3D direction of the two fingers and the force exerted of each finger when they are simultaneously touching the touchscreen. Detection of the 3D direction and the force exerted by each finger, in addition to the distance between the two fingers during the moment of touch determines certain measurements that accurately describe the hand skeleton. Comparison of these measurements with those stored in a database reveals the user's identity.
The two fingers used for biometrics touchscreen can be any two fingers from the same hand, such as the thumb and the index finger, the thumb and the middle finger, or the index finger and the middle finger. More than two fingers of the same hand can also be utilized in detecting the user's identity, such as the thumb, the index finger and the ring finger. The hand can be the left hand or the right hand. The touchscreen can be a touchscreen of a mobile phone, tablet, computer, or the like. The recognition of the user's identity can be done at the initial moment of use of the touchscreen to gain access to the device or the computer. The recognition of the user's identity can be continuous, prolonged period of time and not limited to initial use, which ensures that the one who began the use of the device is still the same person who is interacting with the device. Also for public computers, the recognition and storing of the users identities will be done each time a user identity is changed during use of the same computer.
In one embodiment, the present invention discloses a touchscreen that recognizes the identity of a user touching the touchscreen by two fingers of a hand, wherein the touchscreen is comprised of;
To clarify the function and operation of the present invention,
The thumb and index finger can be any combination of two fingers from the same hand. For example,
The configuration of the two fingers touching the touch panel can take various shapes. For example,
Generally, the touch panel can be the touchscreen of any mobile phone, tablet, or computer equipped with sensors to detect the 3D directions and values of the forces exerted from the two fingers on the touchscreen. For example,
Generally, detecting the first position and the second position of the two points of touch by the two fingers determines the distance between the two positions of touch. Detecting the 3D directions and values of the two forces exerted by the two fingers determines the distribution of the hand force through the two fingers. The combination of the distance of the two positions of touch and the distribution of the hand force through the two fingers represents the uniqueness of the hand skeleton. The uniqueness of the hand skeleton can be described as the unique dimensions of the hand bones and the structure of the hand joints. The structure of the hand joints impacts the force translation from the hand through the two fingers to the touch panel.
It is important to note that each phalange, metacarpal, and carpal of a user's hand has different dimensions, shape, and connectivity with each other. This results in a different distance between the two points of touch touched by of the two fingers associated with different 3D directions of two forces exerted by the two fingers, and a different ratio between these two forces. Using these three parameters (the distance between the two points of touch; the 3D directions of the two forces; and the ratio between the two values of the two forces) enables recognition of the user's identity. Generally, recognition of the user's identity is achieved by using a database that associates each unique combination of the three aforementioned parameters with a unique ID.
To create the database, various methods are utilized. In one embodiment, the database is created by touching the touchscreen at all possible distances between two fingers and storing each distance with the corresponding 3D directions and values of the forces of the two fingers of the user. In another embodiment, the database is created by automatically storing the distances between the two fingers, the 3D directions and values of the forces of the two fingers each time the users touches the touchscreen.
Generally, the present invention is utilized with other computer input devices rather than the touchscreen. For example,
In one embodiment, the present invention detects the user's identity using more than two fingers. For example, the user's identity can be recognized when the user touches the touchscreen with any combination of three, four, or five fingers belonging to the same hand.
In another embodiment, the present invention recognizes the identity of the user when the forces exerted from the fingers on the touchscreen are horizontal forces. For example,
In one embodiment of the present invention, detecting the points of touch, the 3D directions of the forces, and the values of the forces exerted by the fingers on the touchscreen, is achieved by a plurality of force sensors. For example,
Generally, the 3D direction and value of each force exerted on the touch panel can be determined by using methods or tools other than the force sensors. For example, the 3D direction of the force can be captured by a tracking camera. The value of the force exerted by the finger can also be determined by using a special touchscreen that detects the finger pressure or force. Also, the value of the force exerted by the finger can be determined, as known in the art, by measuring the area of finger's touch on the touchscreen, where each different area is associated with a certain finger pressure or force.
The touch panel described in the previous example can be a transparent surface designed to be attached to a touchscreen of a mobile phone, tablet, or a computer. It can also be a built-in component in the touchscreen of the mobile phone, tablet, or computer. It can also be used for touchpads of laptops or the like. In the case of using the present invention with a computer keyboard, the force sensors are positioned inside each button of the keyboard to enable detection of the user's identity when touching two or more buttons by two or more fingers. Also, in case of using the present invention with a computer mouse, the force sensors are positioned inside the left and right buttons of the mouse to detect the user's identity via the two buttons. The force sensors can also be positioned inside the chassis of the computer mouse to detect the user's identity once the chassis is touched by two or more fingers.
Once the points of touch, the 3D directions and values of the forces are detected, the distance between the points of touch and the ratio between the forces are calculated to retrieve the user's identity from the database. In one were to use, more than two fingers, the distances between each two successive points of touch and the angles between the lines that represent these distances are used in determining the user's identity. For example,
In another embodiment, the identity of the user can be recognized if two fingers or more are not touching the touchscreen at the same time. All that is required in this case is to have the hand supported at the same place without movement, while each finger is individually touching the touchscreen. The position of touch of each finger is detected to determine the distance between each two positions of touch. Also, the 3D direction and value of each force exerted from a finger is detected to determine the ratio between the forces. Accordingly, the user's identity can be recognized using the database, as previously described.
Finally, the presented invention can be described as a method for recognizing the identity of a hand skeleton by detecting the distribution pattern of the hand force through at least two fingers of the hand touching a surface, and checking the distribution pattern with a database that associates the distribution pattern with an identifier representing the hand skeleton.
Conclusively, while a number of exemplary embodiments have been presented in the description of the present invention, it should be understood that a vast number of variations exist, and these exemplary embodiments are merely representative examples, and are not intended to limit the scope, applicability or configuration of the disclosure in any way. Various of the above-disclosed and other features and functions, or alternative thereof, may be desirably combined into many other different systems or applications. Various presently unforeseen or unanticipated alternatives, modifications variations, or improvements therein or thereon may be subsequently made by those skilled in the art which are also intended to be encompassed by the claims, below. Therefore, the foregoing description provides those of ordinary skill in the art with a convenient guide for implementation of the disclosure, and contemplates that various changes in the functions and arrangements of the described embodiments may be made without departing from the spirit and scope of the disclosure defined by the claims thereto.
This application is a continuation-in-part of a U.S. patent application Ser. No. 12/587,339, filed Oct. 6, 2009, titled “Touch Sensing Technology”.
Number | Date | Country | |
---|---|---|---|
Parent | 12587339 | Oct 2009 | US |
Child | 14147528 | US |