The invention relates generally to handwriting recognition technologies, and more particularly to three-dimensional (3D) handwriting recognition methods and systems.
Handwriting recognition is a technique by which an intelligence system can recognize characters and other symbols written by hand. This technique has been very popular since it frees users from the keyboard, allowing users to write and draw in a more natural way. With the increasing demands from users, more and more devices now have incorporated handwriting recognition system to give users natural input experience. Handwriting recognition is particularly popular in various handheld devices that provides handwriting recognition of words with complex structures, e.g., Chinese characters, input of which using a conventional keyboard is very time consuming.
A handwriting input device provides a user with a friendly way of inputting information. At present, the minimum requirement for an inputting device is a mouse. To write with the mouse, the user needs to press and hold down the primary mouse button, and then move the mouse pointer to form strokes of a word or character for generating the final word or character. Popular handwriting input devices, such as a pen stylus and tablet, are used on conventional handheld devices such as PDAs, or are connected to a computer through a serial or USB port. Handheld devices often use a pen stylus and a touch screen as a pad to allow users to perform recognition functions. Most handheld devices, such as PDAs, are equipped with this kind of input device.
Another kind of handwriting input device includes a pen that allows users to transfer data into a receiving unit, such as a cellular phone, a PDA or a PC by simply writing or drawing in their own natural handwriting on the comfort and space of a regular piece of paper.
At present, all conventional handwriting input devices adopt a two-dimensional input method. Users have to write on a physical medium, such as a tablet, a touch screen, or a paper pad. This limits the choices available for users. For example, if a user wants to write some comments during a presentation or a lecture, he or she would have to first find a physical medium, e.g., a table, a paper pad, etc. This can cause much inconvenience for the user while standing in the room, giving the presentation or lecture. Also, in a mobile environment, such as in a car, a bus or a subway, it would be very inconvenient to “write” on a physical medium using a stylus.
Therefore, there is a need to provide an improved handwriting recognition system that gives users more flexibility and convenience and frees the users from the physical medium required for two-dimensional handwriting recognition.
The present invention gives users more flexibility and enjoyable writing experience by allowing users to freely write words or characters in a 3D space in a touchless way, without requiring any physical medium such as a pad or a tablet.
According to the invention, there is provided a handwriting recognition system that comprises an input device and a recognition device in communication with the input device. The input device includes a three-dimensional (3D) motion detection sensor configured to generate 3D motion data in response to a 3D motion. In one embodiment, the motion detection sensor measures acceleration of the 3D motion in X, Y and Z axial directions to generate the 3D motion data. The recognition device is configured to receive (e.g., by wireless means) the 3D motion data from the input device and derive corresponding two-dimensional (2D) images for handwriting recognition, based on the 3D motion data.
According to one embodiment of the invention, the recognition device calculates corresponding 3D coordinates based on the 3D motion data, constructs corresponding 3D tracks based the 3D coordinates, and derives the corresponding 2D images from the 3D tracks by mapping the 3D tracks onto a 2D plane for handwriting recognition.
Other objects and attainments together with a fuller understanding of the invention will become apparent and appreciated by referring to the following description and claims taken in conjunction with the accompanying drawings.
The invention is explained in further detail, and by way of example, with reference to the accompanying drawings wherein:
Throughout the drawings, the same reference numerals indicate similar or corresponding features or functions.
In operation, a user moves input device 20 to freely write words or characters in a 3D space, e.g., in the air. Motion detection sensor 22 detects the 3D motion and communicates the 3D motion data and a sampling rate to recognition device 30 for handwriting recognition via a communication interface 28, such as Bluetooth, Zigbee, IEEE 802.11, infrared, or a USB port. The sampling rate may be a predetermined value set by an end user or a manufacturer based on factors such as the processing capability of the system. Alternatively, the sampling rate may be dynamically determined and adjusted based on, for example, the speed of the motion. This can be done, for example, by first determining the speed of the initial motion associated with the handwriting. Then the recognition device can dynamically adjust the sampling rate based on the speed of the motion at the last sampling point. The higher the speed is, the higher the sampling rate will be, and vice versa. By dynamically adjusting the sampling rate, the accuracy of the handwriting recognition can be improved since only the optimal number of sampling points will be used for constructing the word or character. Further, lower power consumption is needed.
Based on the motion data and the sampling rate received from input device 20, processor 32 calculates the corresponding 3D coordinates on the X, Y and Z axes and stores these coordinates in storage device 36. Using the calculated coordinates, processor 32 constructs the corresponding 3D tracks. The 3D tracks will then be projected onto a 2D plane to form 2D images which will be recognized using conventional handwriting recognition software. The final result is displayed on output device 40.
Since 3D writing is a continuous process, control circuit 26 of input device 20 provides a control signal to recognition device 30 via interface 28 to indicate separation of individual words or characters upon receiving a user provided external input. For example, the user may press a control button to cause control circuit 26 to generate the control signal after completion of writing a word or a character.
According to a specific embodiment of the invention, motion detection sensor 22 detects the 3D motion by measuring the acceleration of the movement along the X, Y and Z axes. As an example, the piezoresistive-type tri-axial accelerating sensor commercially available from Hitachi Metals, Ltd., Tokyo, Japan, may be used as motion detection sensor 22. This accelerating sensor in the form of an IC chip has the ability to simultaneously detect acceleration in the three axial directions (X, Y and Z). The sensor is highly sensitive and shock resistant and is a very small and thin semiconductor type 3 axial accelerating sensor.
Calculation of the 3D coordinates is continuously performed based on the incoming 3D motion data until processor 32 detects receipt of a control signal (step 112). The control signal indicates completion of writing a word or a character. The corresponding 3D tracks are constructed using the 3D coordinates (step 116) and are then mapped onto a 2D plane (step 122). Thereafter, conventional 2D handwriting recognition is performed (step 126).
In the above recognition process 100, to map the 3D tracks onto a 2D plane at step 122, it is necessary to first find a proper 2D projection plane. In a preferred embodiment of the invention, a proper 2D projection plane is separately derived for each word or character.
According to the geometry principles, a proper 2D projection plane is a plane, to which the sum of the distance square of each sampling point is minimal. Assume the coordinates of n sampling points are known as follows: (x1, y1, z1), (x2, y2, z2) . . . (xn, yn, zn), and the equation of the plane is Ax+By+Cz+D=0, where (A2+B2+C20). Now it is necessary to determine the values of A, B, C, D. The distance from one point (x1, y1, z1) to the plane is given as
and the sum
represented by F(A,B,C,D) is expressed as:
The values of A, B, C, D can be determined using the Lagrange multiplication method as follows, which is described in Mathematics Analysis by Ouyang Guangzhong, published by Fudan University Press in 1999 in China, which is hereby incorporated by reference. Thus,
F(A,B,C,D)=F′(A,B,C,D)=(Ax1+By1+Cz1+D)2+(Ax2+By2+Cz2+D)2+ . . . +(Axn+Byn+Czn+D)2
under the constraint A2+B2+C2=1. From this equation, the following equation is derived:
G(A,B,C,D)=F′(A,B,C,D)+λ(A2+B2+C2−1),
where λ is a Lagrange multiplier, which is a constant. The partial differential functions are now performed on G(A, B, C, D) with respect to A, B, C and D as follows:
From the above four equations, the following equations are obtained:
where equation (4) can be rewritten as:
By incorporating equation (6) into equations (1), (2), and (3), the following is obtained:
Thus, from the above equations, the values of A, B, C and D can be obtained.
As an example, the following 3D coordinates for a total of 13 sampling points are obtained for a Chinese character:
{(0,0,0),(0.49,−1,0.02),(1,0,0.03),(0,0,0.02),(2,−0.13, 0.01),(3.5,−0.14,0),(3.76,−0.31,−0.01),(2.74,−1.26, 0.01),(1.38,−2.24,0),(2.5,−2,0.01),(1.746,1,0.02),(2,−1,5,0.03),(1.876,−4,5,0.02),}.
By using the Lagrange multiplication method described above, the following are obtained:
A
2
+B
2
+C
2=1, and
From the above equations, the values of A, B, C and D are determined as follows:
A=0.0045, B=0.0023, C=0.9999, D=−0.01777.
Thus, the equation of the 2D projection plane is 0.0045x+0.0023y+0.9999z−0.01777=0. From the equation of the projection plane, Ax+By+Cz+D=0, and the equation of a line that is vertical to the projection plane,
the following equations are derived:
which can be used to obtain the corresponding 2D coordinates for each 3D sampling point. In this example, the following corresponding 2D coordinates are obtained: {(0.0001,0),(0.49,−1),(1,−2.0001),(0,0),(2,−0.13),(3.5001,−0.14),(3.7601,−0.31),(2.7401,−1.26),(1.3801, 2.24),(2.5001,−2),(1.746,1),(2,−1.5001),(1.876,−4.5001)}.
Based on these 2D coordinates, the corresponding 2D image is projected onto the 2D projection plane. The result of this Chinese character is shown in
In the above, the Lagrange multiplication method is used to obtain the values of A, B, C and D. Other methods such as linear regression method may also be used.
According to the invention, the input device and the recognition device can be integrated together as a single unit (e.g., a pen) that operates in the same manner as described above. The final recognition results will be transmitted to an intelligent apparatus such as a PDA, a laptop computer, a PC, etc.
The invention can also be used on a 2D plane if the user so chooses. In such a case, the coordinates calculated will be for 2D plane and the step of mapping the 3D tracks onto a 2D plane is omitted.
While the invention has been described in conjunction with specific embodiments, it is evident that many alternatives, modifications and variations will be apparent to those skilled in the art in light of the foregoing description. Accordingly, it is intended to embrace all such alternatives, modifications and variations as fall within the spirit and scope of the appended claims.
Number | Date | Country | Kind |
---|---|---|---|
02 1 44248 | Sep 2002 | CN | national |
Filing Document | Filing Date | Country | Kind | 371c Date |
---|---|---|---|---|
PCT/IB03/04102 | 9/18/2003 | WO | 00 | 8/29/2005 |
Publishing Document | Publishing Date | Country | Kind |
---|---|---|---|
WO2004/029866 | 4/8/2004 | WO | A |
Number | Name | Date | Kind |
---|---|---|---|
5568565 | Minakata | Oct 1996 | A |
6212296 | Stork et al. | Apr 2001 | B1 |
6226404 | Ikebata | May 2001 | B1 |
6233351 | Feeney et al. | May 2001 | B1 |
6628847 | Kasabach et al. | Sep 2003 | B1 |
6737591 | Lapstun et al. | May 2004 | B1 |
20010004254 | Okahara et al. | Jun 2001 | A1 |
20030215140 | Gounares et al. | Nov 2003 | A1 |
20040003071 | Mathew et al. | Jan 2004 | A1 |
Number | Date | Country |
---|---|---|
WO9521436 | Aug 1995 | WO |
Number | Date | Country | |
---|---|---|---|
20060149737 A1 | Jul 2006 | US |