1. Field of the Invention
The present invention relates to a character input apparatus and, more particularly, to a video-based handwritten character input apparatus.
2. Description of Related Art
Recently, with the progress of science and technology, almost all the electronic products are developed to have light weight, small volume and powerful performance, such as a PDA, a mobile phone and a notebook computer. However, the shrinkage of the volume causes difficulty in engaging with an input device often used in the past and having a bigger size, such as a tablet, a keyboard, a mouse and a joystick, rendering degraded portability. Therefore, it is an important issue on how to conveniently input information in a portable electronic product.
In order to facilitate people to conveniently input information, it is prosperously developed in research for various kinds of interactive human-machine interfaces. The most convenient way of inputting a character is to directly use a gesture motion for operating a computer or to use hand-writing by a fingertip. In order to detect the gesture motion or the position of the fingertip, a glove-based approach of using a glove is submitted, in which a data glove equipped with a sensor is used, capable of accurate perception of a lot of data regarding the gesture motion of a user, including contact and curvature of the finger, the rotation degree of a wrist, and so on. The advantage thereof is capable of obtaining accurate gesture information, but the disadvantages thereof are high cost and limitation of the activity range, while burdening on the user when wearing such equipment on the hand.
Another video-based approach may be classified into two kinds, one being based on establishing a model, and the other being based on shape information of an appearing contour. The approach based on establishing a model uses more than two cameras to take pictures of hand's motions, then to calculate the position in a 3D space of a hand, and further to compare with a 3D model built in advance, thereby obtaining the current hand's motion or the position of the fingertip. However, such an approach requires a lot of calculations, and can hardly be applied in a real time. The existing commonly used approach is based on the shape information of an appearing contour, in which a single camera is used to take pictures of the of hand's motions, the information regarding the edge or shape of a hand is sliced and taken out, and such information is based to verify the hand's gesture or to judge the position of fingertip. Since not many calculations are required, it becomes a popular, commonly-used approach recently.
After obtaining the information of the hand's motions or the track of the handwritten character, it is required to proceed with verification on the hand's motions or the handwritten character. There are three commonly-used ways as follows: a hidden Markov model, a neural network and a dynamic time warp matching algorithm, in which the rate of verification for the way of the dynamic time warp matching algorithm is the highest, while the time taken thereby is rather longer. In view of this, according to the present invention, some basic strokes for constructing a character are defined, including eight direction strokes, eight curvature strokes and two circle strokes, with which a one-dimensional series of all the possible strokes are assembled, and character's comparison is made using the dynamic time warp matching algorithm, capable of tolerating input, deletion and substitution of a stroke, thereby enhancing performance of comparison to obtain the effect of the real-time verification.
An object of the present invention is to provide a video-based character input apparatus, comprising: an image capturing unit, an image processing unit, a one-dimensional feature coding unit, a character recognizing unit, a display unit, a stroke feature database and a character database. Among which, the image capturing unit is used for capturing an image; the image processing unit is used for filtering a moving track of an object in the image, the object may be a fingertip, and the procedures include: firstly detecting a graphic difference, then detecting a skin color, and finally picking out a moving track most corresponding to a point of the object; the stroke feature database stores various strokes and their corresponding codes; the one-dimensional feature coding unit is used for taking a stroke with respect to the moving track and converting the stroke into a coding sequence in a one-dimensional string according to a time sequence and the kinds of strokes include strokes in eight directions, semi-circles, and circles; the character database is used for storing characters, including Chinese, English, digits and symbols; the character recognizing unit is used for proceeding with character comparison between the coding sequence in a one-dimensional string and the character database to find out a character having the most similarity; and the display unit is used for displaying the character found out by the character recognizing unit.
The image capturing unit includes a network camera, an image capturing device in a mobile device, and an image capturing device in an embedded device. The character recognizing unit proceeds with character comparison using a dynamic time warp matching algorithm. Thus, the objective and effect of effectively recognizing video-based handwritten characters and inputting characters can be achieved using the video-based character input apparatus of the invention.
A further object of the invention is to provide a method for inputting a character in a video-based character input apparatus, in which the video-based character input apparatus includes an image capturing unit, an image processing unit, a one-dimensional feature coding unit, a character recognizing unit, a display unit, a stroke feature database for storing various strokes and their corresponding codes and a character database for storing Chinese, English, digits and symbols. According to the method, the image capturing unit is to capture an image; the image processing unit is to filter a moving track of an object in the picture, the object may be a fingertip, and the procedures include: firstly detecting a graphic difference, then detecting a skin color, and finally picking out a moving track most corresponding to a point of the object; the one-dimensional feature coding unit is to take a stroke with respect to the moving track, to search the stroke feature database and to convert the stroke into a coding sequence in a one-dimensional string according to a time sequence, in which the kinds of strokes include strokes in eight directions, semi-circles, and circles; the character recognizing unit is to proceed with character comparison between the coding sequence in a one-dimensional string and the character database to find out a character having the most similarity; and the display unit is to display the character found out by the character recognizing unit.
The image capturing unit includes a network camera, an image capturing device in a mobile device, and an image capturing device in an embedded device. The character recognizing unit proceeds with character comparison using a dynamic time warp matching algorithm. Thus, the objective and effect of effectively recognizing video-based handwritten characters and inputting characters can be achieved using the method for inputting a character in the video-based character input apparatus of the invention.
Other objects, advantages, and novel features of the invention will become more apparent from the following detailed description when taken in conjunction with the accompanying drawings.
FIGS. 2(A)-(B) are schematic diagrams showing coding of kinds of strokes according to a preferred embodiment of the invention;
To facilitate understanding technical contents of the invention, a video-based character input apparatus is particularly submitted and explained as follows. Please refer to
The one-dimensional feature coding unit 12 is used for taking a stroke with respect to the moving track. Please refer to FIGS. 2(A)-(B), which are schematic diagrams showing coding of kinds of strokes according to a preferred embodiment of the invention. The coding of kinds of strokes is based to construct a fundamental stroke for a character model, including strokes in eight directions (0-7 in FIG. 2(A)), eight strokes in semi-circle ((A)-(H) in FIG. 2(B)), and two strokes in circle ((O) and (Q) in FIG. 2(B)), which are all stored in the stroke feature database 15. Based on a one-dimensional on-line model, the one-dimensional feature coding unit 12 converts the stroke into a coding sequence in a one-dimensional string according to a time sequence. The character recognizing unit 13 is used for proceeding with character comparison between the coding sequence in a one-dimensional string and characters stored in the character database 16, including such as Chinese, English, digits and symbols, to find out a character having the most similarity and outputs the same to the display unit 14 for displaying, using a dynamic time warp matching algorithm.
Please refer to
Please refer to
Please refer to
Please refer to
Please refer to
Although the present invention has been explained in relation to its preferred embodiment, it is to be understood that many other possible modifications and variations can be made without departing from the scope of the invention as hereinafter claimed.
Number | Date | Country | Kind |
---|---|---|---|
097140620 | Oct 2008 | TW | national |