The present invention relates to a handwritten music sign recognition device and a program, and particularly relates to technology preferably used to recognize a sign inputted by handwriting by a user using an input means, such as a finger or a pen.
A tablet terminal corresponding to a handwriting input detects trace information resulted from scanning a touch panel with an input means, such as a finger or a pen, and recognizes a sign, a figure or the like based on the detected information.
Patent Literature 1 discloses a method as follows. A specific sign or figure for converting a character is registered in advance. When a character is inputted in a character box in which a handwritten character is inputted for performing character recognition, this character is converted into a target character based on the information registered in advance.
Further, Patent Literature 2 discloses a method to extract a line component from inputted coordinate data.
Moreover, Patent Literature 3 discloses a method as follows. The end point and the starting point of a writing stroke are linked with virtual lines, and a characteristic amount of a bent position generated in this case is detected. Then, the detected characteristic amount and dictionary data are compared to perform character recognition.
Patent Literature 1: Japanese Laid-open Patent Publication No. 2000-194796
Patent Literature 2: Japanese Laid-open Patent Publication No. 1-96770
Patent Literature 3: Japanese Laid-open Patent Publication No. 2004-133951
However, in the techniques described in Patent Literatures 1, 3, it is necessary to register in advance a specific sign or figure for converting a character. Accordingly, it is necessary to retain data by amounts of characters, signs, and figures which are recognizable. Therefore, a storage medium having a large storage capacity is needed.
Further, it is necessary that inputted characters and character candidates which are stored by large amounts in the storage medium are correlated one-to-one. Therefore, a complicated and large-scale program is needed.
Further, it is not possible to recognize a character or sign which is not registered in the storage medium. Moreover, precise sign recognition is not performed unless it is inputted clearly.
Further, in the technique described in Patent Literature 2, when a plurality of inputted coordinate data are in a predetermined range, the plurality of coordinate data are taken as line data linking coordinates of a starting point and coordinates of an end point. When the inputted coordinate data are determined as being out of the predetermined range, coordinates of an end point of coordinate data preceding the coordinate data and coordinates of a first starting point of coordinate data within the predetermined range are extracted and stored as a line component. While recognition of such a line component is performed sequentially, line data are generated. Accordingly, the amount of calculations needed for recognizing the entirety of an inputted sign becomes large.
The present invention has been made in view of the above-described problems, and it is an object thereof to enable easy and accurate recognition of a sign inputted by handwriting by a user with an input means, such as a hand or a pen.
A handwritten music sign recognition device according to the present invention has: a touch information obtaining means which obtains, as position information, a position of an input means on a screen every time the input means moves, from when the input means touches the screen until when the input means is moved away therefrom; a vector information creation means which creates attributes of a vector indicating a trace of movement of the input means as vector information based on the position information obtained by the touch information obtaining means; a stroke information storage means which stores one or a plurality of pieces of stroke information for each object to be recognized, the stroke information including the vector information with respect to a series of one or a plurality of the vectors created by the vector information creation means based on the position information from when the input means touches the screen until when the input means is moved away therefrom; and a stroke characteristic amount derivation means which derives a characteristic amount of a stroke being a trace of movement of the input means from when the input means touches the screen until when the input means is moved away therefrom, based on the vector information included in the stroke information.
A program according to the present invention causes a computer to execute: a touch information obtaining step which obtains, as position information, a position of an input step on a screen every time the input step moves, from when the input step touches the screen until when the input step is moved away therefrom; a vector information creation step which creates attributes of a vector indicating a trace of movement of the input step as vector information based on the position information obtained by the touch information obtaining step; a stroke information storage step which stores one or a plurality of pieces of stroke information for each object to be recognized, the stroke information including the vector information with respect to a series of one or a plurality of the vectors created by the vector information creation step based on the position information from when the input step touches the screen until when the input step is moved away therefrom; and a stroke characteristic amount derivation step which derives a characteristic amount of a stroke being a trace of movement of the input step from when the input step touches the screen until when the input step is moved away therefrom, based on the vector information included in the stroke information.
According to the present invention, a sign inputted by handwriting by a user with an input means, such as a hand or a pen, can be recognized easily and accurately.
Hereinafter, an embodiment of a handwritten music sign recognition device of the present invention will be explained with reference to the drawings.
To begin with, a first embodiment will be explained.
The handwritten music sign recognition device 10 is constituted of a tablet terminal for example. Note that the handwritten music sign recognition device 10 is not limited to the tablet terminal as long as it is a terminal device having a touch panel. The handwritten music sign recognition device 10 may be constituted of, for example, a PC (Personal Computer) which can use a touch panel as a user interface.
In
A program which controls the overall operation of the handwritten music sign recognition device 10 is stored in the flash memory 15 or the ROM 12. The CPU 14 reads this program and deploys the program in the RAM 13 to execute it. Thus, operation of the handwritten music sign recognition device 10 is realized.
The touch display 17 is a computer display having a touch panel. The user performs a handwriting input by using an input means, such as a finger or a pen, on the touch display 17. The handwritten music sign recognition device 10 recognizes the handwritten-inputted object (sign).
The network interface 16 is for storing information of recognized objects (signs) as a file in another piece of hardware, or reading a file of information of stored signs.
The audio output device 18 is for creating music performance information from recognized objects (signs) to enable audio output.
The MIDI output device 19 creates the MIDI information from recognized objects (signs) to enable MIDI output.
As illustrated in
The touch information obtaining part 21 is realized by using, for example, the touch display 17, the RAM 13, the CPU 14, and the flash memory 15, and the like. Further, the touch information obtaining part 21 has an API (Application Programming Interface) which obtains position information every time a movement of the input means, such as a finger or a pen, is made.
The vector information creation part 22, the stroke characteristic amount derivation part 23, the object generation part 24, the stroke recognition part 25, and the object recognition part 26 are realized by using, for example, the RAM 13, the CPU 14, and the flash memory 15, and the like.
<Explanation of Vector, Stroke, and Object>
The relation of vectors, a stroke, and an object will be explained by taking an example of a segno mark with reference to
An object 30 of a segno mark is constituted of an S-shaped stroke 31 (stroke 0), a line stroke 32 (stroke 1), and two point strokes 33, 34 (strokes 2, 3). The line stroke 32, the point stroke 33, and the point stroke 34 are each constituted of one vector.
(Method of Assigning a Vector ID)
In this embodiment, for simplification of processing, the vector information creation part 22 categorizes a vector under any one of following nine types (“down→up”, “up→down”, “lower left→upper right”, “upper right→lower left”, “left→right”, “right→left”, “upper left→lower right”, “lower right→upper left”, “point”).
As illustrated in
Then, the vector information creation part 22 determines a vector ID taking into consideration the positional relation of coordinates x1, y1, x2, y2. Note that in
The vector information creation part 22 categorizes vectors according to (A) to (K) below.
(A) When a vector is |a|>2 and the relation between Y-coordinate y2 of end point and Y-coordinate y1 of starting point of the vector is y2<y1, the vector information creation part 22 defines vector ID of this vector as “down→up”.
(B) When the inclination a of a vector is |a|>2 and the relation between Y-coordinate y2 of end point and Y-coordinate y1 of starting point of the vector is y2>y1, the vector information creation part 22 defines vector ID of this vector as “up→down”.
(C) When the inclination a of a vector is −2≤a<−0.5 and the relation between X-coordinate x2 of end point and X-coordinate x1 of starting point of the vector is x2<x1, the vector information creation part 22 defines vector ID of this vector as “upper right→lower left”.
(D) When the inclination a of a vector is −2≤a<−0.5 and the relation between X-coordinate x2 of end point and X-coordinate x1 of starting point of the vector is x2>x1, the vector information creation part 22 defines vector ID of this vector as “lower left→upper right”.
(E) When the inclination a of a vector is |a|≤0.5 and the relation between X-coordinate x2 of end point and X-coordinate x1 of starting point of the vector is x2>x1, the vector information creation part 22 defines vector ID of this vector as “left→right”.
(F) When the inclination a of a vector is |a|≤0.5 and the relation between X-coordinate x2 of end point and X-coordinate x1 of starting point of the vector is x2<x1, the vector information creation part 22 defines vector ID of this vector as “right left”.
(G) When the inclination a of a vector is 0.5<a≤2 and the relation between X-coordinate x2 of end point and X-coordinate x1 of starting point of the vector is x2>x1, the vector information creation part 22 defines vector ID of this vector as “upper left→lower right”.
(H) When the inclination a of a vector is 0.5<a≤2 and the relation between X-coordinate x2 of end point and X-coordinate x1 of starting point of the vector is x2<x1, the vector information creation part 22 defines vector ID of this vector as “lower right→upper left”.
(I) When the relation between X-coordinate x2 of end point and X-coordinate x1 of starting point of a vector is x1=x2 and the relation between Y-coordinate y2 of end point and Y-coordinate y1 of starting point of the vector is y2<y1, the vector information creation part 22 defines vector ID of this vector as “down→up”.
(J) When the relation between X-coordinate x2 of end point and X-coordinate x1 of starting point of a vector is x1=x2 and the relation between Y-coordinate y2 of end point and Y-coordinate y1 of starting point of the vector is y2>y1, the vector information creation part 22 defines vector ID of this vector as “up→down”.
(K) When the relation between X-coordinate x2 of end point and X-coordinate x1 of starting point of a vector is x1=x2 and the relation between Y-coordinate y2 of end point and Y-coordinate y1 of starting point of the vector is y1=y2, the vector information creation part 22 defines vector ID of this vector as “point”.
In the example illustrated in
First, from the starting point coordinate and the end point coordinate in
Therefore, the vector whose vector Index (index) of the S-shaped stroke 31 (stroke 0) is 0 corresponds to the above (A). Thus, the vector ID of this vector is “down→up”.
Note that in
<Explanation of Each Characteristic Amount of a Stroke>
As described above,
(Starting Point-End Point Distance)
When the starting point-end point distance is smaller than a prescribed value in a stroke, the stroke is considered as a closed figure like a circle. On the other hand, when the starting point-end point distance is larger than the prescribed value in the stroke, the stroke is considered as a figure which is not closed like an arc line.
The coordinates of starting point of the headmost vector of a stroke are (x1, y1), and the coordinates of end point of the final vector of this stroke are (x2, y2). The stroke characteristic amount derivation part 23 obtains a starting point-end point distance Dse with expression (2) below. The starting point-end point distance Dse is a straight distance linking the starting point of the headmost vector of a stroke and the end point of the final vector of this stroke.
[Expression 2]
Dse=√{square root over ((x2−x1)2+(y2−y1)2)} (2)
(Coordinates of Four Points of an Inscribed Rectangle and Center Coordinates of a Stroke)
Coordinates of four points and center coordinates of an inscribed rectangle of a stroke are used for obtaining the positional relation of strokes with each other in a plurality of strokes. The inscribed rectangle of a stroke is a rectangle inscribed on the stroke. The coordinates of four points of the inscribed rectangle of a stroke are coordinates of four corners of the inscribed rectangle of the stroke.
A rectangle coordinate left is the leftmost X-coordinate among starting points or end points of all the vectors constituting the S-shaped stroke 31 (see the field left of the inscribed rectangle coordinates of
A rectangle coordinate right is the rightmost X-coordinate among starting points or end points of all the vectors constituting the S-shaped stroke 31 (see the field right of the inscribed rectangle coordinates of
A rectangle coordinate up is the highest Y-coordinate among starting points or end points of all the vectors constituting the S-shaped stroke 31 (see the field up of the inscribed rectangle coordinates of
A rectangle coordinate down is the lowest Y-coordinate among starting points or end points of all the vectors constituting the S-shaped stroke 31 (see the field down of the inscribed rectangle coordinates of
Center coordinates are expressed with an intermediate X-coordinate between the rectangle coordinate left and the rectangle coordinate right and an intermediate Y-coordinate between the rectangle coordinate up and the rectangle coordinate down (see the fields of center X and center Y of the inscribed rectangle coordinates of
(Connection Type)
A connection type is used for determining whether a sign is smooth, whether a sign has a corner, or the like.
The stroke characteristic amount derivation part 23 determines the connection type of a vector whose connection type is to be determined based on the vector ID of the vector whose connection type is to be determined (target vector ID), the vector ID of a next vector of the target vector (next vector ID), and the relation 700 illustrated in
In this embodiment, as the connection type, one of obtuse angle (=1), right angle (=2), acute angle (=3), and none of these (=0) is determined as the connection type of each vector.
In the S-shaped stroke 31 (stroke 0) of the segno mark of
(Rotation Direction)
A point where the rotation direction of a stroke changes is a hint for identifying a sign.
For example, in the first stroke of “f”, vectors are continuous smoothly like “S”, and the starting point and the end point are separated. “S” has one point where its rotation direction changes. On the other hand, “f” does not have any point where its rotation direction changes. Thus, the stroke recognition part 25 can distinguish “S” and “f” by presence of a point where the rotation direction of a stroke changes.
The stroke characteristic amount derivation part 23 determines the rotation direction of a vector whose rotation direction is to be determined based on the vector ID of the vector whose rotation direction is to be determined (target vector ID), the vector ID of a next vector of the target vector (next vector ID), and the relation 800 illustrated in
In the S-shaped stroke 31 (stroke 0) of the segno mark of
(Rotation Angle)
The coordinates of starting point of the vector whose rotation angle is to be determined are (x11, y11), and the coordinates of end point are (x12, y12). Further, the coordinates of starting point of a next vector of the vector whose rotation angle is to be determined are (x21, y21), and the coordinates of end point are (x22, y22). In this case, the stroke characteristic amount derivation part 23 obtains the angle formed by the two vectors by expression (3) to expression (7) below as a rotation angle Deg (°) of the vector whose rotation angle is to be determined with respect to the next vector of the target vector.
(Left-Right Movement Direction and Up-Down Movement Direction)
The stroke recognition part 25 determines whether a stroke is a sign like a line drawn simply from up to down or left to right by a left-right movement direction and an up-down movement direction.
Then, the stroke characteristic amount derivation part 23 obtains the left-right movement direction and the up-down movement direction from the vector ID.
When the vector ID is down→up (=1), there is no left-right movement direction (=0), and the up-down movement direction is up (=1). Further, when the vector ID is lower right→upper left (=8), the left-right movement direction is left (=1), and the up-down movement direction is up (=1). Note that for the convenience of description, in
(Indexes of Leftmost Point, Rightmost Point, Highest Point, Lowest Point)
For example, the leftmost point Index is 0 (starting point of the first vector constituting a stroke) and the rightmost point Index is N (starting point of the final vector constituting a stroke). In this case, the handwritten music sign recognition device 10 can determine it as a chevron sign with an open bottom if the highest point Index is near N/2.
In the example illustrated in
<Operation Flowchart>
Next, an example of operation of the handwritten music sign recognition device 10 will be explained with reference to a flowchart of
In step S101, the touch information obtaining part 21 determines whether a screen of the touch display 17 is touched or not. When the screen is not touched as a result of this determination, it becomes a standby state. When the screen is touched, the process proceeds to step S102.
When it proceeded to step S102, the touch information obtaining part 21 empties a coordinate array, and then stores the touched coordinates in the coordinate array. After the screen is touched by the input means, until the input means moves away completely from the screen, movement processing of the input means is performed every time the input means is moved. This movement processing of the input means is performed by storing the coordinates of the moved input means in the coordinate array. Note that as described above the input means is, for example, a finger or a pen.
Next, in step S103, the touch information obtaining part 21 determines whether it is moved away from the touched screen or not. As a result of this determination, when the input means is not moved away from the screen, the process returns to step S102. Then, until the input means is moved away completely from the screen, the movement processing of the input means of step S102 is performed every time the input means moves. When the input means is moved away completely from the screen as a result of the determination in step S103, the process proceeds to step S104.
When it proceeded to step S104, the vector information creation part 22 and the stroke characteristic amount derivation part 23 execute stroke generation processing.
First, the vector information creation part 22 newly generates stroke information.
Next, the vector information creation part 22 creates coordinates of a vector with first coordinates being a starting point and second coordinates being an end point. At this time, the vector information creation part 22 sets the vector ID of the created vector. Further, the vector information creation part 22 derives the magnitude of the created vector. The vector information creation part 22 stores vector information indicating attributes of the created vector (coordinates, vector ID, magnitude) in a vector array of the stroke information.
Next, the vector information creation part 22 creates vector information of a vector with the second coordinates being a starting point and third coordinates being an end point, and stores it in the vector array of the stroke information.
In the S-shaped stroke 31 (stroke 0) of
Next, in step S105, the vector information creation part 22 determines whether the screen of the touch display 17 is touched or not within a given length of time. As a result of this determination, when an input of a next stroke is started within the given length of time, the process returns to step S102. Then, the processing of step S102 to S104 is performed, so as to create the stroke information (vector information, characteristic amount) of the new stroke.
In the example illustrated in
When it proceeded to step S106, the object generation part 24 performs object generation processing to newly generate an object. The object generation part 24 stores stroke information (vector information, characteristic amounts) of the stroke which has been generated in the stroke array of the object.
At this time, information as illustrated in
Moreover, the object generation part 24 obtains relative positional relation (position ID) of strokes. The position ID is an ID indicating the position of each stroke with respect to a stroke as a reference.
In this embodiment, the object generation part 24 uses nine types of position IDs below.
(L) When the positions of the stroke as a reference and the stroke for which a position ID is to be obtained are the same, the object generation part 24 defines the position ID of this stroke as “same (=0)”.
(M) When the stroke as a reference is down and the stroke for which a position ID is to be obtained is up (down→up), the object generation part 24 defines the position ID of this stroke as “up (=1)”.
(N) When the stroke as a reference is lower left and the stroke for which a position ID is to be obtained is upper right (lower left→upper right), the object generation part 24 defines the position ID of this stroke as “upper right (=2)”.
(O) When the stroke as a reference is left and the stroke for which a position ID is to be obtained is right (left→right), the object generation part 24 defines the position ID of this stroke as “right (=3)”.
(P) When the stroke as a reference is upper left and the stroke for which a position ID is to be obtained is lower right (upper left→lower right), the object generation part 24 defines the position ID of this stroke as “lower right (=4)”.
(Q) When the stroke as a reference is up and the stroke for which a position ID is to be obtained is down (up→down), the object generation part 24 defines the position ID of this stroke as “down (=5)”.
(R) When the stroke as a reference is upper right and the stroke for which a position ID is to be obtained is lower left (upper right→lower left), the object generation part 24 defines the position ID of this stroke as “lower left (=6)”.
(S) When the stroke as a reference is right and the stroke for which a position ID is to be obtained is left (right→left), the object generation part 24 defines the position ID of this stroke as “left (=7)”.
(T) When the stroke as a reference is lower right and the stroke for which a position ID is to be obtained is upper left (lower right→upper left), the object generation part 24 defines the position ID of this stroke as “upper left (=8)”.
Here, an example of a method to obtain the position ID will be explained.
The object generation part 24 obtains a vector whose starting point is the center coordinates of an inscribed rectangle of the stroke as a reference and whose end point is the center coordinates of an inscribed rectangle of the stroke for which a position ID is to be obtained. Then, the object generation part 24 obtains the vector ID of this vector. For example, when the vector ID is “down→up”, the position ID is “up (=1)”, and when the vector ID is “left→right”, the position ID is “right (=3)”. Note that the vector ID is obtained in accordance with the above-described references (A) to (K).
In the example illustrated in
Next, the object generation part 24 obtains the position ID of the line stroke 32 (stroke 1). In this case, the object recognition part 26 employs the S-shaped stroke 31 (stroke 0) as the stroke as a reference. The object generation part 24 obtains a vector whose starting point is the center coordinates of an inscribed rectangle of the S-shaped stroke 31 as a reference and whose end point is the center coordinates of an inscribed rectangle of the line stroke 32 for which the position ID is to be obtained. Then, the object generation part 24 obtains the vector ID of this vector.
In this case, X-coordinates of the starting point and the end point are the same (see
Thereafter, when the position IDs of the point strokes 33, 34 (strokes 2, 3) are obtained similarly, four position IDs “left (=7)”, “up (=1)”, “left (=7)”, “right (=3)” are obtained. Note that the position IDs of the point strokes 33, 34 (strokes 2, 3) are “left (=7)” and “right (=3)”, respectively.
The object generation part 24 stores these position IDs in the position ID array of a segno mark object as illustrated in
As described above, the object generation part 24 stores stroke information of a target stroke in the stroke array of each stroke of the object. Moreover, the object generation part 24 stores the position ID of the target stroke in the position ID array of each stroke of the object.
Returning to explanation of
An example of the stroke/object recognition processing performed in step S107 of
In step S131, the stroke recognition part 25 performs stroke recognition processing to recognize a stroke. Details of the stroke recognition processing will be described later.
Next, in step S132, the stroke recognition part 25 determines whether there is any other unrecognized stroke or not. When there is an unrecognized stroke as a result of this determination, the process returns to step S131. Then, the stroke recognition part 25 performs the stroke recognition processing on the unrecognized stroke. On the other hand, when there is no unrecognized stroke, the process proceeds to step S133.
Next, in step S133, the object recognition part 26 performs object recognition processing to recognize an object. Details of the object recognition processing will be described later.
As illustrated in
An example of the stroke recognition processing performed in step S131 of
In step S141, the stroke recognition part 25 performs branching of processing according to the number of vectors contained in a stroke to be recognized. When the stroke to be recognized is a stroke constituted of one vector, the process proceeds to step S142. When the stroke to be recognized is a stroke constituted of two vectors, the process proceeds to step S143. When the stroke to be recognized is any other stroke, the process proceeds to step S144.
When it proceeded to step S142, the stroke recognition part 25 performs recognition processing of the stroke constituted of one vector. By the vector ID contained in stroke information of the stroke constituted of one vector for example, the stroke recognition part 25 recognizes what figure this stroke is from among a vertical line, a horizontal line, a rightward obliquely ascending line, a leftward obliquely ascending line, and a point. For example, when the vector ID contained in the stroke information of this stroke is “lower right upper left” or “upper left lower right”, this stroke is the leftward obliquely ascending line.
When it proceeded to step S143, the stroke recognition part 25 performs recognition processing of the stroke constituted of two vectors. In this embodiment, recognizing four signs illustrated in
A relation 1600 illustrated in
In the example illustrated in
Returning to the explanation of
Here, the processing of recognizing signs illustrated in
In step S181, the stroke recognition part 25 refers to a rotation direction among the characteristic amounts of a stroke to be recognized, counts the number of changes of the rotation direction, and performs branching of processing according to the number thereof. When the number of changes of the rotation direction of the stroke to be recognized is 0, the process proceeds to step S182. When the number of changes of the rotation direction of the stroke to be recognized is 1, the process proceeds to step S183. When the number of changes of the rotation direction of the stroke to be recognized is 2, the process proceeds to step S184. In this embodiment, when the number of changes of the rotation direction of the stroke to be recognized is other than 0, 1, and 2, the stroke recognition part 25 determines that there is no sign corresponding to this stroke.
<When the Number of Changes of Rotation Direction is 0>
In step S182, the stroke recognition part 25 performs a determination of other than “S” and “G clef” in the following procedure.
(1) The stroke recognition part 25 determines whether there is an acute angle point or not in the stroke to be recognized based on the connection type of this stroke (see
(2) The stroke recognition part 25 obtains the total of rotation angles of the stroke to be recognized, and recognizes a sign based on the result thereof (see
(2-1) When the Total of Rotation Angles is 180° or Less
When the total of rotation angles of the stroke to be recognized is 180° or less, the stroke recognition part 25 recognizes the sign of this stroke following conditions described in (2-1-1), (2-1-2), (2-1-3), (2-1-4), (2-1-5) below.
(2-1-1)
When there is “right (=2)” in the left-right movement direction of the stroke to be recognized and there is no “left (=1)”, the stroke recognition part 25 recognizes the sign of the stroke to be recognized as in (2-1-1-1) and (2-1-1-2) below (see
(2-1-1-1)
Based on the highest point Index of the stroke to be recognized, when the highest point of the stroke is other than the starting point and the end point of this stroke, the stroke recognition part 25 recognizes that the sign of this stroke as an upper arc line (see
(2-1-1-2)
Based on the highest point Index of the stroke to be recognized, when the highest point of the stroke is the starting point or the end point of this stroke, the stroke recognition part 25 recognizes that the sign of this stroke as a lower arc line (see
(2-1-2)
When there is no “right (=2)” in the left-right movement direction of the stroke to be recognized and there is “left (=1)”, the stroke recognition part 25 recognizes the sign of the stroke to be recognized as in (2-1-2-1) and (2-1-2-2) below (see
(2-1-2-1)
Based on the highest point Index of the stroke to be recognized, when the highest point of the stroke is other than the starting point and the end point of this stroke, the stroke recognition part 25 recognizes that the sign of this stroke as an upper arc line (see
(2-1-2-2)
Based on the highest point Index of the stroke to be recognized, when the highest point of the stroke is the starting point or the end point of this stroke, the stroke recognition part 25 recognizes that the sign of this stroke as a lower arc line (see
(2-1-3)
When there is “down (=2)” in the up-down movement direction of the stroke to be recognized and there is no “up (=1)”, the stroke recognition part 25 recognizes the sign of the stroke to be recognized as in (2-1-3-1) and (2-1-3-2) below (see
(2-1-3-1)
Based on the leftmost point Index of the stroke to be recognized, when the leftmost point of the stroke is other than the starting point and the end point of this stroke, the stroke recognition part 25 recognizes that the sign of this stroke as a left arc line (see
(2-1-3-2)
Based on the leftmost point Index of the stroke to be recognized, when the leftmost point of the stroke is the starting point or the end point of this stroke, the stroke recognition part 25 recognizes that the sign of this stroke as a right arc line (see
(2-1-4)
When there is no “down (=2)” in the up-down movement direction of the stroke to be recognized and there is “up (=1)”, the stroke recognition part 25 recognizes the sign of the stroke to be recognized as in (2-1-4-1) and (2-1-4-2) below (see
(2-1-4-1)
Based on the leftmost point Index of the stroke to be recognized, when the leftmost point of the stroke is other than the starting point and the end point of this stroke, the stroke recognition part 25 recognizes that the sign of this stroke as a left arc line (see
(2-1-4-2)
Based on the leftmost point Index of the stroke to be recognized, when the leftmost point of the stroke is the starting point or the end point of this stroke, the stroke recognition part 25 recognizes that the sign of this stroke as a right arc line (see
(2-1-5)
When none of the (2-1-1), (2-1-2), (2-1-3), and (2-1-4) applies, the stroke recognition part 25 determines that there is no sign corresponding to the stroke to be recognized.
(2-2) When the Total of Rotation Angles Exceeds 360°
When the total of rotation angles of the stroke to be recognized exceeds 360°, the stroke recognition part 25 recognizes the sign of the stroke to be recognized as a multiple circle.
(2-3) When the Total of Rotation Angles is Neither 180° or Less Nor More than 360°
When the total of rotation angles of the stroke to be recognized is out of the ranges described in the (2-1) and the (2-2), the stroke recognition part 25 recognizes the sign of this stroke as a circle or an ellipse. Specifically, the stroke recognition part 25 recognizes whether the stroke to be recognized is the sign of a circle or an ellipse as follows.
First, the stroke recognition part 25 obtains an equation of a line 191 linking the leftmost point and the rightmost point of the stroke to be recognized. Here, the length of the line 191 will be referred to as a major axis.
Next, the stroke recognition part 25 obtains a leg as a normal drawn from each point (starting point and end point of the vector) of the stroke to be recognized down to the line 191.
Next, among the lengths of the points of the stroke to be recognized and the legs of the normals thereof, the stroke recognition part 25 takes the longest one as a half minor axis of an ellipse. In
Next, the stroke recognition part 25 obtains the ratio between the half minor axis×2 (=minor axis) and the major axis.
Next, when the obtained ratio is approximately 1:1, the stroke recognition part 25 recognizes that the sign of the stroke to be recognized is a circle. On the other hand, when the obtained ratio is not approximately 1:1, the stroke recognition part 25 recognizes that the sign of the stroke to be recognized is an ellipse.
<When the Number of Changes of Rotation Direction is 1>
In step S183, the stroke recognition part 25 determines whether the sign of the stroke to be recognized is “S” or not.
First, the stroke recognition part 25 derives the number of vectors before or after the point 201 where the rotation direction of the stroke to be recognized changes. When the derived number of vectors is less than a number (three) by which an arc can be drawn, the stroke recognition part 25 determines that there is no sign corresponding to the stroke to be recognized.
In the example illustrated in
Therefore, the stroke recognition part 25 determines that the number of vectors before or after the point 201 where the rotation direction of “S” changes is a number by which an arc can be drawn.
Note that in the example illustrated in
Further, the stroke recognition part 25 can determine whether it is a vector before the point 201 where the rotation direction of “S” changes or a vector after that based on, for example, the vector Index.
Next, when the absolute value of the total of rotation angles of vectors up to the point where the rotation direction of the stroke to be recognized changes and the absolute value of the total of rotation angles of vectors after the point are both 100° or more, the stroke recognition part 25 recognizes the sign of the stroke to be recognized as “S”. On the other hand, when the absolute values are not 100° or more, the stroke recognition part 25 determines that the sign of the stroke to be recognized is not “S”.
In the example illustrated in
<When the Number of Changes of Rotation Direction is 2>
In step S184, the stroke recognition part 25 determines whether the sign of the stroke to be recognized is a “G clef” or not.
When all of three conditions (U), (V), (W) below are satisfied, the stroke recognition part 25 determines that the sign of the stroke to be recognized is the “G clef”, or otherwise determines that the sign of the stroke to be recognized is not the “G clef”.
(U) There exists one each of first point where the rotation direction of the stroke to be recognized changes from clockwise to counterclockwise and second point where it changes from counterclockwise to clockwise.
(V) First point where the rotation direction of the stroke to be recognized changes from clockwise to counterclockwise is located higher than the center of an inscribed rectangle of the stroke.
(W) Second point where the rotation direction of the stroke to be recognized changes from counterclockwise to clockwise is located lower than the center of an inscribed rectangle of the stroke.
In the example illustrated in
Next, the object recognition processing performed in S133 of the flowchart of
In the following, the object to be recognized which is a segno mark will be explained as an example. In order for a sign drawn on the screen of the touch display 17 to be determined as a segno mark, it is necessary that condition 1, condition 2, and condition 3 below are satisfied.
Condition 1: the object is constituted of four strokes, “S”, “rightward obliquely ascending line”, “point”, and “point”.
Condition 2: the distance between the center of an inscribed rectangle of “S” and the center of an inscribed rectangle of “rightward obliquely ascending line” is less than or equal to a threshold (these centers are located closely).
Condition 3: there are “points” at left and right of the center of an inscribed rectangle of “S” or the center of an inscribed rectangle of “rightward obliquely ascending line”.
In the example illustrated in
Further, the Y-coordinate of the center of an inscribed rectangle of the S-shaped stroke 31 (stroke 0) and the Y-coordinate of the center of an inscribed rectangle of the line stroke 32 (stroke 1) is separated only by 20 (see
The position ID of the point stroke 33 (stroke 2) with respect to the line stroke 32 (stroke 1) is “left (=7)” (see
Thus, in the example illustrated in
In the stroke/object recognition processing, when there is an input of another stroke within a given length of time since the input of one stroke is finished, the stroke recognition part 25 and the object recognition part 26 wait until this input is completed. On the other hand, when there is no input of another stroke within the given length of time since the input of one stroke is finished, the stroke recognition part 25 performs processing of recognizing an unrecognized stroke and registering it in an object, and the object recognition part 26 performs processing of recognizing this object.
Further, instead of doing in this manner, as illustrated in
The handwritten music sign recognition device 10 of this embodiment displays a staff 222 in advance on the screen of the touch display 17. When the object is recognized as a figure (for example a circle or an ellipse) corresponding to a note, the object recognition part 26 recognizes this object as the note. Then, the object recognition part 26 recognizes the pitch of this note based on the position of the recognized note on the staff 222.
In the example illustrated in
Next, an example of a method of recognizing a sharp sign “#” as an example of an object constituted of a plurality of strokes will be explained.
As illustrated in
A sign constituted of a plurality of strokes as illustrated in
The database 230 (dictionary data) for the sharp sign stores “oblique upper right”, “horizontal”, “oblique upper right”, “horizontal” as Stroke 1.
The database 230 (dictionary data) for the sharp sign stores “horizontal”, “oblique upper right”, “oblique upper right”, “horizontal” as Stroke 2.
The database 230 (dictionary data) for the sharp sign stores “vertical”, “vertical”, “vertical”, “vertical” as Stroke 3 and Stroke 4.
In
“Upper left”, “up”, “up”, “upper left” are registered as directions of a vector whose starting point is the barycenter 242 of the stroke Stroke 2 and whose end point is the barycenter 243 of the stroke Stroke 3 (see the field of “2→3” of
“Right”, “right”, “right”, “right” are registered as directions of a vector whose starting point is the barycenter 243 of the stroke Stroke 3 and whose end point is the barycenter 244 of the stroke Stroke 4 (see the field of “3→4” of
The combination of four aligned horizontally in one row in
When the four strokes Stroke 1, Stroke 2, Stroke 3, Stroke 4 satisfy the relations of
A musical note other than the sharp sign and constituted of a plurality of strokes can be recognized by registering in advance the sign of each stroke as illustrated in
As described above, in this embodiment, a music sign is recognized by evaluating characteristic points of vectors constituting the music sign. Therefore, it is possible to accurately recognize a handwritten-inputted music sign without retaining vast amounts of data.
In this embodiment, an example of recognizing a music sign is exemplified, but the handwritten music sign recognition device of the present invention is able to recognize characters, such as “Coda”, and numerals.
Next, a second embodiment will be explained. In the first embodiment, the case that vectors obtained every time the input means is moved are used as they are in the touch information obtaining part 21 is explained as an example. On the other hand, in this embodiment, a combination processing of combining a plurality of vectors into one vector according to predetermined conditions is performed. Thus, the number of vectors can be compressed. In this manner, this embodiment and the first embodiment are different mainly in configuration and processing for performing the combination processing. Specifically, part of the processing in the vector information creation part 22 is mainly different. Therefore, in the explanation of this embodiment, same parts as those in the first embodiment are given the same reference numerals as those given in
In this embodiment, four examples will be explained as the combination processing.
In this example, vector conversion processing of combining a plurality of minute vectors to convert them into a vector having a certain magnitude will be explained.
In this example, there are first to third vector arrays as vector arrays of stroke information.
In the first vector array, vector information of minute vectors (coordinates, vector IDs, magnitudes) is stored. The first vector array is a vector array explained in a first embodiment.
In the second vector array, vector information of minute vectors to be converted is stored.
In the third vector array, vector information of vectors after conversion is stored.
With reference to a flowchart of
In step S261, the vector information creation part 22 selects one non-selected minute vector sequentially from the head minute vector of the first vector array. Then, the vector information creation part 22 copies (adds) vector information of the selected minute vector to the second vector array. At this time, the vector information creation part 22 copies (adds) the vector information of the selected minute vector to the head of empty parts of the second vector array.
Next, in step S262, the vector information creation part 22 generates a test vector whose starting point is the starting point of the first minute vector of the second vector array, and whose end point is the end point of the last minute vector of the second vector array. Then, the vector information creation part 22 obtains the magnitude of the created test vector.
Next, in step S263, the vector information creation part 22 determines whether the magnitude of the test vector exceeds a prescribed value or not. When the magnitude of the test vector exceeds the prescribed value as a result of this determination, the process proceeds to step S264. On the other hand, when the magnitude of the test vector is equal to or less than the prescribed value, the process returns to step S261. Then, in step S261, the vector information creation part 22 copies (adds) the vector information of the next minute vector from the first vector array to the second vector array.
Then, in step S262, the vector information creation part 22 generates a test vector whose starting point is the starting point of the first minute vector of the second vector array, and whose end point is the end point of the minute vector added this time, and obtains the magnitude thereof.
The vector information creation part 22 repeats the processing as described above until the magnitude of the test vector exceeds the prescribed value. Then, as described above, when the magnitude of the test vector becomes equal to or more than the prescribed value, the process proceeds to step S264.
When it proceeded to step S264, the vector information creation part 22 registers vector information (coordinates, vector ID, magnitude) of the test vector whose magnitude is determined to exceed the prescribed value in step S263 in the third vector array, and empties the second vector array.
Next, in step S265, the vector information creation part 22 determines whether there is an unprocessed minute vector or not in the first vector array. When there is no unprocessed minute vector as a result of this determination, the processing by the flowchart of
The stroke characteristic amount derivation part 23, in step S104 of
By executing the flowchart of
In this example, processing of correcting wobbling of a straight vector will be explained.
When a person draw a straight line with the input means, such as a finger or a pen, on the screen of the touch display 17, the straight line may wobble. Unless this wobbling is absorbed to make it one straight line, the stroke recognition part 25 may erroneously determine a stroke which should be determined correctly as a straight line as, for example, an arc line.
In this example, as illustrated in
With reference to a flowchart of
In step S281, the vector information creation part 22 obtains an equation of a line 272 linking the starting point of a first vector 271a and the end point of a last vector 271e.
Next, in step S282, the vector information creation part 22 selects one unselected vector from among vectors 271a to 271e from the first vector 271a to the last vector 271e. Then, the vector information creation part 22 draws a normal down to a straight line 272 obtained in step S281 from the end point of the selected vector, and obtains a distance between a leg of the normal (intersection of the normal and the straight line 272) and the end point of the selected vector.
Next, in step S283, the vector information creation part 22 determines whether the distance obtained in step S282 is equal to or less than a prescribed value or not. When the distance obtained in step S282 is not equal to or less than the prescribed value as a result of this determination, the processing by the flowchart of
On the other hand, when the distance obtained in step S282 is equal to or less than the prescribed value, the process proceeds to step S284.
When it proceeded to step S284, the vector information creation part 22 determines whether there is an unprocessed vector or not. When there is an unprocessed vector as a result of this determination, the process returns to step S282. Then, the vector information creation part 22 performs the processing of the next vector. On the other hand, when there is no unprocessed vector, the process proceeds to step S285.
When it proceeded to step S285, the vector information creation part 22 performs replacement of vectors. Specifically, the vector information creation part 22 generates vector information of a vector whose starting point is the starting point of the first vector of a vector array of stroke information, and whose end point is the end point of the last vector thereof. Then, the vector information creation part 22 deletes all vector information of original vectors stored in the vector array of the stroke information, and stores the vector information of the generated vector in the vector array. Then, the processing by the flowchart of
The stroke characteristic amount derivation part 23, in step S104 of
By executing the flowchart of
In this example, processing of grouping vectors by direction and, when vectors of a same group continue, combining these vectors into one will be explained.
In this example, the vector information creation part 22 groups vectors using vector IDs.
In the example explained with reference to
In
The vector IDs of first three vectors b1, b2, b3 are “left→right”. In this case, as illustrated in the lower diagram of
Then, the vector information creation part 22 replaces vector information of the original three vectors b1, b2, b3 with the vector information of the vector c1. Moreover, the vector IDs of fourth to sixth vectors b4, b5, b6 are “up→down”. Also in this case, the vector information creation part 22 generates a vector c2 whose starting point is the starting point of the fourth vector b4, and whose end point is the end point of the sixth vector b6, and replaces these three vectors with the vector c2.
With reference to a flowchart of
In step S301, the vector information creation part 22 copies the n-th vector from the head in the vector array as a reference vector. The initial value of n is 1.
Next, in step S302, the vector information creation part 22 determines whether the vector ID of the reference vector and the vector ID of the (n+m)-th vector from the head in the vector array are the same or not. The initial value of m is 1.
When the vector ID of the reference vector and the vector ID of the (n+m)-th vector from the head in the vector array are the same as a result of this determination, the process proceeds to step S303. On the other hand, when the vector ID of the reference vector and the vector ID of the (n+m)-th vector from the head in the vector array are different, the process proceeds to step S304.
When it proceeded to step S303, the vector information creation part 22 changes the end point of the reference vector to the end point of the (n+m)-th vector. Then, the vector information creation part 22 recalculates the vector ID of the reference vector based on the position of the changed end point. Moreover, the vector information creation part 22 increments m (adds 1 to m). Then, the process proceeds to step S305, which will be described later.
On the other hand, when it proceeded to step S304, the vector information creation part 22 stores the vector information of the current reference vector in the vector array. At this time, the vector information creation part 22 replaces the vector information of the current reference vector with the vector information of the vector which the reference vector is based on. Moreover, the vector information creation part 22 lets n=n+m. Then, the process proceeds to step S306, which will be described later.
When it proceeded to step S305, the vector information creation part 22 determines whether the processing of all the vectors in the vector array is finished or not. When the processing of all the vectors in the vector array is finished as a result of this determination, the processing by the flowchart of
On the other hand, when the processing of all the vectors in the vector array is not finished, the process returns to step S302 to perform the above-described processing.
Further, when it proceeded to step S306, the vector information creation part 22 determines whether the processing of all the vectors in the vector array is finished or not. When the processing of all the vectors in the vector array is finished as a result of this determination, the processing by the flowchart of
On the other hand, when the processing of all the vectors in the vector array is not finished, the process returns to step S301 to perform the above-described processing.
The stroke characteristic amount derivation part 23, in step S104 of
By executing the flowchart of
In this example, the processing of combining a plurality of vectors in the vicinity of a threshold of inclination into one vector will be explained.
In this example, when a plurality of vectors are categorized by direction, categorizing of vectors is performed by inclination of vector. A person may draw a straight line close to the threshold of inclination with the input means, such as a finger or a pen, on the screen of the touch display 17. In this case, due to shaking of a hand, or the like, the vector ID of a vector may become a vector ID to be given when the inclination exceeds the threshold, and the vector ID of another vector may become a vector ID to be given when the inclination is lower than or equal to the threshold. Then, the vector ID of a vector desired to be actually the same ID may become a different ID, which may result in that vectors cannot be combined by the method of the above-described third example.
Then, in this example, a plurality of vectors in the vicinity of the threshold of inclination are combined into one vector.
Here, when the inclination a of a vector is less than −2.0, the vector ID of this vector is “down→up” (see
The combination processing for determining a vector ID in this manner will be explained with reference to
As illustrated in
As illustrated in
With reference to a flowchart of
In step S321, the vector information creation part 22 generates a list of Index array of vectors to be combined from the vector array. Details of this processing will be described later with reference to the flowchart of
Next, in step S322, the vector information creation part 22 determines whether there is Index or not in the list created in step S321. When Index in the list is empty as a result of this determination, the processing of the flowchart of
On the other hand, when there is Index in the list, the process proceeds to step S323.
When it proceeded to step S323, the vector information creation part 22 generates a combined vector. The combined vector is a vector whose starting point is the starting point of the first vector in the Index array, and whose end point is the end point of the last vector of the Index array. Then, the vector information creation part 22 generates vector information (coordinates, vector ID, magnitude) of the combined vector.
Next, in step S324, the vector information creation part 22 deletes vector information of the vector of Index of the Index array from the vector array.
Next in step S325, the vector information creation part 22 inserts vector information of the combined vector in the position of first Index of the vector array. Thereafter, the process returns to step S321 to generate an Index array of vectors to be combined.
The stroke characteristic amount derivation part 23 derives the characteristic amounts of the stroke based on vector information registered in the vector array in step S104 of
Next, an example of the processing of step S321 of
In step S331, the vector information creation part 22 checks the vector ID of vector ID of vector n and the vector ID (direction) of vector n+1, and determines whether the vectors n, n+1 are next to each other or not. The initial value of n is 1. Further, n denotes the order from the head of vectors in the vector array.
When the vectors n, n+1 are adjacent to each other as a result of this determination, the vector information creation part 22 stores the vector ID of the vector n or vector n+1 as a temporary storage vector ID according to the combination of vectors n, n+1.
Moreover, the vector information creation part 22 sets an X-axis flag indicating whether a threshold of the vector IDs of vectors n, n+1 adjacent to each other is an inclination close to the X-axis or not.
Next, in step S332, the vector information creation part 22 branches the process according to the present temporary storage vector ID and the previous temporary storage vector ID.
Specifically, when the present temporary storage vector ID is not no direction, the process proceeds to step S333 (case 1).
Further, when the present temporary storage vector ID is no direction and also the previous temporary storage vector ID is no direction, the process returns to step S331 (case 2). At this time, the vector information creation part 22 increments n (adds 1 to n).
Further, when the present temporary storage vector ID is no direction and the previous temporary storage vector ID is other than no direction, the processing by the flowchart of
When it proceeded to step S333, the vector information creation part 22 determines whether the previous temporary storage vector ID and the present temporary storage vector ID are the same or not. As a result of this determination, when the previous temporary storage vector ID and the present temporary storage vector ID are the same, the processing by the flowchart of
On the other hand, when the previous temporary storage vector ID and the present temporary storage vector ID are different, the process proceeds to step S334.
When it proceeded to step S334, the vector information creation part 22 determines a threshold angle of vector according to the X-axis flag and the temporary storage vector ID.
The vector information creation part 22 determines the threshold angle (threshold upper limit and threshold lower limit) of vector based on the relation illustrated in
Next, in step S335, when the angle of vector n and the angle of vector n+1 are both within the range of threshold angles determined in step S334 (within the range from the threshold lower limit to the threshold upper limit), the vector information creation part 22 registers Index of vector n+1 in the Index array of vectors to be combined.
At this time, if n=1, the vector information creation part 22 registers Index of vector n in the Index array of vectors to be combined, and then registers Index of vector n+1. Then, the process proceeds to step S336.
On the other hand, when the angle of vector n or the angle of vector n+1 are not within the range of threshold angles determined in step S334 (within the range from the threshold lower limit to the threshold upper limit), the vector information creation part 22 changes the temporary storage vector ID to no direction. Then, the process proceeds to step S336.
When it proceeded to step S336, the vector information creation part 22 checks the present and previous temporary storage vector IDs. When the previous temporary stored vector ID is not no direction and is different from the present temporary storage vector ID, the processing by the flowchart of
When it proceeded to step S337, the vector information creation part 22 substitutes the present temporary storage vector ID for the previous temporary storage vector ID. Further, the vector information creation part 22 increments n (adds 1 to n). Then, the process returns to step S331.
By executing the flowcharts of
As described above, in this embodiment, it is possible to reduce the amount of calculations necessary when a sign for which many vectors are generated by handwriting is recognized.
Note that at least two of the above-described first to fourth examples may be combined.
Next, a third embodiment will be explained. In the first embodiment, the case that the stroke recognition part 25 uses the connection type, which is one of characteristic amounts of a stroke, for determining whether or not there is an acute angle point in the stroke to be recognized is explained as an example.
As explained in the first embodiment, as one characteristic of movement of the input means, such as a human finger or pen, there is a connection type of a vector and a vector inputted next to this vector. In the example explained in the first embodiment, a vector is categorized into a connection type of any one of obtuse angle (=1), right angle (=2), acute angle (=3), and none of these (=0).
In this embodiment, the stroke recognition part 25 divides one stroke into a plurality of strokes at the position of an acute angle of the stroke. Then, the stroke recognition part 25 recognizes each of the divided strokes. In this manner, the stroke recognition part 25 can recognize the shapes of divided individual strokes more simply, and by the positional relation of the individual divided strokes, the sign of the stroke before being divided can be recognized simply and accurately.
Thus, this embodiment is made by adding the above-described processing as processing for the stroke recognition part 25 to recognize a stroke to the first embodiment. Therefore, in the explanation of this embodiment, same parts as those in the first embodiment are given the same reference numerals as those given in
In the character “2” illustrated in the left diagram of
Then, the stroke recognition part 25 recognizes the first to fifth vectors a1 to a5 and the sixth vector a6 as different strokes. Then, as illustrated in the right diagram of
Next, an example of the method of dividing one stroke into a plurality of strokes and recognizing them will be explained with reference to a flowchart of
In step S371, the stroke recognition part 25 obtains the connection type of two adjacent vectors in the order of input, and stores this connection type in a connection type array. When a character “2” is inputted as illustrated in
Next, in step S372, the stroke recognition part 25 divides the original vector array into a vector array of vectors located before an acute angle point and a vector array of vectors located after the acute angle point based on the connection type array. The vector array is created by the number of adding 1 to the number of acute angle points (=number of acute angle points+1). Then, the stroke recognition part 25 deletes the original vector array.
When a character “2” is inputted as illustrated in
In this case, the stroke recognition part 25 stores vector information (coordinates, vector ID, magnitude) of the vectors a1 to a5 in the first vector array of two vector arrays. Further, the stroke recognition part 25 stores vector information (coordinates, vector ID, magnitude) of the vector a6 in the second vector array of the two vector arrays.
Next, in step S373, the stroke recognition part 25 recognizes, in every vector array created in step S372, the shape of a stroke constituted of vectors included in this vector array.
In the example illustrated in
<Regarding First Vector Array>
The stroke recognition part 25 determines whether the change in rotation direction is neither of clockwise and counterclockwise. The stroke recognition part 25 can determine the change in rotation direction by referring to, for example, a rotation direction derived by the stroke characteristic amount derivation part 23. In the example illustrated in
Further, the stroke recognition part 25 derives the total of rotation angles. In the example illustrated in
From the above point, the stroke recognition part 25 can determine that the stroke constituted of the vectors included in the first vector array is an arc line.
Moreover, the stroke recognition part 25 derives following (a) to (h) with respect to the vectors included in the first vector array. Then, the stroke recognition part 25 determines based on the derived result what arc line the stroke constituted of the vectors included in the first vector array is.
(a) number of vectors directed rightward
(b) number of vectors directed leftward
(c) number of vectors directed downward
(d) number of vectors directed upward
(e) index of highest point
(f) index of lowest point
(g) index of leftmost point
(h) index of rightmost point
The stroke recognition part 25 can derive the above (a) and (b) by referring to, for example, a left-right movement direction derived by the stroke characteristic amount derivation part 23.
The stroke recognition part 25 can derive the above (c) and (d) by referring to, for example, an up-down movement direction derived by the stroke characteristic amount derivation part 23.
The stroke recognition part 25 can derive each of the above (e), (f), (g), (h) by referring to, for example, a highest point Index, a lowest point Index, a leftmost point Index, a rightmost point Index derived by the stroke characteristic amount derivation part 23.
In the example illustrated in
From the above conditions, the stroke recognition part 25 recognizes the stroke constituted of the vectors included in the first vector array is an arc line opening leftward.
<Regarding Second Vector Array>
The number of vectors included in the second vector array is one. The vector ID of the vector included in the second vector array is “left→right”. The vector ID is created by the vector information creation part 22. As described above, the vector ID is derived based on the inclination of a vector, and the relation of starting point and end point of the vector.
Therefore, the stroke recognition part 25 recognizes that the stroke constituted of vectors included in the second vector array is a horizontal straight line.
Next, in step S374, the stroke recognition part 25 generates one stroke and recognizes it based on the shape of each stroke and the positional relation of this stroke. In the example illustrated in
As described above, in this embodiment, a stroke is divided at a point where the angles of vectors of a handwritten-inputted sign changes at an acute angle, and each of the divided strokes is recognized individually. Therefore, even when a handwritten-inputted sign is complicated, this sign can be recognized accurately without retaining vast amounts of data.
In this embodiment, an example of recognizing a character “2” is illustrated, but it is also possible to divide one stroke at a point of changing acutely, and to recognize various characters and signs which is each recognizable.
Further, this embodiment can be realized by combining with the second embodiment.
Embodiments of the present invention as have been described can be realized by a computer executing a program. Further, means for supplying a program to a computer, for example a computer readable recording medium, such as a CD-ROM recording such a program, or a transmission medium transmitting such a program, can be applied as an embodiment of the present invention. Further, a program product, such as a computer readable recording medium recording the program can be applied as an embodiment of the present invention. The program, the computer readable recording medium, the transmission medium and the program product are included in the scope of the present invention.
Further, all the above-described embodiments of the present invention merely illustrate specific examples for carrying out the present invention, and the technical scope of the invention should not be construed as limited by these embodiments. That is, the invention may be implemented in various forms without departing from the technical spirit or main features thereof.
The present invention can be used as, for example, a device which automatically recognizes a sign inputted by handwriting by using an input means, such as a finger or a pen.
Number | Date | Country | Kind |
---|---|---|---|
2014-054010 | Mar 2014 | JP | national |
2014-070018 | Mar 2014 | JP | national |
2014-070029 | Mar 2014 | JP | national |
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/JP2015/050875 | 1/15/2015 | WO | 00 |
Publishing Document | Publishing Date | Country | Kind |
---|---|---|---|
WO2015/141260 | 9/24/2015 | WO | A |
Number | Name | Date | Kind |
---|---|---|---|
3541511 | Watanabe | Nov 1970 | A |
3766520 | Patterson | Oct 1973 | A |
3863218 | Oka | Jan 1975 | A |
3979722 | Sakoe | Sep 1976 | A |
3987412 | Morrin, II | Oct 1976 | A |
4040009 | Kadota | Aug 1977 | A |
4124871 | Morrin, II | Nov 1978 | A |
4331955 | Hansen | May 1982 | A |
4468808 | Mori | Aug 1984 | A |
4542526 | Satoh | Sep 1985 | A |
4573196 | Crane | Feb 1986 | A |
4628532 | Stone | Dec 1986 | A |
4653107 | Shojima | Mar 1987 | A |
4685142 | Ooi | Aug 1987 | A |
4718103 | Shojima | Jan 1988 | A |
4891750 | Pastor | Jan 1990 | A |
4979226 | Sato | Dec 1990 | A |
4987603 | Ohnishi | Jan 1991 | A |
5007098 | Kumagai | Apr 1991 | A |
5038382 | Lipscomb | Aug 1991 | A |
5070466 | Anderson | Dec 1991 | A |
5105470 | Will | Apr 1992 | A |
5113452 | Chatani | May 1992 | A |
5265174 | Nakatsuka | Nov 1993 | A |
5313528 | Nishida | May 1994 | A |
5414228 | Yamashita | May 1995 | A |
5459809 | Kim | Oct 1995 | A |
5468077 | Motokado | Nov 1995 | A |
5485531 | Ichinohe | Jan 1996 | A |
5533147 | Arai | Jul 1996 | A |
5579408 | Sakaguchi | Nov 1996 | A |
5588074 | Sugiyama | Dec 1996 | A |
5594855 | Von Ehr, II | Jan 1997 | A |
5596656 | Goldberg | Jan 1997 | A |
5615285 | Beernink | Mar 1997 | A |
5644652 | Bellegarda | Jul 1997 | A |
5665927 | Taki | Sep 1997 | A |
5727140 | Ohtomo | Mar 1998 | A |
5729629 | Dai | Mar 1998 | A |
5734748 | Amano | Mar 1998 | A |
5734750 | Arai | Mar 1998 | A |
5742705 | Parthasarathy | Apr 1998 | A |
5754689 | Niki | May 1998 | A |
5775915 | Quinn | Jul 1998 | A |
5852680 | Yoshida | Dec 1998 | A |
5862251 | Al-Karmi | Jan 1999 | A |
5867172 | Fujisawa | Feb 1999 | A |
5907630 | Naoi | May 1999 | A |
5917501 | Muller | Jun 1999 | A |
5926189 | Beaman | Jul 1999 | A |
5940081 | Guo | Aug 1999 | A |
5940084 | Motokado | Aug 1999 | A |
5940535 | Huang | Aug 1999 | A |
5959634 | Yoshida | Sep 1999 | A |
5982387 | Hellmann | Nov 1999 | A |
5982933 | Yoshii | Nov 1999 | A |
6005976 | Naoi | Dec 1999 | A |
6035063 | Nakashima | Mar 2000 | A |
6041137 | Van Kleeck | Mar 2000 | A |
6051770 | Milburn | Apr 2000 | A |
6052482 | Arai | Apr 2000 | A |
6055332 | Aitani | Apr 2000 | A |
6072114 | Kunimasa | Jun 2000 | A |
6144764 | Yamakawa | Nov 2000 | A |
6201174 | Eller | Mar 2001 | B1 |
6201553 | Morooka | Mar 2001 | B1 |
6226403 | Parthasarathy | May 2001 | B1 |
6232987 | Choi | May 2001 | B1 |
6275237 | Wu | Aug 2001 | B1 |
6275611 | Parthasarathy | Aug 2001 | B1 |
6285785 | Bellegarda | Sep 2001 | B1 |
6289124 | Okamoto | Sep 2001 | B1 |
6292179 | Lee | Sep 2001 | B1 |
6348648 | Connick, Jr. | Feb 2002 | B1 |
6370269 | Al-Karmi | Apr 2002 | B1 |
6373473 | Sakaguchi | Apr 2002 | B1 |
6421461 | Arai | Jul 2002 | B1 |
6464585 | Miyamoto | Oct 2002 | B1 |
6493464 | Hawkins | Dec 2002 | B1 |
6498608 | Dresevic | Dec 2002 | B1 |
6502114 | Forcier | Dec 2002 | B1 |
6538187 | Beigi | Mar 2003 | B2 |
6594023 | Omura | Jul 2003 | B1 |
6597360 | Stamm | Jul 2003 | B1 |
6694056 | Ito | Feb 2004 | B1 |
6721452 | Confer | Apr 2004 | B2 |
6760028 | Salesin | Jul 2004 | B1 |
6937239 | Browne | Aug 2005 | B1 |
6952210 | Renner | Oct 2005 | B1 |
6956968 | O'Dell | Oct 2005 | B1 |
6968083 | Williams | Nov 2005 | B2 |
6970599 | Longe | Nov 2005 | B2 |
6992671 | Corona | Jan 2006 | B1 |
7146046 | Biswas | Dec 2006 | B2 |
7149356 | Clark | Dec 2006 | B2 |
7238876 | Worrall | Jul 2007 | B1 |
7251365 | Fux | Jul 2007 | B2 |
7289123 | Duggan | Oct 2007 | B2 |
7302099 | Zhang | Nov 2007 | B2 |
7330605 | Wang | Feb 2008 | B2 |
7369702 | Abdulkader | May 2008 | B2 |
7432432 | Li | Oct 2008 | B2 |
7495164 | Funaki | Feb 2009 | B2 |
7500187 | Karimoto | Mar 2009 | B1 |
7567238 | Sugimoto | Jul 2009 | B2 |
7729539 | Lin | Jun 2010 | B2 |
7750891 | Stephanick | Jul 2010 | B2 |
7762893 | Kando | Jul 2010 | B2 |
7771279 | Miyamoto | Aug 2010 | B2 |
7885464 | Kawamura | Feb 2011 | B2 |
7927215 | Sekimori | Apr 2011 | B2 |
8024775 | Xu | Sep 2011 | B2 |
8053657 | Sitrick | Nov 2011 | B2 |
8102397 | Perry | Jan 2012 | B2 |
8123601 | Kondo | Feb 2012 | B2 |
8150160 | Al-Muhtaseb | Apr 2012 | B2 |
8445766 | Raveendran | May 2013 | B2 |
8558792 | Momose | Oct 2013 | B2 |
8610744 | Harris | Dec 2013 | B2 |
8744136 | Katsuta | Jun 2014 | B2 |
8781227 | Sun | Jul 2014 | B2 |
8830167 | Ninjouji | Sep 2014 | B2 |
8866820 | Terazono | Oct 2014 | B2 |
8882590 | Shimura | Nov 2014 | B2 |
8884916 | Worfolk | Nov 2014 | B2 |
9070224 | Esfahbod MirHosseinzadeh Sarabi | Jun 2015 | B1 |
9183754 | Tanaka | Nov 2015 | B2 |
9195887 | Shibata | Nov 2015 | B2 |
9274607 | DeLean | Mar 2016 | B2 |
9275274 | Cheng | Mar 2016 | B2 |
9298365 | Takasugi | Mar 2016 | B2 |
9323726 | Joshi | Apr 2016 | B1 |
9339725 | Ohta | May 2016 | B2 |
9424823 | Valente | Aug 2016 | B2 |
9478201 | Begelfor | Oct 2016 | B1 |
9601029 | Gebauer | Mar 2017 | B2 |
9613444 | Makino | Apr 2017 | B2 |
9711117 | Valente | Jul 2017 | B2 |
9851880 | Kim | Dec 2017 | B2 |
9911354 | Oliver | Mar 2018 | B2 |
9922629 | Whiteside | Mar 2018 | B1 |
9934768 | Uemura | Apr 2018 | B2 |
20020049795 | Freeman | Apr 2002 | A1 |
20020064308 | Altman | May 2002 | A1 |
20020071607 | Kawamura | Jun 2002 | A1 |
20020085003 | Nagashima | Jul 2002 | A1 |
20020088334 | Mohammad Beigi | Jul 2002 | A1 |
20020114394 | Ma | Aug 2002 | A1 |
20020180797 | Bachmann | Dec 2002 | A1 |
20030086611 | Loudon | May 2003 | A1 |
20030156145 | Hullender | Aug 2003 | A1 |
20030185444 | Honda | Oct 2003 | A1 |
20030215138 | Raghupathy | Nov 2003 | A1 |
20030215139 | Shilman | Nov 2003 | A1 |
20030215140 | Gounares | Nov 2003 | A1 |
20030215145 | Shilman | Nov 2003 | A1 |
20040006749 | Fux | Jan 2004 | A1 |
20040041811 | Lin | Mar 2004 | A1 |
20040117724 | Szilagyi | Jun 2004 | A1 |
20040140977 | Hakamada | Jul 2004 | A1 |
20050088420 | Dodge | Apr 2005 | A1 |
20050108004 | Otani | May 2005 | A1 |
20050159223 | Tahara | Jul 2005 | A1 |
20050164794 | Tahara | Jul 2005 | A1 |
20060062462 | Li | Mar 2006 | A1 |
20060170683 | Lin | Aug 2006 | A1 |
20060210172 | Sutanto | Sep 2006 | A1 |
20060218171 | Wakeam | Sep 2006 | A1 |
20060242349 | Ewanchuk | Oct 2006 | A1 |
20060245639 | Jiang | Nov 2006 | A1 |
20070183665 | Yuasa | Aug 2007 | A1 |
20070234878 | Worrall | Oct 2007 | A1 |
20080007199 | Kasaoka | Jan 2008 | A1 |
20080117214 | Perani | May 2008 | A1 |
20080165141 | Christie | Jul 2008 | A1 |
20080198165 | Sawazaki | Aug 2008 | A1 |
20080304748 | Koubaroulis | Dec 2008 | A1 |
20090021530 | Ishiguro | Jan 2009 | A1 |
20090136136 | Mori | May 2009 | A1 |
20090138830 | Borgaonkar | May 2009 | A1 |
20090148017 | Inoue | Jun 2009 | A1 |
20090160376 | Yamada | Jun 2009 | A1 |
20090288889 | Carlvik | Nov 2009 | A1 |
20090289902 | Carlvik | Nov 2009 | A1 |
20090309898 | Nakamura | Dec 2009 | A1 |
20100005428 | Ikeda | Jan 2010 | A1 |
20100074527 | Vukosavljevic | Mar 2010 | A1 |
20100137049 | Epstein | Jun 2010 | A1 |
20100283766 | Shieh | Nov 2010 | A1 |
20100292008 | Matsumura | Nov 2010 | A1 |
20110025693 | Merry | Feb 2011 | A1 |
20110046942 | Zheng | Feb 2011 | A1 |
20110066984 | Li | Mar 2011 | A1 |
20110122139 | Lee | May 2011 | A1 |
20110167988 | Berkovitz | Jul 2011 | A1 |
20110203442 | Raveendran | Aug 2011 | A1 |
20110285634 | Lim | Nov 2011 | A1 |
20110298709 | Vaganov | Dec 2011 | A1 |
20120024128 | Takahashi | Feb 2012 | A1 |
20120038735 | Hasegawa | Feb 2012 | A1 |
20120176324 | Akiyama | Jul 2012 | A1 |
20120308127 | Kudoh | Dec 2012 | A1 |
20130120237 | Diverdi | May 2013 | A1 |
20130127898 | DiVerdi | May 2013 | A1 |
20130276619 | Fujita | Oct 2013 | A1 |
20130293554 | Vostrikov | Nov 2013 | A1 |
20130319209 | Good | Dec 2013 | A1 |
20140000438 | Feis | Jan 2014 | A1 |
20140022194 | Ito | Jan 2014 | A1 |
20140035844 | Zhang | Feb 2014 | A1 |
20140043341 | Goel | Feb 2014 | A1 |
20140071076 | Dogra | Mar 2014 | A1 |
20140075549 | Lewis | Mar 2014 | A1 |
20140104201 | Tsutsui | Apr 2014 | A1 |
20140137029 | Stephenson | May 2014 | A1 |
20140160125 | Yoo | Jun 2014 | A1 |
20140171153 | Kienzle | Jun 2014 | A1 |
20140267302 | Lubetkin | Sep 2014 | A1 |
20140300554 | Samuel | Oct 2014 | A1 |
20140300563 | Iwaki | Oct 2014 | A1 |
20140325351 | Hirabayashi | Oct 2014 | A1 |
20140340333 | Takasugi | Nov 2014 | A1 |
20140355884 | Tran | Dec 2014 | A1 |
20140355885 | Shibata | Dec 2014 | A1 |
20150095822 | Feis | Apr 2015 | A1 |
20150139547 | Yamaji | May 2015 | A1 |
20150179156 | Uemura | Jun 2015 | A1 |
20150228259 | Valente | Aug 2015 | A1 |
20150229369 | Honary | Aug 2015 | A1 |
20150235391 | Makino | Aug 2015 | A1 |
20150302242 | Lee | Oct 2015 | A1 |
20150339524 | Yang | Nov 2015 | A1 |
20150346886 | Motoi | Dec 2015 | A1 |
20150346995 | Sugiura | Dec 2015 | A1 |
20150347835 | Baba | Dec 2015 | A1 |
20160048324 | Fukui | Feb 2016 | A1 |
20160070952 | Kim | Mar 2016 | A1 |
20160092021 | Tu | Mar 2016 | A1 |
20160098594 | Sugiura | Apr 2016 | A1 |
20160117548 | Hirakawa | Apr 2016 | A1 |
20160124630 | Qian | May 2016 | A1 |
20160125578 | Ohki | May 2016 | A1 |
20160147434 | Lee | May 2016 | A1 |
20160162174 | Tsutsui | Jun 2016 | A1 |
20160188970 | Matsuda | Jun 2016 | A1 |
20160210452 | Pahud | Jul 2016 | A1 |
20160364607 | VanBlon | Dec 2016 | A1 |
20160364892 | Huang | Dec 2016 | A1 |
20170004122 | Taniguchi | Jan 2017 | A1 |
20170243506 | Bayadzhan | Aug 2017 | A1 |
20170266558 | Rath | Sep 2017 | A1 |
20180060689 | Sklaroff | Mar 2018 | A1 |
20180061093 | Liu | Mar 2018 | A1 |
20180095653 | Hasek | Apr 2018 | A1 |
20180129367 | Hatfield | May 2018 | A1 |
20180158230 | Yan | Jun 2018 | A1 |
20180225535 | Yamamoto | Aug 2018 | A1 |
20190087654 | Xiao | Mar 2019 | A1 |
Number | Date | Country |
---|---|---|
63-095591 | Apr 1988 | JP |
1-96770 | Apr 1989 | JP |
05-135207 | Jun 1993 | JP |
5-225396 | Sep 1993 | JP |
8-180135 | Jul 1996 | JP |
10-55409 | Feb 1998 | JP |
10-105650 | Apr 1998 | JP |
2000-194796 | Jul 2000 | JP |
2001-184458 | Jul 2001 | JP |
2004-133951 | Apr 2004 | JP |
2007-200186 | Aug 2007 | JP |
2008-123182 | May 2008 | JP |
Entry |
---|
Mathcentre, Equations of straight lines, 2009, http://www.mathcentre.ac.uk/resources/uploaded/mc-ty-strtlines-2009-1.pdf (Year: 2009). |
Harvard, Multivariable calculus, 2011, http://www.math.kit.edu/ianm2/lehre/am22016s/media/distance-harvard.pdf (Year: 2011). |
International Search Report, dated Mar. 24, 2015, issued in PCT/JP2015/050875. |
Written Opinion of the International Searching Authority, dated Mar. 24, 2015, issued in PCT/JP2015/050875. |
English translation of the International Preliminary Report on Patentability and English translation of the Written Opinion of the International Searching Authority (PCT/IB/338, PCT/IB/373 and PCT/ISA/237), dated Sep. 29, 2016, for International Application No. PCT/JP2015/050875. |
English Translation of Office Action, dated Apr. 24, 2018, issued in corresponding Japanese Application No. 2014-070029. |
English Translation of Office Action, dated Mar. 6, 2018, issued in corresponding Japanese Application No. 2014-054010. |
Japanese Office Action for corresponding Japanese Application No. 2014-054010, dated Sep. 4, 2018, with English translation. |
Number | Date | Country | |
---|---|---|---|
20160202899 A1 | Jul 2016 | US |