1. Field of the Invention
The present invention relates to an action recognition apparatus, an action recognition system, and an action recognition method.
2. Description of the Related Art
Techniques of automatically recognizing an action of a subject by sensor measurement have been known. See, for example, Japanese Laid-Open Patent Application, Publication No. H10-113343 (to be referred to as “JP H10-113393” hereinafter), paragraph [0028] and FIG. 1; Japanese Laid-Open Patent Application, Publication No. 2008-165277, paragraphs [0005] to [0015], FIG. 1, and FIG. 2; Japanese Laid-Open Patent Application, Publication No. 2008-33544, FIG. 1 and FIG. 6; Japanese Laid-Open Patent Application, Publication No. 2007-310658, paragraphs [0025] and [0028] to [0032], FIG. 2, FIG. 8, and FIG. 12; Japanese Laid-Open Patent Application, Publication No. 2006-209468, paragraphs [0008] to [0015], FIG. 1, FIG. 2, and FIG. 3; Japanese Laid-Open Patent Application, Publication No. 2006-157463, paragraph [0008] and FIG. 1.
The term “subject” used herein means a human, an animal, a machine, or any other object whose state changes. The term “action” used herein means a state in which the subject is moving or changing.
JP H10-113343 discloses a technique as follows. A subject 9 is equipped with, for example, an action detection sensor 81 in the hip, arm, or any other body part which makes a characteristic movement, as shown in
The recognition processing unit 82 recognizes an action of the subject 9 by, for example, utilizing a recognition method using frequency analysis and referencing the dictionary database 83 in which frequency characteristics corresponding to action types (for example, walking) to be recognized of the subject 9. A result of the recognition is outputted to the outside via a recognized result output unit 84.
If the technique disclosed in JP H10-113343 is applied to a case where an action type to be recognized is only “walking”, an action recognition can be realized with a sufficient accuracy for practical use.
However, if the number of action types to be recognized is increased, those similar to “walking” are also likely to be increased. If such action types having similar characteristic amounts are registered in a database as a dictionary, less difference between the similar action types can result in less accuracy of recognition.
In that case, a recognition algorithm may be improved in order to enhance accuracy of recognition. This requires, however, an advanced recognition technique for recognizing a specific action type from among the action types having similar characteristic amounts, thus increasing load of calculation. If all actions made by the subject 9 are to be recognized, a large amount of action types to be recognized make a dictionary enormous. Further, if a difference in characteristic amounts is small, accuracy of recognition is lowered and a recognition algorithm becomes complicated as described above.
In light of the background, the present invention has been made in an attempt to provide an action recognition apparatus, an action recognition system, and an action recognition method, in each of which, even if the number of action types to be recognized is increased, accuracy of the recognition can be prevented from lowering in spite of existence of many similar action types.
In an action recognition apparatus, an action recognition system, and an action recognition method, action types to be recognized are narrowed down prior to a recognition processing; a dictionary and a recognition method for recognizing an action is selected, based on the narrowed-down action types; and then, an action recognition is performed.
Other features and advantages of the present invention will become more apparent from the following detailed description of the invention, when taken in conjunction with the accompanying exemplary drawings.
Exemplary embodiments for carrying out the present invention are described next in detail with reference to the related drawings according to the necessity.
In
The control unit 10 narrows down action types to be recognized, selects a dictionary and a recognition method for recognizing an action based on the narrowed-down action types, and performs an action recognition. The control unit 10 includes an estimation processing unit 11a, a selection processing unit 12, an action detection unit 13, a recognition processing unit 14, and a recognized result output unit 15.
Functions of the control unit 10 are embodied by, for example, developing and executing a program stored in the storage unit 20 of the action recognition apparatus 1a, in a RAM (Random Access Memory) by a CPU (Central Processing Unit).
The estimation processing unit 11a narrows down action types to be recognized and transfers the narrowed-down action types to the selection processing unit 12. In the action recognition apparatus 1a according to the first embodiment, an operator in charge of controlling a work narrows down the action types to be recognized via the input unit 30 and obtains information on the narrowed-down action types.
Based on the action types narrowed down by the estimation processing unit 11a, the selection processing unit 12 selects a dictionary to be referenced and a recognition method to be implemented by the recognition processing unit 14 from a recognition method/dictionary database 21 stored in the storage unit 20 and transfers the dictionary and the recognition method to the recognition processing unit 14. The term “action type” used herein means, for example, if contents of a work is an “installation”, an element of an action taken by a subject and characterizing the work contents, such as “walking” and “taking object in and out”, or the like (see
The action detection unit 13 is wiredly or wirelessly connected to an action detection sensor (not shown) attached to an arm, a waist, or any other parts of a subject (herein, an operator) whose action is a target of recognition. The action detection unit 13 obtains information detected by the action detection sensor. The action detection sensor is not limited to an acceleration sensor, but may be, for example, an angular velocity sensor, a position sensor, a displacement sensor, or any other sensor as long as it can measure an amount of physical change caused by movements of the subject's body. Further, the action detection sensor may have a memory therein. This allows the action detection unit 13 to obtain information stored in the memory from the input unit 30.
Description herein is made assuming that a well-known three-axis acceleration sensor is used as the action detection sensor. The three-axis acceleration sensor detects a force applied to a built-in spindle using strain sensors in the X, Y, and Z directions and measures an amount of acceleration using the obtained strain values.
The recognition processing unit 14 performs an action recognition processing based on the acceleration information of a subject's action detected by the action detection unit 13, according to the dictionary and the recognition method corresponding to the action types and selected by the selection processing unit 12. The recognition processing unit 14 then outputs the recognized information to the recognized result output unit 15.
The recognized result output unit 15 transfers the information on the recognized action type which is outputted as a result of the action recognition processing by the recognition processing unit 14, to the output unit 40.
The storage unit 20 is embodied by a hard disk, a flash memory, or the like. The storage unit 20 stores therein a recognition method/dictionary database 21.
The recognition method/dictionary database 21 stores therein a method of recognizing information obtained from an action detection sensor and a dictionary including information on characteristics of each action type to be referenced, corresponding to the recognition method. In the first embodiment, the recognition method/dictionary database 21 stores therein, for example, a method of recognizing data on acceleration such as frequency analysis, pattern matching, acceleration dispersion, and inclination angle, and information on characteristics of each action type to be referenced, corresponding to the recognition method (see
The input unit 30 is embodied by a keyboard, a touch panel, a memory card reader, or the like, into which information from the outside is inputted.
The output unit 40 is embodied by a display device for displaying a result of an action recognition processing, such as a liquid crystal display monitor, a drive device for outputting the processing result as information to an external storage medium, or the like.
Next is described a processing performed by the action recognition apparatus la according to the first embodiment with reference to
Description herein is made assuming that a three-axis acceleration sensor is attached to a right arm of a subject, and information is obtained by the action detection unit 13 by recognizing an action of the subject based on acceleration changes in the subject's right arm by FFT (Fast Fourier Transform), which is one of the recognition methods using frequency analysis.
As shown in
The recognition processing unit 14 converts the obtained acceleration data from the time series data to frequency distribution data by means of FFT (step S202). The recognition processing unit 14 then extracts a peak frequency from the converted frequency distribution data (step S203). The extracted peak frequency is a characteristic amount of the acceleration data obtained in step S201 in time series.
The recognition processing unit 14 retrieves an action type having the highest probability of occurrence of the extracted peak frequency using the recognition method/dictionary database 21 (step S204). The recognition processing unit 14 outputs the retrieved action type to the recognized result output unit 15 as a recognized result (step S205).
In
This is the problem occurred in a conventional action recognition apparatus. If a plurality of action types have similar characteristic amounts, the conventional action recognition apparatus has disadvantageously mixed recognition results, thus lowering accuracy of the recognition. For this reason, the action recognition apparatus 1a according to the first embodiment can narrow down the number of action types to be recognized by removing an unrelated action type which has a similar characteristic amount prior to an action recognition processing.
For example, if which work an operator as a subject carries out and that the work does not include the “tightening screw” is previously known, the action type “tightening screw” is omitted from action types to be recognized.
In the action recognition apparatus 1a according to the first embodiment, if an extracted peak frequency is in a position indicated by an arrowhead 43, the filled circle 40 is in a position more clearly belonging to “walking” because the curve 31 of “tightening screw” shown in
The recognition processing unit 14 obtains an acceleration data obtained from the action detection unit 13 by the prescribed window width (step S401). The obtained acceleration data in time series is referred to as I(n). The “n” is the number of the acceleration data obtained by the window width. If an acceleration sensor used has a single-axis, I(n) is represented as a vector having the element number “n”.
The recognition processing unit 14 acceleration computes a degree of similarity (a distance of the vector) of between an acceleration data of a time series pattern and a pattern registered in the recognition method/dictionary database 12 (to be described hereinafter as a dictionary pattern) (step S402). For simplification, assuming that the dictionary pattern has the same element number Pi(n) as that of the obtained acceleration data, a distance Di of the vector is indicated by Expression 1 as follows:
Di=|I(n)−Pi(n)| Expression 1
wherein “i” is a serial number of the dictionary.
The recognition processing unit 14 retrieves a dictionary pattern (i.e, an action type) having the shortest distance, using Expression 2 as follows (step S403):
W(I(n))=min Di=min|I(n)−Pi(n)| Expression 2
Herein, W(I(n)) is the shortest distance from the dictionary pattern. An action type of the dictionary pattern having the shortest distance is outputted as a result of recognition (step S404).
In
In the above description, Expressions 1 and 2 are exemplified assuming that, for simplification, the element number of the dictionary pattern is the same as that of the obtained acceleration data. However, the action recognition can also be made using DP (Dynamic Programming) matching or Hidden Markov model, an advanced technique of the DP matching, in which the element numbers are different.
Next are described in detail the operations of the action recognition apparatus 1a according to the first embodiment shown in
For example, an operator in charge of controlling a work of interest inputs an action type as narrowed-down information into the estimation processing unit 11a via the input unit 30 (step S601). The estimation processing unit 11a transfers the narrowed-down information to the selection processing unit 12.
The operator in charge determines an operation of which action type an operator as a subject is to perform and manually inputs the determined action type by inputting appropriate data in the input unit 30 embodied by, for example, a keyboard. Instead, the operator may input schedule information based on a current time and date in cooperation with a step management system or a schedule management system to be described later. More specifically with reference to a dictionary structure of
The selection processing unit 12 selects a recognition method and a dictionary based on the narrowed-down information inputted by the estimation processing unit 11a (step S602). For example, as shown in
In some cases, the input of the narrowed-down information eliminates the need for a recognition processing using the pattern matching of respective action types of, for example, “turning screw with driver”, “tightening screw with wrench”, “operating a jack”, and “documentation” shown in
Upon selection of the recognition method and the dictionary by the selection processing unit 12, the recognition processing unit 14 obtains the corresponding appropriate recognition method and a dictionary from the recognition method/dictionary database 21 and starts collection of acceleration data from the action detection unit 13 (step S603) The recognition processing unit 14 carries out an action recognition using the selected recognition method and dictionary if the number of the collected data reaches a window width sufficient to perform a calculation (step S604).
Only selection of a dictionary has been described above. However, selection of a necessary recognition method (or a processing procedure using a plurality of recognition methods) is herein assumed to be similarly made using the narrowed-down information inputted by the estimation processing unit 11a. Such selection makes it possible to carry out an action recognition with techniques using the FFT shown in
The recognition processing unit 14 transfers a result of the action recognition to the recognized result output unit 15. The recognition processing unit 14 outputs the recognition result to a liquid crystal display monitor or the like of the output unit 40 under controls of the recognized result output unit 15 (step S605).
After the output of the recognition result, the recognition processing unit 14 determines whether or not any change in an action type to be recognized by the operator is necessary (step S606). This is because a new action type which may appear according to newly obtained acceleration data because action contents may change over time. If a change in the action types is necessary (if Yes in step S606), the processing returns to step S601 and continues an input of narrowed-down information. If a change in the action types is not necessary (if No in step S606), the processing returns to step S603 in which the recognition processing unit 14 collects and obtains acceleration data.
Next are described the recognition result which the recognized result output unit 15 of the action recognition apparatus 1a displays in the output unit 40 as shown in step S605 of
In an area within a dotted circle 75 of
Meanwhile,
As described above, the action recognition apparatus 1a according to the first embodiment enhances accuracy of recognition by narrowing down action types to be recognized. Further, the action recognition apparatus la can reduce load of calculation because an unnecessary recognition processing can be omitted.
As shown in
The action recognition apparatus 1b according to the second embodiment is similar to the recognition apparatus 1a according to the first embodiment shown in
The step management device 60 includes: a work instruction information database (which may also be referred to as a schedule information database) 62 for storing steps of manufacturing a product in time series; a step management unit 61 for managing register, update, or the like of the work instruction information stored in the work instruction information database 62; and a communication unit 63 for communicating data with the action recognition apparatus 1b. The work instruction information shows a type of a work step (or work contents) outputted by the day or by the hour (see
The action recognition apparatus 1b includes a schedule correspondence database (schedule correspondence database) 22 (22b) provided in the storage unit 20, in addition to the configuration of the action recognition apparatus 1a according to the first embodiment of
The correspondence database 22b stores therein a correspondence relation between work contents included in the work instruction information obtained from the step management device 60 and an action type (see
An example of the work instruction information which is managed by the step management device 60 in the work instruction information database 62 is shown in
As shown in
For example, the operator A 90 has the work schedule of a trial assembly work 92 on Day 1, a welding work 93 from Day 2 to Day 3, and a painting work 94 from Day 4 to Day 5. Such a work schedule is registered in the work instruction information database 62 in time series. Note that
The work instruction information is outputted by the day or the hour. For example, a work instruction to the operator A on Day 4 is a painting work. The estimation processing unit 11b references the correspondence database 22b based on the work instruction information obtained from the step management device 60 via the communication unit 63, estimates a possible action type to be performed by the operator A, and transmits the possible action type to the selection processing unit 12.
As shown in
For example, a work to be performed by the operator A 90 on Day 4 is the painting work as shown in
In
In
If recognition of action types during an off-work period is desired, the action recognition apparatus 1b registers a possible action type estimated to be made before the start of the work or during the lunch break, in the correspondence database 22b. The action recognition apparatus 1b switches action types to be recognized for each prescribed time for switching work contents, using necessary processings by the estimation processing unit 11b and the selection processing unit 12.
As described above, in the action recognition system 100 according to the second embodiment, the action recognition apparatus 1b obtains the work instruction information for scheduling work contents in time series from the step management device 60, narrows down action types to be recognized based on the correspondence database 22b, and performs a recognition processing with the narrowed-down action types. This enhances accuracy of recognition. Further, by obtaining scheduled work contents in time series from the step management device 60, the action recognition apparatus 1b can selectively switch the most suitable recognition method and dictionary at a given point of time by the selection processing unit 12 and perform a recognition processing based on the selection.
In the action recognition system 100 according to the second embodiment, workload for performing the recognition processing can be reduced, because a manual input for narrowing down action types in advance of the recognition processing is not necessary.
In the second embodiment, description has been made assuming that, in the configuration of the action recognition system 100, the action recognition apparatus 1b is separate from the step management device 60. However, another configuration is also possible. The step management unit 61 and the work instruction information database 62 of the step management device 60 may be built in the action recognition apparatus 1b, which makes a single action recognition apparatus. Such a configuration can also obtain effects and advantages similar to those in the second embodiment.
In the action recognition system 100 according to the second embodiment, the work instruction information outputted by the step management device 60 is exemplified as the schedule information. However, another configuration is also possible in which a schedule management system for managing a schedule of actions of a person or movements of an object in time series is used as the schedule information.
For example, action types may be narrowed down using a scheduler indicating that a person conducts a work at or away from office. Another way of narrowing down the action types is to obtain, from a scheduler, information on whether a person is at or away from home, or on actions outside home (e.g. workout at a gym or mountain climbing). A still another way of narrowing down the action types is to utilize schedule information which is determined regardless of a person' s schedule, such as a train timetable. Such schedule information is used to determine whether a person makes an action on a train or at a station before getting on or getting off a train.
Besides the schedule management system, the action types can also be narrowed down by using pattern information which is information on common practice of a person such as a lifestyle pattern.
For example, if the pattern information on wake-up time, bedtime, and mealtime is used, the action types can also be narrowed down by using possible actions during sleep (e.g. roll-over and breathing conditions), during meal, or the like. This allows an enhanced accuracy of recognition. During a sleep time, a recognition algorithm for recognizing actions such as roll-over and breathing (for example, a pattern recognition of sleep apnea syndrome or the like) can be used to recognize an accurate sleep period. During a meal time, a recognition algorithm for recognizing actions such as movements of chopsticks or a fork can be used to recognize an accurate meal period.
As shown in
The action recognition apparatus 1 according to the third embodiment 1c is similar to the action recognition apparatus 1a according to the first embodiment shown in
The position detection device 70 is embodied by, for example, a device capable of detecting an absolute position on the earth as represented by the GPS (Global Positioning System), a positioning system in which a plurality of receivers receive radiowave from a transmitter and an arrival time of radiowave or a field intensity is utilized for detecting a position, a ranging system in which radiowave transmitted from a transmitter is received and a distance from the transmitter is estimated, or the like.
The position detection device 70 includes; a position detection unit 71 for detecting positional information by receiving radiowave from a transmitter attached to an operator; and a communication unit 72 for transmitting the positional information detected by the position detection unit 71, to the action recognition apparatus 1c.
The action recognition apparatus 1c also includes a correspondence database (which may also be referred to as a position correspondence database) 22c provided in the storage unit 20, in addition to the configuration of the action recognition apparatus la according to the first embodiment of
The correspondence database 22c stores therein positional information obtained from the position detection device 70 and work contents in a device at a position indicated by the positional information (see
In
The correspondence database 22c stores therein information indicating a correspondence relation between a type of a machine installed in the area and a section in which the machine is installed (that is, the positional information) as shown in
The correspondence database 22c stores therein, in addition to the information shown in
The position detection device 70 receives radiowave transmitted from the respective beacons 125, 126 attached to the operators 141, 142, via the receivers 121 to 124, measures distances from the respective beacons 125, 126 to the receivers 121 to 124, and detects in which sections 131 to 136 each of the operators 141, 142 is present.
The estimation processing unit 11c references the correspondence database 22c based on the positional information on the operator as a subject obtained from the position detection device 70 and narrows down action types to be recognized.
For example, if the position detection device 70 detects that the operator 141 is present in the section 133, the estimation processing unit 11c references the correspondence database 22c as shown in
The estimation processing unit 11c references the correspondence database 22c as shown in
The position detection device 70 may be an position sensor shown in
The example of
As described above, in the action recognition system 200 according to the third embodiment, the action recognition apparatus 1c obtains positional information of a subject from the position detection device 50, narrows down action types to be recognized, based on the correspondence database 22c, and performs a recognition processing with the narrowed-down action types. This enhances accuracy of recognition.
In the action recognition system 200, an appropriate action type can be estimated, making use of a result outputted by the position detection device 70. Thus, workload for performing a recognition processing can be reduced, because a manual input for narrowing down action types in advance of the recognition processing is not necessary.
In the third embodiment, description has been made assuming that, in the configuration of the action recognition system 200, the action recognition apparatus 1c is separate from the position detection device 70. However, another configuration is also possible. The position detection unit 71 of the position detection device 70 may be built in the action recognition apparatus 1c, which makes a single action recognition apparatus. Such a configuration can also obtain effects and advantages similar to those in the third embodiment.
In the action recognition system 200 according to the third embodiment, action types are narrowed down based on a position of a machine which is present at its manufacturing site. However, action types may be narrowed down based on a place or characteristics of a machine. For example, if a position sensor is attached to a vehicle and an operator is recognized to be on the vehicle, action types can be narrowed down to those related to a vehicle operation. For another example, if a subject who is equipped with a GPS or any other position detection device is recognized to be in an amusement park, action types can be narrowed down to those related to amusement rides. This further enables to recognize on which amusement ride the subject enjoys.
In the third embodiment, description has been made assuming that actions types to be recognized of a subject are narrowed down using the positional information. In the second embodiment, meanwhile, the actions types are narrowed down using the step management (which may also be referred to as a scheduler). However, both the scheduler and the positional information may be used for narrowing down the actions types. This allows the action types to be further narrowed down, enhances accuracy of recognition by the recognition processing unit 14, and reduces load of calculation.
As shown in
The action recognition apparatus 1d according to the fourth embodiment is similar to the action recognition apparatus 1b according to the second embodiment of
In order to narrow down action types, the action recognition apparatus 1d has a configuration in which a time for delimiting each work is detected, after which action types of the each work are narrowed down. For this purpose, the action recognition apparatus 1d includes: a characteristics database 23 provided in the storage unit 20; and a delimiting recognition processing unit 16 provided in the control unit 10, in addition to the configuration of the action recognition apparatus 1b according to the second embodiment.
The characteristics database 23 stores therein a characteristic action type of each work. For example, if the work contents is polishing, the characteristics database 23 stores therein “grinder” which is a characteristic action representative of the polishing.
The delimiting recognition processing unit 16 detects delimiting of consecutive works, using a characteristic action type stored in the characteristics database 23 and schedule information for scheduling work contents in time series. Details of the delimiting recognition processing unit 16 are described later.
Next are described in detail the operations of the action recognition system 300 according to the fourth embodiment shown in
In
The obtained work types are, for example, works to be done today in one day, as shown in
The selection processing unit 12 retrieves a characteristic action of each work from the characteristics database 23 (step S172). The characteristics database 23 stores therein a characteristic action type representative of a work. In
The characteristic action retrieved by the selection processing unit 12 and read by the characteristics database 23 is transferred to the recognition processing unit 14. The recognition processing unit 14 performs an action recognition processing based on the obtained characteristic action (step S173). The recognition processing unit 14 obtains a result of recognition and transfers the result to the recognized result output unit 15. The recognized result output unit 15 outputs the recognition result of characteristic actions 192, 193, 194 shown in
The delimiting recognition processing unit 16 detects a starting time and an ending time of the characteristic action recognized by the recognition processing unit 14. That is, the delimiting recognition processing unit 16 detects a time period between the starting time and the ending time (both of which may also be referred to as delimiting times) recognized by the recognition processing unit 14, during which the characteristic action was performed (step S174). The time period between the starting time and the ending time is thus regarded as a work time period during which each work is implemented.
More specifically, the delimiting recognition processing unit 16 determines delimiting times indicated by arrowheads 196, 197, 198, 199 in
As described above, the action recognition system 300 according to the fourth embodiment, the recognition processing unit 14 performs an action recognition based on a characteristic action type of each work contents stored in the characteristics database 23. This enables the delimiting recognition processing unit 16 to detect a delimiting time between continuous works. The delimiting recognition processing unit 16 determines delimiting times, for example, as indicated by the arrowheads 196, 197, 198, 199 in
In the action recognition system 300 according to the fourth embodiment, an estimated work time obtained from the step management device 60 as the work instruction information can be compared to an actual work time. The compared result can be fed back to the step management device 60, which enhances accuracy of step management made by the step management device 60.
In the action recognition system 300 according to the fourth embodiment, a delimiting time of each work is detected based on a characteristic action. However, the delimiting time may be detected using clustering or any other suitable technique.
In the fourth embodiment, description has been made assuming that, in the configuration of the action recognition system 300, the action recognition apparatus id is separate from the step management device 60. However, another configuration is also possible. The step management unit 61 and the work instruction information database 62 of the step management device 60 may be built in the action recognition apparatus 1d, which makes a single action recognition apparatus. Such a configuration can also obtain effects and advantages similar to those in the fourth embodiment.
The embodiments according to the present invention have been explained as aforementioned. However, the embodiments of the present invention are not limited to those explanations, and those skilled in the art ascertain the essential characteristics of the present invention and can make the various modifications and variations to the present invention to adapt it to various usages and conditions without departing from the spirit and scope of the claims.
Number | Date | Country | Kind |
---|---|---|---|
2008-328404 | Dec 2008 | JP | national |
This application claims the benefit of Japanese Patent Application No. 2008-328404 filed on Dec. 24, 2008, the disclosure of which is incorporated herein by reference.