This application also claims priority to Taiwan Patent Application No. 101149021 filed in the Taiwan Patent Office on Dec. 21, 2012, the entire content of which is incorporated herein by reference.
The present disclosure relates to a workflow monitoring and analysis system and method, and more particularly, to a method and system capable of generating a task syntagm to be used in a hybrid automation means.
With rapid advance of technology, there are more and more handheld electronic devices that are becoming available and commonly used in our daily lives, such as smart phones, tablet computers and notebook computers. Nevertheless, also because of their variety in design and style, many recent electronic devices can not be produced completely by an automation process, but still require plenty of manpower for assembly.
As a consequence, for increasing production and reducing cost, there are more and more studies in the industry trying to design a hybrid automation system capable of combining tasks that are needed to be accomplished accurately and rapidly and being executed by robots with tasks that are high complicated and needed to be performed by human into a same production line, and thereby, enjoying the benefic of both robotic assembly and manual assembly simultaneously.
However, in most workflows enabled in current hybrid automation systems there is no visual recognition apparatus being provided for monitoring the movement of both robots and human, whereas the movement of human operators are especially difficult to measure and quantified. In most cases, certain kinds of artificial intelligence will be needed just to identify the meaning of operator's hand movements, otherwise the workflow including alternating manual procedures and automated robotic procedures can not be performed smoothly. Therefore, it is in need of an improved workflow monitoring and analysis apparatus and method adapted for hybrid automation.
In an exemplary embodiment, the present disclosure provides a workflow monitoring and analysis method, which comprises the steps of: generating at least one three-dimensional joint coordinate according to an image information so as to be used for generating at least one workpiece posture information accordingly, and further according to workpiece posture information to generate a task posture information; generating at least one three-dimensional track information according to a movement information so as to be used for generating at least one workpiece track information accordingly, and further according to the workpiece track information to generating a task track information; and generating a task syntagm according to the task posture information and the task track information.
In another exemplary embodiment, the present disclosure provides a workflow monitoring and analysis, which comprises: a detection module; at least one first image capturing module; at least one second image capturing module; and a workflow analysis unit, electrically and respectively coupled to the at least one first image capturing module and the at least one second image capturing module, and further comprised of: an image recognition module, a detection zone posture module, a detection zone movement module and a task model database; wherein, the at least one first image capturing module is used for capturing images of the detection module so as to generate an image information accordingly; the at least one second image capturing module is used for capturing movements of the detection module so as to generate a movement information accordingly; the image recognition module is used for identifying and recognizing the image information and the movement information; the detection zone posture module is used for receiving the image information to be used for generating at least one feature point accordingly; the detection zone movement module is provided for receiving the movement information so as to be used for generating at least one three-dimensional track information accordingly; the task model database is enabled to generate at least one task posture information according to the at least one feature point and is also being enabled to generate at least one task track information according to the at least one three-dimensional track information, and then according to the at least one task posture information and the at least one task track information, the workflow model database is enabled to generate a task syntagm.
Further scope of applicability of the present application will become more apparent from the detailed description given hereinafter. However, it should be understood that the detailed description and specific examples, while indicating exemplary embodiments of the disclosure, are given by way of illustration only, since various changes and modifications within the spirit and scope of the disclosure will become apparent to those skilled in the art from this detailed description.
The present disclosure will become more fully understood from the detailed description given herein below and the accompanying drawings which are given by way of illustration only, and thus are not limitative of the present disclosure and wherein:
In the following detailed description, for purposes of explanation, numerous specific details are set forth in order to provide a thorough understanding of the disclosed embodiments. It will be apparent, however, that one or more embodiments may be practiced without these specific details. In other instances, well-known structures and devices are schematically shown in order to simplify the drawing.
Please refer to
In addition, an exemplary first image capturing module 20 is a charge-coupled device (CCD), and there can be one or a plurality of such first image capturing modules 20 being installed in the apparatus of the present disclosure that are used for acquiring images of the detection module 1. One the other hand, the second image capturing module 21 can be a depth camera, such as an infrared camcorder or camera, and similarly there can be one or a plurality of such second image capturing modules 21 being installed in the apparatus of the present disclosure that are used for capturing image depth or moving track of the detection module 1
The workflow analysis unit 3. which is electrically and respectively coupled to the first image capturing module 20 and the second image capturing module 21, is composed of an image recognition module 30, a detection zone posture module 31, a detection zone movement module 32, a workpiece posture module 33, a workpiece movement module 34, a task model database 35 and an output module 36.
Wherein, the task model database 35 has a plurality of base shape information and a plurality of mid-level shape information registered therein. In an embodiment, the plural primitive shape information can include a number of primitive hand gestures that can be distinguished from one another by the extent of finger bending, whereas the mid-level shape information can include hand gestures of reversing palm for instance. However, each of those base shape information and mid-level shape information is composed of a plurality of three-dimensional joint coordinates.
Operationally, the first image capturing module 20 is enabled to acquire images of the detection module 1 so as to generate an image information; the second image capturing module 21 is enabled to capturing depth or movement of the detection module 1 so as to generate a movement information; the image recognition module 30 is used for identifying and recognizing the image information and the movement information; the detection zone posture module 31 is used for receiving the image information of the detection module 1 to be used for detecting and generating at least one feature point from the detection zones 10 accordingly, and then the detection zone posture module 31 is further being enabled to receive at least one primitive shape information from the task model database 35 to be used in a comparison with the feature point while consequently transmitting the comparison result to the task model database 35 for enabling the task model database 35 to perform the following procedures: acquiring a mid-level shape information according to the comparison result in a successive approximation approach manner; performing an angle calibration procedure upon the comparison result according to the mid-level shape information so as to generate an angle calibration result; generating at least one three-dimensional joint coordinate according to the angle calibration result; and generating at least one task posture information according to the at least one three-dimensional joint coordinate. Thereafter, the workpiece posture module 33 is enabled to generate at least one workpiece posture information according to the image information and then transmitting the workpiece posture information to the task model database 35.
Moreover, the detection zone movement module 34 is provided for receiving the movement information so as to be used for performing a measurement upon a region accordingly, and then the detection zone movement module 34 is enabled to generate at least one three-dimensional track information according to the aforesaid three-dimensional joint coordinate and the measurement result relating to the region. Thereafter, the workpiece movement module 34 is enabled to generate at least one workpiece track information according to the movement information and than transmitting the workpiece track information to the task model database 35. Thereby, the task model database 35 to generate at least one task track information according to the three-dimensional track information and is able to further define a task posture and a task track, and the like.
In addition; the task model database 35 is enabled to generate at least one task model data according to the task track, the task posture, the workpiece posture information and the workpiece track information so as to be used in the generating of the task syntagm according to the combination of more than one so-generated task model data. The task syntagm is used for recognizing and describing movements of an on-line operator, and can substantially a kind of makeup language. The task syntagm is outputted by the output module 36.
Please refer to
In an embodiment, a detection zone posture module 31 is used to perform a comparison upon the image information of S1 so as to generate at least one three-dimensional joint coordinate accordingly, and thereby, the at least one three-dimensional joint coordinate is used in the generation of at least one task posture information. Please refer to
S11: the image information is inputted into a detection zone posture module 31;
S12: a region measurement operation is performed by the detection zone posture module 31 according to the image information, whereas when the detection module 1 is a glove that is worn on a hand, the region measurement operation is a measurement performed on an area where the hand is moving;
S13: the detection zone module 31 is enabled to acquire at least one feature point out of the detection zones 10 in the image information, and in an embodiment, there can be 20 to 60 feature points, whereas each of the feature points can be established using a color distribution analysis means to determine whether the colors in the detection zones 10 are included in a color histogram that is stored in the task model database 35 and if so, the detection zones are identified as target regions to be used for establishing feature points;
S14: the task model database 35 is enabled to provide at least one primitive data to the detection zone posture module 31, and then the detection zone posture module 31 compares the at least one primitive data with the feature points so as to generate and transmit a comparison result to the task model database 35, in that the comparison is a similarity test comparing the orientations and positions between the feature points of the image information to those of the primitive data, and in an embodiment, if an image geometrical distance between the feature point of the image information and the related feature point of the primitive data is shorter than a specific distance, the similarity between the image information to the primitive data can be established, and moreover, the image geometrical distance is defined to be the Euclidean distance between two feature points, that is the real distance between two points in a space;
S15: the task model database 35 is enabled to acquire a mid-level shape information according to the comparison result in a successive approximation approach manner;
S16: the task model database 35 is enabled to perform an angle calibration procedure upon the comparison result according to the mid-level shape information so as to generate an angle calibration result;
S17: the task model database 35 is enabled to generate at least one three-dimensional joint coordinate according to the angle calibration result; and
S18: the task model database 35 is enabled to generate at least one task posture information according to the at least one three-dimensional joint coordinate while enabling the at least one three-dimensional joint coordinate to be stored in the task model database 35.
In addition, in an embodiment, a detection zone movement module 32 is used to generate at least one three-dimensional track information according to the movement information, and thereby, the at least one three-dimensional track information is used in the generation of at least one task track information. Please refer to
S19: the detection zone movement module 32 is enabled to receive the movement information of S1;
S20: the detection zone movement module 32 is enabled to perform a region measurement operation according to the movement information;
S21: the detection zone movement module 32 is enabled to fetch the three-dimensional joint coordinate of S2 according to the measurement result;
S22: an evaluation is made by the detection zone movement module 32 according to the three-dimensional joint coordinate to determine whether the movement information is a task starting coordinate or is a task terminating coordinate, and if it is determined to be either the task starting coordinate or the task terminating coordinate, a three-dimensional track information is generated accordingly and the flow proceeds to step 23, otherwise, the flow proceeds back to step 21 for the fetching of the three-dimensional joint coordinate; and moreover, the three-dimensional track information includes actions and the time stamps of those actions and thereby, the exact action and position of the detection module 1 at the time t can be identified;
S23: the detection zone movement module 32 is enabled to fetch another three-dimensional joint coordinate for updating the same, and then the flow proceeds back to step 21 for repeating;
S24: an evaluation is made to determine whether the flow is ended, and if so, the flow proceeds to step 25 for enabling the task model database 35 to generate a task track information according to the three-dimensional track information while allowing the task track information to be stored in the task model database 35; otherwise, the flow proceeds back to step S21.
To sum up, the present disclosure provides a method and apparatus for detecting postures and movements of a detection module, and then strengthening the semantic accuracy of movement recognition and correctness of movement identification through the use of a posture reconstruction technique, so as to generate a task syntagm accordingly.
With respect to the above description then, it is to be realized that the optimum dimensional relationships for the parts of the disclosure, to include variations in size, materials, shape, form, function and manner of operation, assembly and use, are deemed readily apparent and obvious to one skilled in the art, and all equivalent relationships to those illustrated in the drawings and described in the specification are intended to be encompassed by the present disclosure.
Number | Date | Country | Kind |
---|---|---|---|
101149021 A | Dec 2012 | TW | national |
Number | Name | Date | Kind |
---|---|---|---|
5423554 | Davis | Jun 1995 | A |
5454043 | Freeman | Sep 1995 | A |
5852672 | Lu | Dec 1998 | A |
5875108 | Hoffberg et al. | Feb 1999 | A |
6128003 | Smith et al. | Oct 2000 | A |
6222465 | Kumar et al. | Apr 2001 | B1 |
6256033 | Nguyen | Jul 2001 | B1 |
6477239 | Ohki et al. | Nov 2002 | B1 |
6507675 | Lee | Jan 2003 | B1 |
6564144 | Cherveny | May 2003 | B1 |
6963827 | Elyea et al. | Nov 2005 | B1 |
7225414 | Sharma et al. | May 2007 | B1 |
7239718 | Park et al. | Jul 2007 | B2 |
7353082 | Pretlove et al. | Apr 2008 | B2 |
7402743 | Clark et al. | Jul 2008 | B2 |
7565295 | Hernandez-Rebollar | Jul 2009 | B1 |
7606411 | Venetsky et al. | Oct 2009 | B2 |
7853041 | Shamaie | Dec 2010 | B2 |
8139110 | Nishihara | Mar 2012 | B2 |
9318108 | Gruber et al. | Apr 2016 | B2 |
20030219704 | Bleckley et al. | Nov 2003 | A1 |
20040057619 | Lim | Mar 2004 | A1 |
20050234310 | Alwan | Oct 2005 | A1 |
20050256611 | Pretlove et al. | Nov 2005 | A1 |
20060143034 | Rothermel et al. | Jun 2006 | A1 |
20070219738 | Weiher et al. | Sep 2007 | A1 |
20080004904 | Tran | Jan 2008 | A1 |
20080044084 | Lee | Feb 2008 | A1 |
20080247506 | Maschke | Oct 2008 | A1 |
20090006172 | Singh et al. | Jan 2009 | A1 |
20090173779 | Szesko et al. | Jul 2009 | A1 |
20100194679 | Wu et al. | Aug 2010 | A1 |
20100194744 | Glueck et al. | Aug 2010 | A1 |
20100234994 | Shi | Sep 2010 | A1 |
20110026833 | Sugino et al. | Feb 2011 | A1 |
20110270123 | Reiner | Nov 2011 | A1 |
20110274311 | Lee et al. | Nov 2011 | A1 |
20110295399 | Plociennik et al. | Dec 2011 | A1 |
20120062725 | Wampler, II et al. | Mar 2012 | A1 |
20130218340 | Hager | Aug 2013 | A1 |
20140300722 | Garcia | Oct 2014 | A1 |
Number | Date | Country |
---|---|---|
102053575 | May 2011 | CN |
10214346 | Nov 1998 | JP |
2006099158 | Apr 2006 | JP |
393629 | Jun 2000 | TW |
I281024 | May 2007 | TW |
201009766 | Mar 2010 | TW |
201030630 | Aug 2010 | TW |
201203131 | Jan 2012 | TW |
M443156 | Dec 2012 | TW |
Entry |
---|
Kimura et al., “Task-model based human robot cooperation using vision”, Proceedings of the 1999 IEEE/RSJ International Conference on Intelligent Robots and Systems, 1999, IEEE. |
Lambrecht et al., Markerless Gesture-Based Motion Control and Programming of Industrial Robots, Emerging Technologies & Factory Automation (ETFA), 2011 IEEE 16th Conference, 2011, pp. 1-4. |
Chen, Realization of Finger Recognition on Virtual Mouse Application, Institute of Industrial Design of National Yunlin University of Science & Technology, 2010. |
Fernandez-Y-Fernandez et al, Integrated Development Environment Gesture for modeling workflow diagrams, Congreso Internacional de Investigación e Innovación en Ingenieria de Software, 2012, pp. 1-8, Guadalajara, Jalisco, del 25 al 27 de abril. |
Kimura et al., Task-Model Based Human Robot Cooperation Using Vision, Intelligent Robots and Systems, 1999. IROS '99. Proceedings, 1999 IEEE/RSJ International Conference, 1999, pp. 701-706, vol. 2. |
Chen, Application of Mixture of Experts Model to Sign Language Gesture Recognition, Southern Taiwan University Of Science and Technology, 2010. |
Che-Wei Liu et al., Static Hand Posture Recognition Based on an Implicit Shape Model, Department of Electrical Engineering College of Electrical Engineering and Computer Science, National Taiwan University Master Thesis, 2010, Taiwan. |
Wen-Hui Chen et al., A Generic Framework for the Design of Visual-based Gesture Control Interface, 5th IEEE Conference on Industrial Electronics and Applicationsis, 2010, p. 1522-1525. |
Yu, Intelligent Hand Gesture Recognition System Design Based On Hand Contour, 2010. |
Intellectual Property Office, Ministry of Economic Affairs, R.O.C., “Office Action”, Nov. 21, 2014, Taiwan. |
Number | Date | Country | |
---|---|---|---|
20140180743 A1 | Jun 2014 | US |