This application claims the benefit of Japanese Application No. 2017-203978 filed in Japan on Oct. 20, 2017, the contents of which are incorporated herein by this reference.
This invention relates to a work support system, an image pickup apparatus, a wearable apparatus and a work support method for supporting predetermined work, a work procedure of which is prescribed.
Conventionally, various industrial products (manufactured articles) have widely been prevalent. Examples of types of work for various industrial products (manufactured articles) include various types of work such as work of manufacturing and assembling in a manufacturing process, work of using such manufactured articles (handling work) and maintenance work (e.g., maintenance, inspection, adjustment and repair) for the manufactured articles. In many cases, respective predetermined work procedures, methods, etc., are clearly prescribed for the respective types of work. Some of these types of work are performed by means of manual operation by operators.
Each of these types of work includes a wide variety of tasks for each manufactured article, which is an object. Each of the tasks may involve a complicated procedure, or as necessary, need a confirmation task. Therefore, in order to reliably perform prescribed work, an operator accurately memorizes a whole predetermined work procedure or proceeds the work while referring to a manual such as an operating manual.
However, a lot of skills are required to reliably memorize all of a predetermined series of tasks and always perform the tasks correctly. When an operator pursues work referring to a manual such as an operating manual, for example, the operator stops the work each time the operator refers to references, resulting in hindrance of smooth work and consumption of a length of time beyond necessity.
Therefore, for work support systems for supporting an operator when the operator performs predetermined work, a work procedure of which is clearly prescribed, various proposals have been made by, for example, Japanese Patent Application Laid-Open Publication No. 2017-75001, etc.
The work support system disclosed in Japanese Patent Application Laid-Open Publication No. 2017-75001 above is a work support system for guiding a work procedure for a mobile crane. The work support system is configured using, for example, a wearable-type observation apparatus in order to correctly guide an operator according to a predetermined work procedure.
The work support system includes a remote operation apparatus including a detection apparatus configured to detect a state of a work machine, a transmission section configured to transmit the state of the work machine detected by the detection apparatus, an operation section configured to remotely operate the work machine, a display section configured to display a work procedure and a reception section configured to receive the state of the work machine transmitted by the transmission section, and the work support system is configured so that the display section provided in the remote operation apparatus displays a prescribed work procedure based on the received state of the work machine.
Such configuration enables the above work support system to arbitrarily display a detailed explanation of a relevant work content of each work process on a display screen of the observation apparatus, and thus, an operator can always follow a work procedure correctly without memorizing the work procedure in advance.
A work support system according to an aspect of the present invention includes: an image pickup apparatus configured to acquire image data; a control apparatus configured to acquire pieces of process information chronologically managed in a database; and a recording apparatus configured to record the image data acquired by the image pickup apparatus, in association with meta data including a piece of process information corresponding to the image data, from among the pieces of process information acquired by the control apparatus, and shooting time information.
An image pickup apparatus according to an aspect of the present invention includes: an image pickup unit configured to acquire image data; a communication apparatus configured to communicate with a database in which work information including image information relating to predetermined work is accumulated in advance; a control apparatus configured to acquire pieces of information on work performed; and a recording apparatus configured to record the image data acquired by the image pickup unit, in association with meta data including a piece of work information corresponding to the image data from among the pieces of work information acquired by the control apparatus.
A wearable apparatus according to an aspect of the present invention includes: an image pickup unit configured to acquire image data corresponding to an operator's visual line; a communication apparatus configured to communicate with a database in which work information including image information relating to predetermined work is accumulated in advance; a control apparatus configured to acquire pieces of work information on work performed; and a recording apparatus configured to record the image data acquired by the image pickup unit, in association with meta data including a piece of work information corresponding to the image data from among the pieces of work information acquired by the control apparatus.
A work support method according to an aspect of the present invention includes: a process information acquisition step of acquiring particular pieces of process information from a plurality of pieces of process information chronologically managed in a database; an image acquisition step of acquiring image data; and a recording step of recording the image data, in association with meta data including a piece of process information corresponding to the image data and shooting time information.
Benefits of this invention will be further clarified by the following detailed description.
The present invention will be described below based on the embodiment illustrated in the drawings. Each of the drawings used for the below description is a schematic one, and in order to illustrate respective components in sizes that are large enough to be recognized in the drawings, the components may be illustrated so as to be different in, e.g., scale and dimensional relationship between members. Therefore, the present invention is not limited to the illustrated forms in terms of, e.g., the number of the components, shapes of the components, ratios in size among the components and relative positional relationships among the respective components indicated in each drawing.
First, a schematic configuration of a work support system according to an embodiment of the present invention will be described below with reference to
The work support system 1 according to the present embodiment is a work support system that provides work support for reliably performing predetermined sequential work including a plurality of tasks according to a prescribed procedure and method and can record evidence relating to each of the tasks performed.
Normally, a strict work procedure is prescribed and the work procedure needs to be followed faithfully for this type of work. For example, in, e.g., the medical field, where, e.g., work for cleaning an instrument used for various kinds of treatment, surgery or the like is performed, a series of tasks including a task of cleaning a predetermined part of an object instrument or the like and a subsequent observation task of confirming a state of the cleaning of the predetermined part is repeated. Then, this type of work needs to be smoothly and efficiently performed to reduce risks of contamination.
Furthermore, in recent years, a system capable of recording evidence information such as recording, as evidence information indicating that each of a series of tasks has reliably been performed as prescribed, e.g., a state of an object before and after the task and progress of the task in the form of images has been desired as a countermeasure for responding to various types of troubles, for example.
Therefore, the work support system 1 includes an image pickup apparatus and a control apparatus configured to control the image pickup apparatus. In the present embodiment, an example of the work support system 1 including an image pickup apparatus (wearable terminal apparatus 20) and a control apparatus 10 is illustrated as in
Based on a collection of pieces of information relating to a plurality of tasks determined in advance (database) and a plurality of image data relating to actually-performed tasks (acquired image data), such a configuration allows the work support system 1 according to the present embodiment to perform confirmation and determination relating to the tasks such as displaying individual work contents included in the relevant sequential work and a work procedure to provide a work guide and determining whether or not the work procedure has correctly been followed, and provide work support such as performing evidence recording if the work has been performed correctly, and issuing a warning to an operator to remind the operator to perform the work correctly if the work has been performed erroneously.
More specifically, the work support system 1 according to the present embodiment includes a wearable terminal apparatus 20, which is an image pickup apparatus as well as a wearable apparatus, and a control apparatus 10 configured to control the wearable terminal apparatus 20.
The control apparatus 10 controls the wearable terminal apparatus 20, performs confirmation and determination for each of tasks performed, based on image data (still image data and movie data) acquired by the wearable terminal apparatus 20, and performs operation control such as task recording (evidence recording).
For such purpose, the control apparatus 10 includes, e.g., a control section 11, a communication section 12, a database 13, an evidence recording section 14, a display section 15 and a clock section 16.
In addition to the components stated above, the control apparatus 10 according to the present embodiment includes various components included in a general control apparatus, for example, an operation section although illustration of such components are omitted. Various components, illustration of which are omitted here, are parts that are not directly related to the present invention and thus detailed description of such components will be omitted.
The control section 11 is a component section including a plurality of control circuits, etc., configured to perform overall control of the entire work support system 1 according to the present embodiment. The plural control circuits, etc., included in the control section 11 specifically include, e.g., a work determining section 11a, a display control section 11b, a guide section 11c, an object determining section 11d, a size determining section 11e and a falsification preventing section 11f.
The work determining section 11a is a component section including, e.g., a control circuit configured to, based on, e.g., image data acquired by the wearable terminal apparatus 20, information accumulated in advance in the database 13 and determination result information acquired by the object determining section 11d and the size determining section 11e, perform, e.g., confirmation or determination for each of items relating to work performed. Examples of the determinations performed here include, e.g., determination of a type of a tool to be used for work and determination of a type of each task.
The display control section 11b is a component section including, e.g., a control circuit for controlling, e.g., the display section 15 included in the control apparatus 10 or a display section 25 included in the wearable terminal apparatus 20, the display section 25 operating in cooperation with the control apparatus 10, to provide predetermined display (e.g., warning display or guide display in addition to image display).
The guide section 11c is a component section including, e.g., a control circuit configured to, based on image data acquired by the wearable terminal apparatus 20 and information accumulated in advance in the database 13, create guide information for providing, e.g., guide display relating to work performed or guide display for a task to be performed next.
The object determining section 11d and the size determining section 11e are component section including, e.g., control circuits configured to, based on image data acquired by the wearable terminal apparatus 20 and information accumulated in advance in the database 13, detect and determine predetermined conditions such as a type, a part, a shape, a size, a number and/or a state of a predetermined object included in the image data. As a size determining method other than the aforementioned method, for example, use of a method in which a size of an object, an image of which is formed on a light-receiving surface of an image pickup device, is calculated using an autofocus mechanism, based on a relationship between a distance between the relevant image pickup apparatus and the object and a size of the image pickup device, is conceivable.
The falsification preventing section 11f is a processing circuit section for, e.g., protecting, and preventing falsification of, image data generated and recorded in the wearable terminal apparatus 20 (image pickup apparatus) and transferred to the control apparatus 10 via communication sections (22, 12), by subjecting the image data to particular data processing.
For example, the falsification preventing section 11f is a circuit section configured to subject image data recorded by the wearable terminal apparatus 20 and transferred to the control apparatus 10 to, e.g., particular data processing for permitting reproduction of the image data only if a particular image display apparatus or dedicated application software is used.
In addition to the data processing such as above, the falsification preventing section 11f performs, for example, data encryption processing and falsification detection hush function creation processing and other data processing such as file format conversion processing and data locking processing. For a technique for preventing falsification of data recorded on a recording medium, a conventionally well-known technique is used and thus further detailed description of such technique will be omitted.
The communication section 12 is a transmission/reception section including, e.g., a control circuit configured to perform data communication with (information transmission/reception to/from) the communication section 22 of the wearable terminal apparatus 20. Here, the communication section 12 may have a configuration using wireless communication or may have a configuration using wired communication.
The database 13 is a storage apparatus (auxiliary storage apparatus) configured to accumulate various kinds of information determined in advance. In the present embodiment, an example in which the database 13 is included in the control apparatus 10 (more specifically, for example, an internally-fixed recording medium (e.g., a built-in memory) or a removable recording medium (e.g., a removable card-type memory)). However, a configuration of the database 13 is not limited to these forms. For example, the database may be configured by an external storage apparatus having an independent form. In this case, a configuration in which various types of data can be transmitted/received by predetermined wireless or wired connection means may be provided between the database configured independently and the control apparatus 10.
The database 13 includes, for example, a work information database 13a and an image database 13b. Here, the work information database 13a contains work information including various kinds of information relating to a plurality of types of work. Here, work information includes individual pieces of detailed information relating to each type of work such as work names (e.g., a general name/process names/detailed task names), a work object, work parts, work time information (e.g., start/end times and/or a total length of time), surrounding environment information (e.g., work location/building name/room name/workbench position) and an operator name.
Such work information can be broken down to a plurality of processes included in each type of work. Furthermore, each process is detected as a work content (work element) that can be confirmed via an image, which enables checking or ensuring reliability of the work.
Reliability of each process or each task is determined by confirming whether or not each work element matches a content of an actual picked-up image of a task corresponding to the work element, details of which will be described later. Each work element is information that is not separate from information of a relevant process but information included in the information of the relevant process.
Each work content represented by work information is a collection of “work elements” resulting from the work content being broken down into objects or movements that can easily be determined based on results of image pickup by the image pickup apparatus (image data). Inclusion of such plural “work elements” in the database enables breaking down image determination into simple processes.
In other words, clearing each of these work elements (that is, if a result of determination is OK) enables determining whether or not each process is properly being performed or properly providing guide display such as advice related to each process at a proper timing Whether or not acquired images are images of respective relevant processes can be determined. Furthermore, if determinations that respective processes have properly been performed (results of “OK”) can be made in a particular order, the work to be recorded can be ensured in total. For, e.g., determination before and after a task, the work element information may be used as necessary.
The image database 13b is a collection of image data related to various kinds of information recorded in the work information database 13a (for example, image information of images relating to a plurality of tasks from a plurality of field of views) (details will be described later; see
The evidence recording section 14 is a component section including a control circuit or the like and a recording medium or the like configured to record, e.g., image data acquired by, e.g., the wearable terminal apparatus 20 and work information related to the image data, in association with each other. Images from more fields, ranges and points of views may be made available for reference.
The display section 15 is a component unit including, e.g., a display panel and a drive circuit for the display panel, the display panel and the drive circuit being capable of displaying a screen for various settings, displaying, e.g., image data recorded in the evidence recording section 14 or image data transferred from the wearable terminal apparatus 20, and displaying, e.g., various kinds of information (meta data) accompanying image data that is being displayed.
The clock section 16 is an internal clock called “real-time clock (RTC)”. The clock section 16 is a circuit section used for, for example, providing date information and time information to, e.g., a data file, timing an interval between predetermined instruction timings during control processing, and temporal control.
The wearable terminal apparatus 20 is, for example, as illustrated in
In other words, the wearable terminal apparatus 20 is an image pickup apparatus of a type that enables acquisition of work information corresponding to the operator's visual line. In other words, an image displayed based on image data acquired by the wearable terminal apparatus 20 can be used as evidence of the operator viewing such a scene.
The wearable terminal apparatus 20 not only mainly acquires movie data, but also acquires still image data at an arbitrary timing. Movie data acquired by the wearable terminal apparatus 20 is used to grasp, e.g., a flow of successive tasks performed by an operator and passage of time consumed for the tasks and record such time data, etc., as evidence information. In the case of still image data, the still image data may be of a type with information based on which temporal change can be determined, for example, data information representing a temporal change of an image such as a motion vector, added. When the still image data is used as evidence information, it is preferable that the still image data be still image data that can be verified later by a human being.
Such images have an advantage of good searchability when the image is displayed or printed because of the image having a large amount of information and an advantage of easiness of finding a falsification. Linking various kinds of information to image data enables enhancement in significance and value of the image and further enables the image data to be treated as evidence or a trail. An image has a large amount of information and thus is effective for trail information recording for determination of a particular situation, with an image determination function and image recording brought in cooperation with each other.
Here,
In the present embodiment, the wearable terminal apparatus 20 is an image pickup apparatus configured to operate alone under the control of the control section 21 and also operates under the control of the control section 11 of the control apparatus 10 via communication with the control apparatus 10.
In addition to movie data acquisition, the wearable terminal apparatus 20 in the present embodiment has, for example, a function that can arbitrarily or automatically acquire still image data simultaneously with movie data or still image data alone with a predetermined operation or a predetermined action performed by an operator as a trigger.
The wearable terminal apparatus 20 includes, e.g., the control section 21, the communication section 22, an operation section 23, an image pickup section 24, the display section 25, a recording section 26 (not illustrated in
The control section 21 is a component section including, e.g., a control circuit configured to perform overall control of the entire wearable terminal apparatus 20. The control section 21 is configured by, for example, a processor using, e.g., a central processing unit (CPU). The control section 21 operates according to programs stored in a memory (not illustrated) to perform control of the respective component sections.
In other words, the control section 21 controls the communication section 22 to perform control to transmit/receive various kinds of information to/from (communicate with) the control apparatus 10. The control section 21 performs control of the respective component sections in response to an operation input from the operation section 23. The control section 21 controls the image pickup section 24 to perform control to acquire, e.g., image data. The control section 21 controls the display section 25 to perform control to display various kinds of information on a screen of a display panel 25b (described later; see
In addition to these types of control, the control section 21 controls the wearable terminal apparatus 20 in cooperation with the control apparatus 10 upon receipt of a control signal from the control section 11 of the control apparatus 10.
The control section 21 includes, e.g., a plurality of circuit sections, for example, an image determining section 21b and a display control section 21c.
The image determining section 21b is a determination processing circuit configured to perform predetermined determination based on image data acquired by the image pickup section 24. The image determining section 21b is a circuit section that is equivalent to or simpler than various determination sections (11a, 11d, 11e) included in the control section 11 of the control apparatus 10 described above.
In other words, the image determining section 21b functions as a work determining section configured to determine a work content in image data (still image data or movie data) acquired by the image pickup section 24, based on work information in the database 13. Therefore, if various determination sections (11a, 11d, 11e) are provided in the control section 11 of the control apparatus 10, the image determining section 21b can be omitted.
Here, if the image determining section 21b is provided on the wearable terminal apparatus 20 side, a result of determination by the image determining section 21b alone may be transmitted to the control section 11 of the control apparatus 10 via the communication sections (22, 12). Upon receipt of such determination result, the control section 11 of the control apparatus 10 performs control based on the received determination result information. Therefore, this configuration has an advantage of enabling communication load reduction.
The display control section 21c is a control circuit section for controlling the display section 25 to provide predetermined display on the display panel 25b (see
The communication section 22 is a transmission and reception section including, e.g., a control circuit configured to control communication with (information transmission/reception to/from) the control apparatus 10. For the communication section 22, wireless or wired predetermined communication means is employed.
Furthermore, the communication section 22 can communicate with the database 13 via the control apparatus 10 and refer to work information in the database 13. A configuration in which the communication section 22 and the database 13 directly communicate with each other may be employed.
The operation section 23 is a component section including, e.g., a plurality of operation members for operating the wearable terminal apparatus 20 and a plurality of electric components corresponding to the plurality of operation members, the plurality of electric components being configured to generate respective predetermined operation input signals (detailed configuration of the operation section 23 is not illustrated).
The image pickup section 24 is a component section including, e.g., an image pickup optical system 24a (see
For the image pickup device, for example, a photoelectric conversion element such as a CCD (charge-coupled device) image sensor or a CMOS (complementary metal-oxide semiconductor)-type image sensor is employed. Such an image pickup section as above can provide an image close to an operator's visual line (range of vision or field of vision) during work and thus makes it easy to confirm detailed tasks using both hands (including visual check).
The display section 25 is a component section including, e.g., a display control circuit configured to, in addition to displaying an image based on image data acquired by the image pickup section 24, be capable of displaying various warnings or various predetermined messages or guides (details will be described later), etc., in such a manner that such warnings or predetermined messages or guides, etc., are superimposed on the image, and displays, e.g., a menu for various settings, etc., for the wearable terminal apparatus 20 by appropriately switching to a predetermined operation mode.
The display section 25 includes, e.g., a light guiding section 25a and the display panel 25b. The display panel 25b is a component section including, e.g., a display screen for providing various types of display and a display control circuit for achieving such display. For the display panel 25b, for example, a liquid-crystal display (LCD) section or an organic electroluminescence (organic EL or OEL) display section is employed.
The light guiding section 25a is a component section configured to guide an image displayed on the display panel 25b to a predetermined display surface.
Here, as illustrated in
More specifically, for example, in
Inside the light guiding section 25a, a light refracting member (prism or the like) configured to refract and guide light guided in the arrow X2 direction by the light guiding section 25a in the arrow X1 direction is disposed at an end portion (see reference numeral 25x). Therefore, light emitted from the display surface of the display panel 25b is guided in the arrow X2 direction by the light guiding section 25a and displayed at a position that can be viewed by one of the eyes of the operator (see reference numeral 25c in
The voice input section 27 is a component section including, e.g., an electric component configured to receive an input of voice information and a drive circuit configured to drive the electric component. The voice input section 27 is controlled by the control section 21 so as to collect voice information in a surrounding environment at a predetermined timing. For the voice input section 27, for example, a microphone is employed. More specifically, for example, the voice input section 27 is configured to generate a predetermined instruction signal upon receipt of the operator's voice instruction. In other words, the wearable terminal apparatus 20 is configured to be capable of treating voice inputted from the voice input section 27 as an operation command. Therefore, in this case, the voice input section 27 functions as operation means for performing a predetermined operation.
For an arrangement of the voice input section 27, for example, it is desirable that the voice input section 27 be arranged in the vicinity of the mouth of the operator (see
The voice output section 28 is a component section including, e.g., an electric component configured to output voice information and a drive circuit configured to drive the electric component. The voice output section 28 is controlled by the control section 21 to output voice information (for example, voice information such as a warning or a message, or a voice guide) at a predetermined timing. For the voice output section 28, for example, a sound generating apparatus such as a speaker is employed.
As an arrangement of the voice output section 28, for example, it is desirable that the voice output section 28 be arranged in the vicinity of the operator’ ear (see
The support portion 29 (see
For the voice output section 28, in addition to the function that outputs voice information, an electric component having a function that generates vibration may further be employed. In this case, at the time of an output of voice information, vibration of a predetermined cycle is generated, enabling a warning to the operator to be more clearly transmitted. The cycle of the vibration is changed according to the type of the warning, enabling vibration to be generated according to each of a plurality of types of warnings. A vibration generating section may be provided as an independent component.
Although not illustrated, the wearable terminal apparatus 20 according to the present embodiment includes a battery that serves as a power supply source. The battery generates power necessary for driving the wearable terminal apparatus 20 and supplies the power to the respective component sections under the control of the control section 21.
The wearable terminal apparatus 20 also includes the clock section 20a (see
Other components, illustration and description of which have been omitted, are equivalent to relevant components in a conventional general image pickup apparatus configured to be capable of acquiring image data.
A manner in which an operator wears the wearable terminal apparatus 20 configured as above is, for example, as illustrated in
In
In this case, it is assumed that the operator 100 wears the support portion 29 of the wearable terminal apparatus 20, for example, in the vicinity of the right ear (for the operator 100 himself/herself).
As described above, when an operator wears the wearable terminal apparatus 20, the image pickup optical system 24a of the image pickup section 24 is directed in a direction that substantially corresponds to the direction of the operator's visual line (arrow X1 direction). When the operator 100 wears the wearable terminal apparatus 20 and changes the direction of the visual line by, e.g., turning his/her head, the direction in which the image pickup optical system 24a of the image pickup section 24 faces is also changed so as to follow the change of the visual line. Therefore, image data acquired by the image pickup section 24 when the operator 100 wears the wearable terminal apparatus 20 is image data of a predetermined area including a direction in which the operator looks (direction of the visual line). Here, an image pickup area for image data acquired by the image pickup section 24 is the frame line indicated by reference numeral 24c in
Furthermore, in a state in which the operator 100 wears the wearable terminal apparatus 20, display on the display panel 25b of the display section 25 is provided in the predetermined area indicated by the reference numeral 25c in
Although the display area 25c is illustrated in such a manner that the display area 25c is included within the left eye's field of vision 110L in
Operation of the work support system 1 according to the present embodiment configured as above will be described below.
Although various types of work in various fields can be assumed as work performed using the work support system 1 according to the present embodiment, the below description will be provided as a specific example of work taking, as an example, maintenance work relating to, e.g., cleaning a predetermined medical instrument or the like from among types of work of treating a medical instrument or the like. More specifically, work for cleaning and sterilizing a surgical instrument or a treatment instrument, which is a used medical instrument (for example, an instrument having a shape of scissors; hereinafter abbreviated as “scissors”) (hereinafter referred to as “instrument cleaning/sterilizing work”) is described as an example.
First, a schematic flow of instrument cleaning/sterilizing work generally performed will briefly be described with reference to the flowchart in
As illustrated in
Subsequently, in step S1102, the operator confirms a state of a maintenance environment. Here, a maintenance environment refers to an environment of a site (e.g., a room or a table) for performing the work. Confirmation of a state of an environment refers to confirmation of whether or not the environment is in a clean state, that is, whether or not the environment is suitable for cleaning/sterilizing work. In a normal case, a dedicated maintenance environment (e.g., a room or a table) is provided.
Subsequently, in step S1103, the operator performs a first task in a first process of the maintenance work.
Here, it is assumed that maintenance work is considered as sequential work to be completed by performing a plurality of processes each including a plurality of tasks. In other words, predetermined maintenance work is completed by performing a plurality of work processes in a predetermined procedure. Each of the plurality of work processes further includes a plurality of specific task actions, and each work process is completed by performing a plurality of relevant tasks in a predetermined procedure.
In step S1103, the first task in the first process is, for example, a task of confirming a maintenance tool. Here, a maintenance tool is, for example, a cleaning brush for cleaning instruments. In this case, a plurality of types of cleaning brushes are provided as maintenance tools. It is necessary that tools be provided for respective object parts of predetermined instruments so as to be optimum for the respective maintenance object instruments and that tools optimum for respective processes and respective tasks in maintenance work be selected. Therefore, before each task, it is necessary to confirm a tool to be used.
Here, at a timing of the operator confirming a tool to be used, an image of the tool to be used is automatically acquired as evidence information. Accompanying information such as time information of a task start timing is recorded at the same time. Here, the accompanying information is recorded as accompanying information attached to image data, for example, meta data in an image file format such as Exif (exchangeable image file format). Furthermore, e.g., a task of the operator looking around to acquire image data of a work environment may be included.
Next, in step S1104, the operator performs a second task in the first process. The second task is, for example, a task of cleaning a predetermined first part (specific part, for example, blade portions of scissors) of the maintenance object instrument, but a specific example will be omitted. At this timing, a movie during the task is automatically acquired as evidence information. A still image may be acquired at a predetermined timing in this case too. Simultaneously, accompanying information such as time information (e.g., task start time and completion time or a total task time period) is also recorded. Such accompanying information, etc., may be recorded in such a manner that the pieces of accompanying information, etc., is added to the acquired movie data or the still image data in the same file.
Subsequently, in step S1105, the operator performs a third task in the first process. As in the above, the third task is, for example, a task of cleaning a predetermined second part (specific part, for example, a joint portion of the scissors) that is different from the first part of the cleaning maintenance object instrument, but a specific example of third task will be omitted. Here, evidence information similar to the above (image data and various accompanying information, etc.) are also recorded.
In the next step S1106, whether or not a plurality of tasks prescribed in the first process have ended is confirmed. In the work example illustrated in the flowchart in
Therefore, if it is determined at that point of time that the tasks in the first process have not ended, the work returns to the processing in step S1104 and processing that is similar to the above is repeated. Upon completion of all the tasks in the first process being confirmed, the work proceeds to processing in the next step S1107.
Next, in step S1107, process evidencing processing for each of the tasks in the first process is performed. Here, process evidencing processing is a task of the operator confirming completion of each of the tasks in the first process and clearly indicating the confirmation as evidence. Specific examples of such task include, e.g., a checklist for, for each of task items to be confirmed, the task items being listed in advance for each task, confirming and recording whether or not the operator has performed the task.
For image data and various accompanying data, etc., acquired as evidence information in each of the tasks in the relevant process, processing for organization, such as associating the individual pieces of information with the respective tasks, and recording the resulting pieces of information is performed (see
Next, in step S1108, whether or not all the work processes in the relevant maintenance work (instrument cleaning/sterilizing work) has ended is confirmed. Here, if it is confirmed that all the work processes have not ended, the work proceeds to processing in step S1110. If it is confirmed that all the work processes have ended, the work proceeds to processing in the next step S1109.
In step S1110, the operator performs a task of changing the maintenance tool currently in hand to a maintenance tool to be used for the next work process. Then, the work proceeds to the processing in step S1103 above.
On the other hand, if it is confirmed that all the work processes have ended the operator performs final evidencing processing for the relevant maintenance work (instrument cleaning/sterilizing work) in step S1109. Here, final evidencing processing is, for example, a task of compiling the checklists created for the respective processes in the relevant work and creating a comprehensive and final work checklist and creating an evidence information file of the comprehensive and final work checklist. In other words, the final evidencing processing is processing for compiling the pieces of evidence information for the respective processes, the pieces of evidence information being created in the processing (process evidencing processing) in step S1107 above, and filing the evidence information for the relevant work.
The evidence information file is, for example, as illustrated in
An evidence information file 50 is, for example, a data collection having a list form in which, e.g., information on a maintenance object instrument (index information), a plurality of work processes 52 determined for the maintenance object instrument, and information on contents of a plurality of tasks performed in each work process 52, and a plurality of confirmation items according to the respective contents of the tasks are listed.
A plurality of confirmation items 51 included in the evidence information file 50 include, for example, as illustrated in
Conventionally, a plurality of corresponding image data are acquired for the image data, etc., recorded as pieces of evidence information at the time of completion of each task in sequential work or, for example, at a plurality of timings such as a timing immediately before start of each task, a predetermined timing during the task and a timing of completion of the task. These image data are conventionally acquired by manual operation performed by an operator. The work support system 1 according to the present embodiment is configured so that various kinds of evidence information including these image data, etc., are automatically recorded.
Furthermore, in this case, the work support system 1 according to the present embodiment is configured so that various kinds of evidence information related to image data are automatically recorded, as accompanying information, in a predetermined area in the image data. For example, for an image data file acquired during a task in a predetermined process of predetermined work, accompanying information related to the image data, for example, meta data in an image file format such as Exif (exchangeable image file format) (reference numeral 55 in
Next, operation of use of the work support system 1 according to the present embodiment will be described below.
As illustrated in
Here, illustration of the image pickup apparatus (wearable terminal apparatus 20) that operates in cooperation with the control apparatus 10 is omitted in
The maintenance work (instrument cleaning/sterilizing work) described as an example here is work of cleaning and sterilizing scissors 200, which are a used medical instrument, treatment instrument or the like. Therefore, the workbench 150 includes a cleaning basin 150a. A water faucet 153 is installed above the cleaning basin 150a. When a handle 153a of the water faucet 153 is turned in an opening direction, running water 154 flows out from an outlet 153b.
In
The operator 100 brushes and thus cleans a maintenance object instrument (scissors 200) in the cleaning liquid (for example, water) held in the cleaning basin 150a, using a predetermined tool (cleaning brush 300).
In this case, image data acquired by the wearable terminal apparatus 20 is, for example, the image illustrated in
In the example image display illustrated in
Next, operation when predetermined maintenance work (instrument cleaning/sterilizing work) is performed using the work support system 1 according to the present embodiment will be described below.
As described above, the work support system 1 according to the present embodiment is a system configured to, when an operator 100 performs predetermined work, operate with the image pickup apparatus (wearable terminal apparatus 20) and the control apparatus 10 in cooperation with each other.
Here, an outline of operation of the work support system 1 according to the present embodiment will briefly be described as follows.
In the work support system 1 according to the present embodiment, the image pickup apparatus (wearable terminal apparatus 20) arbitrarily picks up images of states of a plurality of tasks sequentially performed by an operator. The respective image data thus acquired are transmitted to the control apparatus 10.
The control apparatus 10 receives the image data acquired by the wearable terminal apparatus 20 and performs predetermined determination processing based on the image data, etc., and various kinds of information accumulated in advance in the database. Here, the determination processing performed by the control apparatus 10 includes, for example, determinations of a plurality of confirmation items prescribed according to the content of the relevant work, for example, determination of a maintenance object instrument, determination of, e.g., a type of a tool to be used, and determination of, e.g. task actions (e.g., brushing, water cleaning, wiping and chemical spraying).
Then, the control apparatus 10 performs various types of control processing based on results of the determination processing. For example, if a task action performed matches a prescribed task (is a correct task), e.g., a confirmation check table is created and in addition, evidence recording is performed. The evidence recording is, for example, processing for recording evidence information to be included in an evidence information file, which is illustrated in
For example, if a result of the determination is that a task action performed deviates from a prescribed task (is an erroneous task), e.g., a warning to such effect is displayed and e.g., a guide for performing the task correctly again is displayed.
Furthermore, the control apparatus 10 receives image data from the wearable terminal apparatus 20 and performs processing for adding various kinds of corresponding evidence information to the image data, and as necessary, performs various types of data processing such as predetermined falsification prevention processing.
As illustrated in
In the work information database in
“First process: brushing” process;
“Second process: cleaning under running water” process;
“Third process: wiping” process; and
“Fourth process: chemical sterilization” process.
The “First process: brushing” process is a process of brushing in, e.g., a predetermined basin using a predetermined tool (e.g., a brush).
The “Second process: cleaning under running water” process is a process of cleaning under running water.
The “Third process: wiping” process is a process of wiping off water and dirt using a predetermined tool (e.g., paper or cloth).
The “Fourth process: chemical sterilization” process is a process of sterilization by using (spraying) a predetermined chemical (e.g., alcohol).
Here, “scissors cleaning/sterilizing work” is described as an example. Therefore, types and numbers of processes to be performed differ depending on the object instrument.
In the work information database in
Here, examples of tasks included in the “First process: brushing” process in the scissors cleaning/sterilizing work are:
“Tool confirmation” task as “First task”;
“First part brushing” task as “Second task”; and
“Second part brushing” task as “Third task”.
Specific information of each task is recorded in the next work information column “D”.
The above also indicates that each of such processes consists of simple tasks. Consequently, instructions given to an operator are simplified, what motion the operator needs to make for each task and which tool to be used for the task are clarified, the tasks are simplified, and monitoring and confirmation for the tasks can thus be performed by simple processing.
In other words, specific information clearly indicating each task (for example, text data and/or image data) is recorded in the work information database in
As described above, the database 13 includes a plurality of pieces of process information managed chronologically and the operator may perform tasks in such order. As described above, as a result of process information acquisitions in which particular process information is acquired from particular work being sequentially performed, an operator can perform the work in a correct procedure. As a result of each process being broken down, individual simple tasks are provided as work contents, and thus even if an operator has no high skills or is not accustomed to the tasks or the process, the operator can reliably perform the work. The work contents can be called “confirmation work elements” resulting from the work being broken down to a level at which each work content can easily be determined objectively using images (still images and a movie).
Information in work information column [D] further includes movie data information [D1] obtained using the wearable terminal apparatus 20 and still image data information [D2].
The “Tool confirmation” task, which is “First task” included in the “First process: brushing” process is a task of confirming a tool to be used in the relevant work process. For that purpose, in work information [D], movie data in which an action of holding a tool (brush) up to a front face of the wearable terminal apparatus 20 (within an image pickup range) is recorded (for example, data called, e.g., motion data) is recorded as movie data information [D1]. In
Still image data information [D2] includes, e.g., an enlarged image of the tool (brush) and an image of a surrounding environment of the operator 100 (stated “Image of surroundings” in
Here, when the operator performs the “Tool confirmation” task, the following task is performed. In other words, first, the operator 100 performs an action of taking a predetermined tool (cleaning brush 300) to be used for the work to be performed from that time in his/her hand and holding the predetermined tool up to the front face of the wearable terminal apparatus 20. This action is recorded as movie data and recorded at the same time as still image data by the wearable terminal apparatus 20. For the “Tool confirmation” task, the operator 100 holds the predetermined tool (cleaning brush 300) in his/her right hand 100R up to the front face of the wearable terminal apparatus 20 (within the image pickup range). At this time, the wearable terminal apparatus 20 continues acquiring movie data from a point of time of start of the work and simultaneously acquires still image data at a predetermined timing (for example, at a timing of holding the tool to be used up to the front face of the wearable terminal apparatus 20 (within the image pickup range)).
Here, although for still image data, an example in which an image of a gesture that is a motion of holding the tool up is picked up is described, shooting may be performed via a particular operation section operation or a command similar to such operation given by the operator's voice using a voice recognition technique. Automatic shooting may be performed via a remote instruction or at a time at which the task should end. The image data thus obtained by the image pickup apparatus is information related to a task based on a piece of the process information chronologically managed in the work database and a result of the task, and thus, the control apparatus for the image pickup apparatus or the system records the image data acquired by the image pickup apparatus in the recording section, in association with meta data including the piece of process information corresponding to the image data, from among the pieces of work information acquired by the control apparatus, and shooting time information, as a trail of the task.
In this case, it is important that respective processes are classified chronologically in the database, and information on a length of time assumed for each process is stored in the form of data in the database, which enables detection of, e.g., a delay or a trouble in the work, a work mistake or a process skip. A result of the detection can also be read from the image data.
Here, a length of time taken for a task in a particular process may be calculated by a difference between a time at which a start of the task is shot and a time at which an end of the task is shot. Since the processes are managed chronologically, what situation has taken place in which process can be traced while visually comparing images related to the respective processes.
As described above, an effect that is different from an effect conventionally obtained by merely incorporating shooting time information into an image file can be obtained. In other words, whether or not respective processes of work have been performed in an assumed procedure can be determined via images. Verification capability of images can also be enhanced.
It should be understood that the “process” can be broken down into further work contents and a tool, a movement, a motion, etc., matching the work contents are determined and results of the determination may be associated with relevant image data, as meta data.
It is possible to determine, by images, whether or not actual task matches an expected work process (task confirmation). That is, the matching can be determined by breaking down a series of motions recorded in the picked-up images for each work process to the extent that the determination by the images is possible. In such a case, information on the results of such task confirmation, that is, desired result or not, for example, can be associated with the picked-up image data as meta data. Further, each task may be supervised, monitored, subjected to determination of whether or not the task is correct or documented using the relevant image by comparing an picked-up image and an image pickup object or a motion (work element) particular to the relevant work process.
An image of the action (action of holding the cleaning brush 300 up to the front face of the wearable terminal apparatus 20) performed by the operator 100 is picked up by the wearable terminal apparatus 20. At this time, the movie data and the still image data acquired by the wearable terminal apparatus 20 are continuously transferred to the control apparatus 10.
The control apparatus 10 performs predetermined image determination processing based on the received movie data. For example, upon detection of the prescribed action (action of holding the brush up), the control apparatus 10 controls the wearable terminal apparatus 20 to perform an operation of picking up still image data. Consequently, an enlarged image (still image) of the tool to be used (cleaning brush 300) is recorded. The still image acquired at this time is, for example, the image illustrated in
Furthermore, subsequently to or simultaneously with the above, the operator 100 performs an action of looking around the surrounding environment (for example, a work room or a workbench) to acquire still image data using the wearable terminal apparatus 20 (not illustrated).
The image data, etc., thus acquired contributes to determination of whether or not the tool to be used (brush) is correct for use in the relevant task and determination of whether or not the performed task action is correct by referring to the image data, etc., accumulated in the database 13. Then, if a result of the determination is that the performed action is correct, the acquired image data, etc., are recorded as evidence information.
In this case, for determination of whether or not the tool to be used (brush) is correct for use in the relevant task, for example, a well-known individual recognition technique is employed. Here, examples of the well-known individual recognition technique include, in addition to methods in which a predetermined tag such as a barcode and an RFID (radio frequency identifier) is attached to an object for identification, e.g., an object pattern recognition technique in which an image of a fine pattern in a surface of an object (object pattern) is recognized based on image data obtained as a result of the object being shot by an image pickup apparatus, the object pattern recognition technique enabling, even in the case of industrial products uniformly manufactured, identifying each of the industrial products. When any of these individual identification techniques is employed, for example, determination is performed based on the still image data of the enlarged image of the tool to be used (brush), which has been obtained as described above, and unique tool information registered in advance in the database.
The “First part brushing” task, which is “Second task” included in the “First process: brushing” process is a task of brushing a first part of the object instrument “scissors”. Here, the first part is “Blade portions”. For work information in this case, as movie data information [D1], data of the number of times of brushing, a length of time of brushing, etc., to be detected based on a task image are included in the database. An enlarged image of the first part of the tool (brush), etc. is included as still image data information [D2].
Likewise, the “Second part brushing” task, which is “Third task” included in the “First process: brushing” process, is a task of brushing a second part of the object instrument “scissors”. Here, the second part is a “Joint portion”. Work information in this case is substantially similar to the work information for “Second task” above. In other words, as movie data information [D1], data of the number of times, a length of time, etc., of brushing is included in the database, and an enlarged image of the second part of the tool (brush), etc. is included as still image data information [D2].
In the “Second process: cleaning under running water” process, e.g., a water faucet is confirmed as a tool and, e.g., a flow of water (running water) from the faucet is confirmed. For work information, a prescribed length of time of the task for the object part is recorded.
In the “Third process: wiping” process, confirmation of a prescribed dedicated wiping tool such as paper or cloth for the wiping task is performed by, for example, confirming, e.g., a brand of the wiping tool based on an image of, e.g., a package of the wiping tool. For work information, a prescribed number of times, a length of time, etc., of the task for the object part are recorded.
In the “Fourth process: chemical sterilization” process, e.g., a brand of a chemical to be used for a sterilization task is confirmed based on an image of, e.g., a bottle of a prescribed dedicated bottle, for example, alcohol. For work information, a prescribed number of times, a length of time, etc., of the task for the object part are recorded.
Operation when predetermined maintenance work (instrument cleaning/sterilizing work) is performed using the work support system 1 according to the present embodiment will be described below with reference to the flowcharts in
First, it is assumed that in the work support system 1 according to the present embodiment, the respective apparatuses (the control apparatus 10 and the wearable terminal apparatus 20) are in a powered-on state. In this state, an operator 100 starts predetermined work in the predetermined environment illustrated in
In other words, the control apparatus 10 is in a powered-on state and is waiting for receiving a predetermined instruction signal. At this time, the control apparatus 10 performs setting of desired work to be performed from that time, in advance (step S100 in
The setting of the work in the control apparatus 10 does not necessarily need to be performed on the control apparatus 10 side. For example, as described later, it is possible that the work setting information is transmitted to the control apparatus 10 when an operator performs the setting on the image pickup apparatus (wearable terminal apparatus 20) side, and that the setting is performed on the control apparatus 10 side upon receipt of the work setting information.
In this state, the operator 100 starts the predetermined “scissors cleaning/sterilizing work” at a predetermined work site. First, the operator 100 takes a maintenance object instrument (scissors 200) in the “scissors cleaning/sterilizing work” in his/her hand and holds the maintenance object instrument up to the front face of the wearable terminal apparatus 20 (within the image pickup range).
Then, the control section 21 of the wearable terminal apparatus 20 determines whether or not a predetermined work setting operation has been performed. Here, a work setting operation is an operation for selecting and setting desired work to be performed from that time from among work types accumulated in advance in the database 13. This operation is performed by, for example, a predetermined operation being performed using the operation section 23 or the voice input section 27 of the wearable terminal apparatus 20. More specifically, for example, upon a work setting operation by the operator using the operation section 23 or the voice input section 27, first, display of a work list menu appears on the display section 25. The operator selects desired work in the list display using the operation section 23 or the voice input section 27. When the predetermined work is thus selected and set (step S202 in
In the work setting operation performed here, besides the above-indicated example, work setting may be determined by automatically detecting a predetermined action performed by an operator, for example. More specifically, for example, an operator performs an action of holding a predetermined instrument to be subjected to work to be performed from that time, for example, scissors 200 up to the front face of the wearable terminal apparatus 20 (hereinafter abbreviated as “action of holding the instrument up”). The wearable terminal apparatus 20 determines whether or not such action has been performed (step S202 in
In step S203 in
In step S101 in
In step S102 in
In step S204A in
Image data acquired by the image pickup operation performed here may include, in addition to an enlarged image data of the scissors 200, for example, image data of a surrounding environment (for example, a work room or a workbench). Acquisition of surrounding environment data enables confirmation of, e.g., whether or not the work is performed in a correct (prescribed) work environment.
Subsequently, in step S204B in
The control apparatus 10 receives the result of the image pickup by the wearable terminal apparatus 20 (image data of the scissors 200) (step S102 in
In step S204A in
Referring back to
In step S103 in
Subsequently, in step S104 in
Subsequently, in step S105 in
Here, the operator 100 starts a maintenance tool confirmation task, which is the “First task” of the “First process”. As described above, a task to be performed by the operator 100 in the “First task: tool confirmation” task is an action of taking a proper tool to be used in the relevant task (cleaning brush 300) in his/her hand and holding the tool (cleaning brush 300) up to the front face of the wearable terminal apparatus 20 (within the image pickup range).
Then, in step S205 in
In other words, the processing for detecting the relevant action (action of holding the cleaning brush 300 up to the front face of the wearable terminal apparatus 20), which is performed in the processing in step S205, is processing for detecting a point of time of a start of the first task (tool confirmation) in the first process (brushing). In other words, the processing is processing for detecting a timing of a start of the first process. Then, upon detecting a point of time of a start of the relevant action, the wearable terminal apparatus 20 starts timing at that timing. The timing processing is processing for acquiring evidence information such as information of a time that is a point of time of a start of a task and/or information of a total length of time consumed for the task. The timing processing is performed by one or both of the clock section 16 of the control apparatus 10 and the clock section 20a of the wearable terminal apparatus 20.
In the wearable terminal apparatus 20, “Image data during the task” is automatically acquired after a lapse of a predetermined length of time after the “action of holding the tool up”. At this time, the control apparatus 10 may cause the display section 25 of the wearable terminal apparatus 20 to display a guide to the effect that, for example, “an image during the task will be picked up”.
On the display section 25 of the wearable terminal apparatus 20, a reference time for the task that the operator is performing may be displayed, a length of time passed from the point of time of the start may be displayed or a prescribed length of time for the task may be displayed in a countdown form. The operator can know a reference length of time to be consumed for the relevant task (here, the second task (first part brushing) in the first process (brushing)) by viewing the time display.
Furthermore, at predetermined timing before an end of the prescribed length of time, information of a procedure to be performed before an end of the task, for example, a guide “hold the object instrument (scissors in the present example) up to the front of the image pickup apparatus” is displayed on the display section 25 of the wearable terminal apparatus 20.
When the operator views the display and follows the procedure, the wearable terminal apparatus 20 acquires enlarged image data of the object after the end of the task.
Referring back to
In step S105 in
In step S111 in
Subsequently, in step S112 in
In step S211 in
Subsequently, in the processing in step S212 in
In step S113 in
In the next step S114, if a result of the comparison indicates no problem, that is, it is confirmed that the task performed by the operator 100 (acquired image) is correct (the prescribed task) (OK), the control apparatus 10 proceeds to processing in the next step S116. If a result of the comparison is not OK, the control apparatus 10 proceeds to processing in step S115.
In step S115 in
In step S221 in
In step S222, the control section 21 of the wearable terminal apparatus 20 controls, e.g., the display section 25 and the voice output section 28 to provide warning display, guide display or the like in a predetermined form.
Here, the warning display, the guide display or the like includes e.g., voice information display using the voice output section 28 in addition to visual information display using the display section 25. Subsequently, the wearable terminal apparatus 20 returns to the processing in step S201 above and enters a waiting state while performing the live view image pickup operation.
In step S116, the control apparatus 10 confirms whether or not the process has ended. This confirmation is performed, for example, with reference to the database 13 or according to a predetermined instruction signal from the wearable terminal apparatus 20, for example, a signal of detection of a predetermined action or an instruction signal generated as a result of a predetermined operation using the operation section 23 or the voice input section 27. Here, if it is confirmed that the process has ended, the control apparatus 10 terminates the above-described timing processing and then proceeds to processing in the next step S117. If it is confirmed that the process has not ended, that is, if the process has not ended but has a next task, the control apparatus 10 returns to the processing in step S113 above.
In step S117, the control apparatus 10 performs evidencing processing for the ended process. The evidencing processing is processing for, e.g., recording the acquired image data, etc., in a predetermined form. For example, the evidencing processing is processing for creating a data file of image data and accompanying information including various kinds of evidence information. Subsequently, the control apparatus 10 proceeds to processing in step S121.
Subsequently, in step S121, the control apparatus 10 confirms whether or not all the processes for the work have ended. Here, if an end of all the processes is confirmed, the control apparatus 10 proceeds to processing in the next step S122. Also, if no end of all the processes is confirmed, the control apparatus 10 proceeds to processing in step S123.
Here, if all the processes for the work have not ended, the operator 100 swaps (changes) the maintenance tool in his/her hand (tool used in the previous task) for (to) a maintenance tool to be used in a next task (for example, a second task).
In this case, in step S123 in
If an end of all the processes is confirmed in the processing in step S121 above and the control apparatus 10 proceeds to the processing in step S122, in step S122, the control apparatus 10 performs final evidencing processing. The final evidencing processing is, e.g., processing for creating files of the plurality of image data, etc., acquired as predetermined position files and creating, e.g., a checklist based on an OK signal outputted in the processing in step S114 above. Subsequently, the control apparatus 10 ends all the processing and returns to a waiting state.
As described above, according to the above-described embodiment, the work support system 1 causes the image pickup apparatus (wearable terminal apparatus 20) and the control apparatus 10 to operate in cooperation with each other, enabling providing work support for reliably performing predetermined sequential work including a plurality of tasks according to a prescribed work procedure and work method and thus enabling preventing mistakes in the work.
A predetermined action during work, a used tool and conditions such as a surrounding environment can also be recorded in the form of images and thus evidenced. Therefore, it is possible to ensure after the work that predetermined work was reliably performed as prescribed.
In this case, various kinds of evidence information are associated with image data acquired by the image pickup apparatus (wearable terminal apparatus 20) and recorded as, for example, accompanying information for the image data. Consequently, for example, even after a lapse of a considerable length of time from an end of the work, evidence information of the time of the work can easily be confirmed merely by checking a plurality of image data acquired during the work.
Image data acquired during work is effectively utilized, mistakes in work are prevented based on a plurality of image data acquired and work information accumulated in advance in the database, and if a work mistake occurs, a warning is given to an operator and a correct procedure is guided, enabling support for performing reliable work.
Work mistakes (for example, an omission in a procedure or a mistake in order in the procedure and negligence of a prescribed task) can reliably be prevented, and thus, anyone can smoothly and reliably perform prescribed work and evidence of progress of work can also easily be recorded.
For using the work support system, work for recording evidence information relating to sequential work is automated and support for presenting, e.g., a timing for acquiring evidence information as work guide information to an operator during the work is provided And when image data is recorded as evidence information, accompanying meta data including various kinds of evidence information is recorded in the relevant image data file, enabling more efficient work. Concurrently, the acquired evidence information can more efficiently be managed.
Furthermore, according to the present embodiment, an evidence information system that reliably monitors work, records a result of the work and creates images of a series of motions in predetermined work as evidence can be configured.
Therefore, for example, even when a problem occurs in, e.g., a relevant work object after the work, a person can easily trace a flow of the relevant sequential work while viewing already-acquired images, and abundant information in the images ensures mechanical traceability, enabling ensuring more steady and solid reliability for, e.g., work object products.
The example work (instrument cleaning/sterilizing work) described in the above embodiment has been described in terms of a case where all the work processes from the first to fourth processes are performed at the same site. However, a work method according to the present invention is not limited only to the method of such a case, but it is possible that some of the work processes are performed at a different site instead of the same site, depending on the work, for example.
For example, in the example work (instrument cleaning/sterilizing work) described in the above embodiment, it is possible that the first process and the second process are performed at the workbench 150 including the cleaning basin 150a, and that the third process and the fourth process are performed at another workbench (not illustrated) located at a site that is different from the site of the workbench 150.
This is because the first process and the second process are cleaning work performed in the vicinity of a water facility, but the third process and the fourth process are work desirably performed at a dry site. In such a case, on another workbench (not illustrated), the work of the third process and the fourth process is performed and an auxiliary camera that is equivalent to but separate from the auxiliary camera 30 may be disposed. Such a separate auxiliary camera may be configured to also operate in cooperation with the control apparatus 10 and also operate under the control of the control apparatus 10.
Although the above embodiment describes an example using the wearable terminal apparatus 20 only as an image pickup apparatus, the present invention is not limited to the form of this example, but for example, a configuration in which a plurality of image pickup apparatuses are used and the plurality of image pickup apparatus are systematically controlled by a single control apparatus 10 is possible.
In this case, for example, a mode in which an image pickup apparatus (which may be a normal camera having a general form) having a form that is different from the form of the wearable terminal apparatus 20 is provided and the image pickup apparatus is, for example, mounted on the workbench 150 as an auxiliary camera and installed so as to be directed to an operator is conceivable.
In the case of such a mode, the auxiliary camera can acquire work information that is different from work information acquired by the wearable terminal apparatus 20 (work information corresponding to the operator's visual line).
In other words, the auxiliary camera can acquire information from a point of view that is different from a point of view of an image corresponding to the operator's visual line, or an objective standpoint, and the system thus can obtain more information.
More specifically, for example, the auxiliary camera is arranged at a position at which a work object and the vicinity of the hands of an operator treating the object are at least included in an image pickup area or is arranged in such a manner that an image of the entire operator is included in the image pickup area so that the operator during the work can be confirmed as well.
Consequently, for example, whether or not the operator wears correct (prescribed) equipment, more specifically, for example, a mask, gloves, face protection equipment, etc., can be confirmed and evidence information of such confirmation can be acquired.
As a mode using an auxiliary camera, in addition to the above-described example in which an auxiliary camera is mounted on the workbench 150, various modes of arrangement of an auxiliary camera are conceivable.
For example, it is conceivable that an auxiliary camera is disposed in the liquid in the cleaning basin 150a.
In the case of this modification, an auxiliary camera 30A disposed in a liquid is an image pickup apparatus having a waterproof function. If the auxiliary camera 30A is provided in the liquid in a cleaning basin 150a as described above, work in the liquid can more clearly be recorded.
In the case of such a configuration as above, it is necessary to use a wired cable for communication between the auxiliary camera 30A and a control apparatus 10 or expose a wireless communication antenna on a water surface.
For the form of the wearable terminal apparatus 20, in addition to the forms of the wearable terminal apparatuses 20 in the embodiment and the modification described above, for example, a configuration in which an image pickup function is incorporated integrally in face protection equipment or the like to be worn by an operator without fail when the operator performs work can be employed. A wearable terminal apparatus having such a configuration enables removing an operator's burden of wearing a separate wearable terminal apparatus because the operator has to wear the face protection equipment without fail before work and also prevents the operator from making a mistake of forgetting to wear a wearable terminal apparatus.
Incorporation of functions, which are on the level equivalent to the functions of the control section of the above-described control apparatus 10, into the apparatus (wearable terminal apparatus 20) enables provision of a wearable apparatus including not only an image pickup section configured to acquire visual line image data in order to acquire a piece of work information corresponding to an operator's visual line, but also a database in which pieces of work information including pieces of image information from a plurality of fields of vision, the pieces of image information relating to predetermined work, are accumulated in advance, a communication section configured to acquire an auxiliary image from an auxiliary camera apparatus provided at a position that is different from a position of the operator, and a work determining section configured to search for or determine a piece of work information in the database based on first image data and second image data, and a system in which the database is provided outside the apparatus and the apparatus searches the database may be employed. The same applies to the auxiliary camera.
As described above, according to the present embodiment, an evidence information system that reliably monitors work, records a result of the work and creates an image of evidence can be configured, which causes increased reliability because a person can easily trace the work viewing images (e.g., when a problem occurs in a work object) and abundant information in images ensures mechanical traceability.
Each of the processing sequences described in the above embodiment can be changed in proceeding as long as such a change is contrary to the nature of the processing sequence. Therefore, in each of the above processing sequences, for example, the order of respective processing steps to be performed may be changed, a plurality of processing steps may be performed simultaneously or the order of respective processing steps may be different each time a processing sequence is executed. In other words, even though the operation flows in the claims, the description and the drawings are described using terms such as “first” and “next” for the sake of convenience, the use of such terms does not mean that it is essential to perform the operation flows in the stated order. It should also be understood that the respective steps included in these operation flows may arbitrarily be omitted as long as such omission does not affect the spirit of the invention.
From among the techniques described here, many of the control and functions described mainly with reference to the flowcharts can often be set by software programs, and the above control and functions can be provided by the software programs being read and executed by a computer.
Here, for simplicity of description, the description has been given solely on a case where a control section configured to perform overall control, but each of individual devices may have a control function of such a control section. A wearable apparatus may have such a control function or a wearable apparatus can be considered as a part of a control apparatus. What has been referred to as an auxiliary camera may have the function of the control apparatus and control a wearable apparatus, or a system in which a control apparatus includes an auxiliary camera may be employed. Furthermore, it is possible that these devices report progress of respective steps via communication and that some of the devices jointly function as a control apparatus.
Each of the aforementioned software programs is electronic data, an entirety or a part of the electronic data being stored or recorded in, e.g., a storage medium or a storage section, more specifically, a storage medium, for example, a portable medium such as a flexible disk, a CD-ROM or a non-volatile memory, or a hard disk or a volatile memory, in advance in a product manufacturing process, as a computer program product. These recording mediums may be arranged in a dispersed manner in order to share control among the devices. Aside from the above, the software programs can be distributed or provided before product shipment or via a portable medium or a communication channel. Even after product shipment, a user can make these software programs operate by downloading the software programs via, e.g., a communication network or the Internet and or installing the software programs in a computer from a storage medium, enabling easy provision of an image pickup apparatus according to the present embodiment.
It should be understood that the present invention is not limited to the above-described embodiment but various modifications and applications are possible without departing from the spirit of the invention. Furthermore, the above embodiment includes various phases of the invention, and various aspects of the invention can be extracted by appropriate combination of a plurality of components disclosed. For example, even if several components are deleted from all the components in the above-described embodiment, the configuration with the components deleted can be extracted as an aspect of the invention as long as such configuration can solve the problem to be solved by the invention and provide effects of the invention. Furthermore, components in different embodiments may appropriately be combined. This invention is not limited by any particular embodiment of the invention except that the invention is limited by the accompanying claims.
Number | Date | Country | Kind |
---|---|---|---|
2017-203978 | Oct 2017 | JP | national |