This application claims priority under 35 U.S.C. §119 to patent application no. DE 10 2014 203 749.2, filed on Feb. 28, 2014 in Germany, the disclosure of which is incorporated herein by reference in its entirety.
The present disclosure relates to a method for monitoring at least one interior space of a building, to a corresponding device and to a corresponding computer program product, and to an assistance system for at least one interior space of a building.
Emergency home call systems are technical systems by means of which elderly or disabled persons in particular may place an emergency call to a switchboard. An emergency call may be activated manually by means of a pushbutton or automatically by means of sensors, e.g. fall sensors.
Against this background, the approach presented here presents a method for monitoring at least one interior space of a building, furthermore a device which employs this method and, finally, a corresponding computer program product and, finally, an assistance system for at least one interior space of a building in accordance with the main claims. Advantageous embodiments emerge from the respective dependent claims and the following description.
In accordance with embodiments of the present disclosure, at least one interior space of a building can be monitored in respect of a presence of a situation which is defined as relevant for the monitoring, in particular by analyzing image data. In particular, a camera system or an optical sensor system can be used or provided here for interior space monitoring. Here, use can be made of e.g. an optical sensor in the home or domestic environment, wherein an assistance system or assistance functions can be implemented by automated evaluation of sensor signals, for example per domicile or per room, from one or more sensors.
Advantageously, at least one interior space of a building can be monitored reliably and accurately in accordance with embodiments of the present disclosure. Here, monitoring-relevant situations can be reliably identified and distinguished from one another in this case, in particular by automated optical monitoring. An identification of situations within the meaning of maintaining, or deviating from, normal situations can be improved. Here, in particular, it is also possible to increase robustness of the monitoring such that an occurrence of false alarms can be reduced. Moreover, multifaceted assistance and comfort systems, which are based on reliable situation identification or room monitoring, can be realized.
A method for monitoring at least one interior space of a building is presented, wherein the method includes the following step:
Comparing recorded image data, which represent the at least one interior space, with reference data, which represent a reference situation (defined as normal or abnormal), in order to generate monitoring information dependent on a comparison result, which monitoring information represents a monitoring-relevant situation in the at least one interior space.
Within the comparison step, the monitoring information can be generated in this case if the image data deviate from reference data which represent a reference situation defined as normal. The monitoring information can also be generated in the comparison step if the image data at least partly correspond to reference data which represent a reference situation defined as abnormal. Here, the monitoring-relevant situation can be a situation which, in accordance with a designated monitoring target, is defined as being relevant. The method can include a step of reading the recorded image data and the reference data. The method can also include a step of recording the image data.
Here, infrared image data recorded by an infrared camera can be used as the recorded image data in the comparison step. The infrared camera can be a camera or thermal camera for depicting thermal radiation, as is used for e.g. temperature measuring instruments or night vision instruments. The infrared camera can be embodied to be effective in the far infrared. The infrared image data can represent image data recorded in the far infrared range. Such an embodiment provides the advantage that an infrared image protects the privacy of occupants of the building to significantly greater extent than an image in the visible light range. Moreover, infrared cameras are insensitive to brightness differences and also operate in the darkness. Significantly more information can be extracted from the images of an infrared sensor or an infrared camera by means of image processing than from signals of simple motion detectors, for example information for identifying a person, counting objects or persons, directional information, temperature information etc.
Therefore, compared to some known sensors, such as e.g. motion detectors or cameras recording in the visible light range, what can advantageously be achieved by using infrared image data or by means of an infrared camera is that, for example, encroachments into the privacy of persons are avoided, a sensitivity in relation to brightness changes is removed or reduced, an identification of stationary persons is made possible, etc. In particular, computational complexity during the image processing can be reduced compared with cameras in the visible light range, e.g. using CCD or CMOS technology, and a sensitivity in relation to changing light conditions can be reduced. Infrared cameras can provide meaningful image data even in the case of darkness. Infrared cameras avoid an encroachment into the privacy of the occupants. Moreover, infrared cameras can be available in a cost-effective manner. Compared to conventional motion detectors, which are based on e.g. radar, ultrasound or infrared technology, infrared cameras can provide sensor data with high information content. Furthermore, it is possible to distinguish between different persons or occupants of the building, e.g. husband, wife, care staff, by using infrared image data. Moreover, pets can be reliably identified and such immunity to pets can reduce susceptibility to faults during the monitoring and can increase robustness. Also, the movement direction of persons or objects can be identified when using infrared image data; for example, it is possible to distinguish between entering and leaving the at least one room. Furthermore, a stationary state of an object, for example in the case of an unconscious person, can also be identified when using infrared image data.
In accordance with one embodiment, the image data can be compared with reference data in the comparison step, which reference data represent at least one object pattern, in order to identify at least one object represented by the image data, which represents a person, an animal or an article. Therefore, object identification can be performed using reference data and recorded image data. Here, objects can be identified and distinguished from one another such that robustness and accuracy of the method can be increased. In particular, it is also possible to distinguish between individual persons. Here, the high accuracy is an important quality feature, since a robust classification of situations is possible, in particular also in the case of households with a number of occupants.
Here, first image data can be compared with second image data, which are recorded offset by a time interval in relation to the first image data, in the comparison step in order to determine a position, a movement, a speed and, additionally or alternatively, a behavior of a person identified in the image data. To this end, a relationship or deviation between the first image data and the second image data can be determined in the region of the person identified in the image data. Such an embodiment offers the advantage that, in particular, there can be identification, including speed and movement direction, of getting-up processes and monitoring of activities of the day-to-day life of at least one person.
Therefore, in particular, an improvement of classification of situations, within the meaning of situational awareness, for the earliest possible identification of inactivity or deviation from normal behavior and a reduction of false alarms can be achieved, for example, inter alia, by the identification of persons, improved distinction between entry and exit of a room, and the identification that a domicile is left or that occupants return home again, and as a result of a pet immunity, and by an extension of the monitoring to households with a number of persons. For such an extension, single person models can be learnt, which can then be used separately or in any combination during application. The term situational awareness can be understood to mean automatic evaluation of sensor data for determining a current state of occupants and/or their surroundings.
Moreover, first image data can be compared with second image data, which are recorded offset by a time interval in relation to the first image data, in the comparison step in order to determine the difference between values of at least one pixel represented in the first image data and in the second image data. Here, the monitoring information can be generated dependent on a comparison of the difference with a threshold. Such an embodiment offers the advantage that, in particular, dangerous or potentially dangerous monitoring-relevant situations can be identified reliably and in an uncomplicated manner, and can be distinguished from harmless situations.
Here, the threshold can be related to the time interval, to a maximum value of the at least one pixel, to a minimum value of the at least one pixel, to a value gradient of the at least one pixel, to a mean value of a plurality of pixels and, additionally or alternatively, to a position of the at least one pixel. Such an embodiment offers the advantage that, due to the multiplicity of reference values, to which the threshold can relate, it is also possible to reliably identify a multiplicity of situations.
Furthermore, the method can include a step of generating the reference data using predefined pattern data, compared image data, at least one item of monitoring information and, additionally or alternatively, surroundings data of the at least one interior space. Here, the reference data can be trained in the generation step. By way of example, the surroundings data can relate to weather data or the like. Such an embodiment offers the advantage that reference data, which enable precise identification of monitoring-relevant situations, which is matched to a multiplicity of applications, are available or can be provided.
Moreover, the method can include a step of emitting warning information and, additionally or alternatively, action information for rectifying the monitoring-relevant situation dependent on the monitoring information. Here, the warning information and, additionally or alternatively, the action information can be generated using the monitoring information. The warning information can be embodied to cause an alarm in the case of processing by a suitable device within, or outside of, the building, wherein the alarm may include an automatically performable command, a message, an acoustic alarm signal and, additionally or alternatively, an optical alarm signal or the like. The action information can be embodied to cause output of an acoustic message and, additionally or alternatively, an optical message within, or outside of, the building in the case of processing by a suitable device. Such an embodiment offers the advantage that, in response to identification of a critical monitoring-relevant situation, e.g. a motionless person, an alarm can be triggered automatically or countermeasures can be introduced or requested.
The approach presented here furthermore develops a device which is embodied to perform or implement the steps of one variant of a method presented here in appropriate apparatuses. This embodiment variant of the disclosure in the form of a device can also quickly and efficiently achieve the objective underlying the disclosure.
In the present case, a device can be understood to mean an electrical instrument, which processes sensor signals and, dependent thereon, outputs control signals and/or data signals. The device can include an interface, which can be embodied in terms of hardware and/or software. In the case of a hardware-type embodiment, the interfaces can for example be part of a so-called system ASIC, which contains very different functions of the device. However, it is also possible for the interfaces to be dedicated integrated circuits or to at least partly consist of discrete components. In the case of a software-type embodiment, the interfaces can be software modules which, for example, are present on a microcontroller in addition to other software modules.
Furthermore, an assistance system for at least one interior space of a building is presented, wherein the assistance system includes the following features:
at least one camera, which is arranged in the at least one interior space, wherein the at least one camera is embodied for recording and providing image data; and
an embodiment of the aforementioned device, which is connected to the at least one camera in a data transmission-capable manner.
The assistance system can be an emergency home call system or the like. In the building, at least one camera per interior space can be arranged in at least one interior space. In conjunction with the assistance system, one embodiment of the monitoring device mentioned above can advantageously be employed or used. Here, a multiplicity of monitoring functions can advantageously be carried out by using the at least one camera and the monitoring device, which monitoring functions are conventionally carried out by a plurality of sensors, e.g. fire alarms, gas sensors, motion detectors, cameras etc. Hence, assistance, safety and comfort functions can be realized which, for example, are also known by the term ambient assisted living.
In accordance with one embodiment, the device can be embodied as part of the at least one camera or as a separate device which is separate from the at least one camera. The separate device can be arranged in the building. Such an embodiment offers the advantage that the main monitoring functions can be carried out in the building itself, and so a connection to the outside is not mandatory.
Moreover, the assistance system can include a base station arranged in the building and a server arranged separately from the building. Here, the base station can be connected to the at least one camera in a data transmission-capable manner. The server can be connected to the base station in a data transmission-capable manner. Here, the base station can be arranged in the building. Moreover, the server can be connected in a data transmission-capable manner to at least one further base station in at least one further building. Here, the device can be embodied as part of the base station or as part of the server. Such an embodiment offers the advantage that monitoring or image evaluation can take place centrally for one building or for a plurality of buildings.
A computer program product or computer program comprising program code, which can be stored on a machine-readable medium or storage medium such as a semiconductor storage, a hard disk drive or an optical storage, and which is used for performing, implementing and/or actuating the steps of the method according to one of the embodiments described above, in particular if the program product or the program is carried out on a computer or a device, is also advantageous.
The approach presented here is explained in more detail below in an exemplary manner on the basis of the attached drawings. In detail:
In the following description of expedient exemplary embodiments of the present disclosure, the same or similar reference signs are used for the elements which are depicted in the various figures and have a similar effect, wherein a repeated description of these elements is dispensed with.
In accordance with one exemplary embodiment of the present disclosure, infrared image data recorded by an infrared camera are used as the recorded image data in the comparison step 120.
Here, the generation step 110 can be performed prior to and, additionally or alternatively, after the comparison step 120. Optionally, the generation step 110 and, additionally or alternatively, the emission step 130 can also be bypassed. Therefore, the method 100 may include a sequence of steps which, in accordance with one exemplary embodiment, comprises the generation step 110, the comparison step 120 and the emission step 130, which, in accordance with a further exemplary embodiment, comprises the comparison step 120, the generation step 110 and the emission step 130, which, in accordance with an even further exemplary embodiment, comprises the generation step 110, the comparison step 120, the generation step 110 and the emission step 130, which, in accordance with a further exemplary embodiment, comprises the comparison step 120 and the emission step 130 and which, in accordance with an even further exemplary embodiment, comprises the comparison step 120.
In accordance with one exemplary embodiment, the image data can be compared with reference data in the comparison step 120, which reference data represent at least one object pattern, in order to identify at least one object represented by the image data, which represents a person, an animal and/or an article.
Here, in accordance with one exemplary embodiment, first image data can be compared with second image data, which are recorded offset by a time interval in relation to the first image data, in the comparison step 120 in order to determine a position, a movement, a speed and, additionally or alternatively, a behavior of a person identified in the image data.
Moreover, in accordance with one exemplary embodiment, first image data can be compared with second image data, which are recorded offset by a time interval in relation to the first image data, in the comparison step 120 in order to determine the difference between values of at least one pixel represented in the first image data and in the second image data. Here, in the comparison step 120, the monitoring information can be generated dependent on a comparison of the difference with a threshold. The threshold can be related to the time interval, to a maximum value of the at least one pixel, to a minimum value of the at least one pixel, to a value gradient of the at least one pixel, to a mean value of a plurality of pixels and, additionally or alternatively, to a position of the at least one pixel.
In accordance with the exemplary embodiment of the present disclosure depicted in
By way of example, the camera 210 is embodied as an infrared camera. Here, the camera 210 is arranged in an interior space or room (which has not been shown merely for representation-dependent reasons) of the building 250. The camera 210 is embodied for recording and providing image data. Moreover, the camera 210 is connected to the device 220 in a data transmission-capable manner, for example by means of a communication interface in the form of a wire, a wireless connection or the like. The device 220 is connected to the camera 210 and the base station 230 in a data transmission-capable manner, for example by means of communication interfaces in the form of wires, wireless connections or the like. Here, the base station 230 is connected to the device 220 and the server 240 in a data transmission-capable manner, for example by means of communication interfaces in the form of wires, wireless connections or the like. Even though this is not explicitly shown in
In accordance with the exemplary embodiment of the present disclosure depicted in
A first one of the cameras 210 is arranged in a first interior space 301 and embodied for recording and providing image data which represent or depict the first interior space 301. A second one of the cameras 210 is arranged in a second interior space 302 and embodied for recording and providing image data which represent or depict the second interior space 302. A third one of the cameras 210 is arranged in a third interior space 303 and embodied for recording and providing image data which represent or depict the third interior space 301. A fourth one of the cameras 210 is arranged in a fourth interior space 304 and embodied for recording and providing image data which represent or depict the fourth interior space 304. By way of example, the first interior space 301 is a hall, the second interior space 302 is a bedroom, the third interior space 303 is a bathroom and the fourth interior space 304 is a kitchen diner or a living room with an open plan kitchen.
In accordance with the exemplary embodiment of the present disclosure depicted in
In accordance with one exemplary embodiment, the cameras 210 are embodied as infrared cameras. Expressed differently, the cameras 210 or the optical sensors are based on infrared camera modules. Here, the cameras 210 or infrared cameras are preferably sensitive in the far infrared range. Far infrared sensors detect, in particular, inherent thermal radiation from persons and objects, i.e. a received signal or recorded image data is/are dependent on the temperature of an emitting surface.
Expressed differently,
The process of object identification, exemplified in the flowchart 500, includes a block 501, in which image recording takes place. Subsequently, image preprocessing or image segmentation takes place in a block 502. In a branching block 503, to which an optional entry 504 into the process leads, a determination is performed as to whether there is a large change in image data statistics. If there is a large change in the image data statistics, block 503 is followed by block 505, in which there is further processing dependent on the application case or dependent on the use case. If there is no large change in the image data statistics, block 503 is followed by block 506, in which a number N of objects is determined in the image data.
This is followed by block 507, in which an index i is set to 1 (i:=1). In a subsequent branching block 508, a check is carried out as to whether the index i is less than or equal to the number N of objects (i≦N). If this condition is not satisfied, the process returns to the image recording in block 501. If the condition is satisfied, block 508 is followed by block 509, in which a classification and, optionally, object tracking is performed for the object with the index i.
Subsequently, this is followed by a branching block 510, in which a decision is made as to whether the object is an animal. If the object is an animal, the process continues from block 510 to block 511, in which there is special further processing for the animal case. If the object is not an animal, block 510 is followed by a branching block 512, in which a decision is made as to whether the object is a human. If the object is a human, the process continues from block 512 to block 513, in which there is special further processing for the human case. Block 513 can lead to a process shown in
In a much simplified implementation, it would likewise be possible to emit blocks 506 to 514 and to return directly to block 501 after block 503, i.e. only evaluate the image data statistics with a corresponding reaction.
As a result of correction of the return position, the sequence of 506 and 507 is irrelevant. Therefore, it would also be possible to combine 506 and 507 into one block.
The process of classifying situations depicted in the flowchart 600 can include the block of special further processing for the human case, shown in
In block 601 there is a classification of a body position. In a subsequent branching block 602, a check is carried out as to whether a person is lying. If, in block 602, a determination shows that the person is lying, a branching block 603 follows, in which a check is carried out as to whether the person is lying at an untypical spot. If the person is lying at an untypical spot, the process continues from block 603 to block 604, in which an alarm is triggered. If the person is not lying at an untypical spot, the process continues from branching block 603 to branching block 605, where a determination is carried out as to whether there is a large decrease in the body temperature of the person. If this is answered in the affirmative, an alarm is triggered in block 606. However, if there is no strong decrease in the body temperature, branching block 605 is followed in the process by branching block 607, in which a check is carried out as to whether the person is e.g. upright in bed and/or whether said person's feet are on the floor. If this is answered in the affirmative, there is further processing—dependent on application—in block 608 and e.g. an alarm is triggered or a light is switched on. If the check in branching block 607 leads to a negative result, the cyclical process is continued in block 609. The sequence of blocks 603, 605 and 607 is only exemplary and can be varied as desired.
If, in branching block 602, a determination shows that the person is not lying, the process continues at branching block 610, in which a check is carried out as to whether a person enters the room. If this is the case, there is further processing—dependent on application—in block 611. If no person enters the room, branching block 610 is followed by branching block 612, in which a determination is carried out as to whether the person leaves the room or the building or the domicile. The sequence of blocks 610, 612 and 616 is only exemplary and can be varied as desired.
If the determination in block 612 is answered in the affirmative, there is establishment in branching block 613 as to whether a temperature distribution is conspicuous and/or whether a piece of clothing is inadequate. If this is the case, an alarm for a control room or a warning to a person is triggered in block 614. If the temperature distribution is inconspicuous and/or if the clothing is adequate, there is further processing—dependent on application—in block 615.
If the determination in block 612 is answered in the negative, a check is carried out in branching block 616 as to whether a temperature distribution is conspicuous and/or whether the clothing is wet. If this is the case, an alarm for a control room or a warning to a person is triggered in block 617. If the temperature distribution is inconspicuous and/or if the clothing is not wet, there is a continuation of the cyclical process in block 618.
With reference to
In accordance with one exemplary embodiment, at least one camera 210 or device 220 is used for implementing an assistance system 200 which reliably identifies the current situation of occupants of a building 250 (“situational awareness”). A process of identifying the occupants by the at least one camera 210, the device 220 or the assistance system 200 is depicted in
In accordance with one exemplary embodiment, at least one camera 210 or device 220 is used for implementing an assistance system 200 which reliably identifies fallen or motionless persons and triggers an alarm. A functionality of identifying such a situation for example follows the process depicted in
In accordance with one exemplary embodiment, at least one camera 210 or device 220 is used for implementing an assistance system 200 for persons who are at great risk of falling, which assistance system can e.g. identify getting-up processes out of a bed, from a chair or a sofa and can then alert care staff. The care staff can then accompany the person to the goal, e.g. bathroom, kitchen, in order to minimize the risk of falling. Illumination can also be switched on on the basis of the monitoring information. The assistance system 200 can be embodied to derive a getting-up process e.g. from a combination of the sitting-up and feet-on-the-floor person-related events, comparable with the process from
In accordance with one exemplary embodiment, at least one camera 210 or device 220 is used for implementing an assistance system 200 which is embodied to monitor activities of daily life. By way of example, this includes time and duration of personal hygiene, e.g. washing hands in the washbasin in the bathroom, taking in hot meals, e.g. eating at the table, using kitchen implements, such as stove, refrigerator, sink, etc., and time and duration of social contacts, e.g. visits, and monitoring of therapies, e.g. regular running exercises in the case of persons with a limp. By means of sensor information or the image data, such activities are quantifiable by the assistance system 200 or the monitoring method 100. Additionally, the assistance system 200 or the method 100 can check a state of the clothing. In particular, an evaporating liquid leads to a local temperature decrease in the region of an outer surface of the clothing. This is detectable to the assistance system 200 as a change in image data. Therefore, the assistance system 200 can be embodied to identify wetness due to spilled liquids during food intake, e.g. drinks, liquid food, or due to incontinence, wherein the process from
In accordance with one exemplary embodiment, at least one camera 210 or device 220 is used for implementing an assistance system 200 which is embodied to identify risks. By way of example, an alarm can be triggered if burning articles, such as candles, cigarettes, etc., or active electric appliances, e.g. iron, stove, television, are identified and there was simultaneous detection that no person is situated in the domicile or the building 250. An alarm may also be triggered when specific temperature thresholds, e.g. 140° C., are exceeded. Moreover, a warning in relation to hot water, e.g. from a kettle or hot shower water, or in relation to food and drink that is too hot, is possible. Additionally, the at least one camera 210 or device 220 can be used to detect a presence of unauthorized persons in the building 250. An unauthorized person can be identified by recognition/non-recognition or identification by image processing, unusual behavior or panic by the occupant/occupants and/or by an unusual situation, for example if a person enters the building 250 even though no further person is expected at this time and this day of the week.
In accordance with one exemplary embodiment, the assistance system 200 is usable as a mobile assessment system. What this means is that an assistance system 200 comprising a suitable number of cameras 210 and a recording apparatus is set up for a certain period of time with a person in a building. Recorded monitoring information is employable to determine how much, and what type of, help this person is to receive, e.g. for estimating the care stage.
In accordance with one exemplary embodiment, the at least one camera 210 or device 220 of the assistance system 200 is employable as a movement motivator, like in the case of video games. Using this, movements/activity can be fed back interactively for analyzing an activity and for motivating a user.
Considered overall, the described monitoring method 100 or the assistance system 200 inter alia constitutes an essential improvement of the so-called situational awareness. Here, an improved automatic identification of monitoring-relevant situations can be achieved, such as e.g. leaving, and returning to, a domicile. The situational awareness can also be improved for households comprising more than one person. It is possible to identify not only inactivity (situational awareness) but also deviations of activities from a defined or learned normal case, such as e.g. slower, quicker, other running routes, etc. Moreover, additional assistance and comfort functions can be realized in conjunction with the monitoring method 100 or the assistance system 200, such as e.g. a robust fall identification and identification of motionless persons, an identification of getting-up processes, e.g. from a bed, for notifying care staff or for controlling the light to minimize the risk of a fall, e.g. when going to the bathroom, i.e. going to the bathroom with attendants, monitoring of activities of daily living (ADL) such as e.g. personal hygiene, food intake, social interactions, e.g. the number of contacts with other persons, furthermore alarm systems on the basis of the measurement of temperatures, e.g. body temperature, temperature of articles and appliances, e.g. the stove, for a fire alarm or the like, a light or heating control depending on persons present, monitoring of whether, and for how long, windows and doors are open, and an identification of activities such as movement, movement speed, body posture, and therefore also the possibility of providing a warning in the case of deviations from normal behavior/normal situations, or else for transmitting a request to a user to modify a behavior or situation. The monitoring method 100 or the assistance system 200 can also find use for identifying gestures (for example, but not restricted to, in the dark as well), e.g. for triggering an emergency call if the monitoring shows that a person can no longer get up.
The exemplary embodiments, which are described and shown in the figures, are only selected in an exemplary manner. Different exemplary embodiments can be combined with one another, either completely or in relation to individual features. Also, an exemplary embodiment can be complemented by features of a further exemplary embodiment.
Furthermore, the method steps presented here can be carried out repeatedly and in a different sequence to the one described.
If an exemplary embodiment comprises an “and/or” link between a first feature and a second feature, this should be read in such a way that the exemplary embodiment includes both the first feature and the second feature in accordance with one embodiment and includes either only the first feature or only the second feature in accordance with a further embodiment.
Number | Date | Country | Kind |
---|---|---|---|
10 2014 203 749.2 | Feb 2014 | DE | national |