The invention concerns in general the technical field of visual indication. Especially the invention concerns systems for generating visual indication.
Typically, when an elevator call is allocated for a user, the elevator call information, e.g. allocated elevator car and/or destination floor, is indicated for the user by means of, e.g. a display. However, the if the elevator call is allocated already when the user is on the way to the elevator, e.g. when the user accesses via an access control gate device, such as a security gate or turnstile, the user may forget the elevator call information before arriving to the elevator.
Thus, there is need to develop further solutions in order to improve the indication of elevator call information.
The following presents a simplified summary in order to provide basic understanding of some aspects of various invention embodiments. The summary is not an extensive overview of the invention. It is neither intended to identify key or critical elements of the invention nor to delineate the scope of the invention. The following summary merely presents some concepts of the invention in a simplified form as a prelude to a more detailed description of exemplifying embodiments of the invention.
An objective of the invention is to present an indication system and a method for generating an indication. Another objective of the invention is that the indication system and the method for generating an indication enables an on-demand indication of information for a user.
The objectives of the invention are reached by an indication system and a method as defined by the respective independent claims.
According to a first aspect, an indication system for generating an indication is provided, wherein the indication system comprises: at least one indication device, at least one detection device configured to monitor at least one area inside a building to provide monitoring data, and a control unit configured to: detect based on the monitoring data obtained from the at least one detection device at least one predefined gesture of an identified user for which an elevator car has been allocated, and control the at least one indication device to generate a visual indication on a floor of the building in a vicinity of the identified user in response to the detection of the at least one predefined gesture of the identified user.
The operation of the at least one detection device may be based on object recognition or pattern recognition.
The monitored at least one area may comprise a lobby area of the building, a landing area on at least one floor of the building, and/or one or more other areas on at least one floor of the building.
The visual indication may comprise elevator car allocation information and/or destination guidance information.
Alternatively or in addition, the visual indication may be indicated during a predefined period of time or until a detection of a predefined second gesture of the identified user.
The monitoring may comprise tracking movements and gestures of the identified user within the at least one monitoring area.
Moreover, the control unit may further be configured to: detect based on the tracked movements of the identified user that the identified user exits the building, and generate an instruction to an elevator control system to cancel all existing elevator car allocations for said identified user.
Furthermore, the control unit may further be configured to control the at least one indication device to generate a visual indication on a floor of the building in a vicinity of the identified user, wherein the visual indication may comprise an elevator car allocation cancel information.
According to a second aspect, a method for generating an indication is provided, wherein the method comprising: monitoring, by at least one detection device, at least one area inside a building to provide monitoring data; detecting, by a control unit, based on the monitoring data obtained from the at least one detection device at least one predefined gesture of an identified user for which an elevator car has been allocated; and controlling, by the control unit, at least one indication device to generate a visual indication on a floor of the building in a vicinity of the identified user in response to a detection of the at least one predefined gesture of the identified user.
The operation of the at least one detection device may be based on object recognition or pattern recognition.
The monitored at least one area may comprise a lobby area of the building, a landing area on at least one floor of the building, and/or one or more other areas on at least one floor of the building.
The visual indication may comprise elevator car allocation information and/or destination guidance information.
Alternatively or in addition, the visual indication may be indicated during a predefined period of time or until a detection of a predefined second gesture of the identified user.
The monitoring may comprise tracking movements and gestures of the identified user within the at least one monitoring area.
Moreover, the method may further comprise: detecting, by the control unit, based on the tracked movements of the identified user that the identified user exits the building; and generating, by the control unit, an instruction to an elevator control system to cancel all existing elevator car allocations for said identified user.
Furthermore, the method may further comprise controlling, by the control unit, the at least one indication device to generate a visual indication on a floor of the building in a vicinity of the identified user, wherein the visual indication may comprise an elevator car allocation cancel information.
Various exemplifying and non-limiting embodiments of the invention both as to constructions and to methods of operation, together with additional objects and advantages thereof, will be best understood from the following description of specific exemplifying and non-limiting embodiments when read in connection with the accompanying drawings.
The verbs “to comprise” and “to include” are used in this document as open limitations that neither exclude nor require the existence of unrecited features. The features recited in dependent claims are mutually freely combinable unless otherwise explicitly stated. Furthermore, it is to be understood that the use of “a” or “an”, i.e. a singular form, throughout this document does not exclude a plurality.
The embodiments of the invention are illustrated by way of example, and not by way of limitation, in the figures of the accompanying drawings.
The indication system 100 comprises at least one detection device 102 for providing monitoring data, at least one indication device 104, and a control unit 106. The control unit 106 may be external entity or it may be implemented as a part of one or more other entities of the indication system 100. In the example of
The at least one detection device 102 is configured to monitor at least one area inside a building in order to provide the monitoring data. The monitored at least one area may comprise e.g. a lobby area of the building, a landing area on at least one floor of the building, and/or one or more other areas, e.g. corridors or rooms, on at least one floor of the building. In the example of
The at least one detection device 102 may comprise at least one optical imaging device, e.g. at least one camera. The at least one detection device 102 may enable detection, tracking, and/or identification of a user 108 at a distance away from the at least one detection device 102. The distance may be e.g. between 0 to 10 meters from the at least one detection device 102 and preferably between 1 to 2 meters, 1 to 3 meters or 1 to 5 meters. The at least one detection device 102 may be arranged to a wall, a ceiling and/or to a separate support device arranged within the at least one monitored area. In the example of
The at least one detection device 102 is configured to provide the monitoring data to the control unit 106. The control unit 106 is configured to detect at least one predefined gesture of an identified user 108 for which an elevator car A-D has been allocated. The allocation of an elevator car A-D for the user 108 and the identification of the user 108 may be provided by any known methods. Preferably, the allocation of the elevator car for the user 108 and the identification of the user 108 is provided already, when the user 108 is on the way to the elevator A-D, e.g. when the user 108 accesses the building or when the user passed through an access control gate device, such as a security gate. The access control gate devices allow access of identified authorized users through the access control gate device. The access control may be based on using keycards; tags; identification codes; e.g. PIN code, ID number, barcodes, QR codes, etc.; and/or biometric technologies, e.g. fingerprint, facial recognition, iris recognition, retinal scan, voice recognition, etc. The access control gate device may be communicatively coupled to the elevator control system enabling the elevator car allocation for the identified user 108 in response to the identification of an authorized user 108. The control unit 106 of the indication system 100 may obtain the elevator car allocation information and destination guidance information from the access control gate device, the elevator control system, and/or a database to which the elevator car allocation information and destination guidance information are stored.
The detection of the at least one predefined gesture of the identified user is based on the monitoring data obtained from the at least one detection device 102. The control unit 106 may utilize machine vision in the detection of the at least one predefined gesture. The predefined gestures of the identified user 108 may e.g. comprise, but is not limited to, lower a look in front of feet, a wave of hand, a toss of head, or any other gesture of the user 108.
The control unit 106 is configured to control the at least one indication device 104 to generate a visual indication 110 on a floor of the building in a vicinity of, i.e. close to, the identified user in response to the detection of the at least one predefined gesture of the identified user 108. For example, the visual indication 110 may be generated on the floor in front of the feet of the user 108 as shown in the example
The at least one indication device 104 may comprise one or more projector devices configured to project the generated visual indication 110 on the floor in a vicinity of the identified user 108. In the example of
The visual indication 110 may be indicated during a predefined period of time. The predefined period of time may be such that the identified user 108 has time to see the visual indication, for example, but not limited to, the predefined period of time may be between 5 to 10 seconds. Alternatively, the visual indication 110 may be indicated until a detection of a predefined second gesture of the identified user 108. The predefined second gesture may be dependent on the previously detected predefined gesture and/or a counter gesture to the previously detected predefined gesture. According to an example, of the previously detected predefined gesture is lowering the look on the floor in front of his feet, the predefined second gesture of the identified user 108 may be raising the look from the floor. Alternatively, if the previously detected predefined gesture is a wave of hand or a toss of head, the predefined second gesture of the identified user 108 may be a wave of hand or a toss of head into another direction.
According to an example embodiment of the invention, the monitoring may comprise tracking movements and gestures of the identified user 108 within the at least one monitoring area. This enables tracking the movement of the identified user 108 within the monitoring area and every time the identified user 108 performs the predefined gesture, the control unit 106 may be configured to control the at least one indication device 104 to generate the visual indication on the floor in a vicinity of the identified user 108, e.g. in front of the identified user 108, irrespective of the location of the identified user 108 as long as the identified user 108 resides within the monitored area. This enables that the indication of the elevator car allocation information and/or destination guidance information may follow the user 108 to the destination of the user 108.
According to an example embodiment of the invention, the control unit 106 may further be configured to detect if the identified user 108 exits the building based on the tracked movements of the identified user 108. In response to the detection of the exit of the identified user 108, the control unit 106 may be configured to generate an instruction to the elevator control system to cancel all existing elevator car allocations for said identified user 108. This reduces amount of unnecessary elevator car allocations and thus improves the operation of the elevator system 120.
According to an example embodiment of the invention, in response to the cancelling the existing elevator car allocations for said identified user 108, the control unit 106 may further be configured to control the at least one indication device 104 to generate the visual indication 110 on the floor of the building in a vicinity of the identified user 108, e.g. in front if the identified user 108, wherein the generated visual indication comprises an elevator car allocation cancel information.
Next an example of the method according to the invention is described by referring to
At a step 302, the at least one detection device 102 monitors at least one area inside the building in order to provide the monitoring data. The monitored at least one area may comprise e.g. a lobby area of the building, a landing area on at least one floor of the building, and/or one or more other areas, e.g. corridors or rooms, on at least one floor of the building. According to an example embodiment of the invention the operation of the at least one detection device 102 may be based on object recognition or pattern recognition. The at least one detection device 102 provides the monitoring data to the control unit 106.
At a step 304, the control unit 106 detects at least one predefined gesture of an identified user 108 for which an elevator car A-D has been allocated. The allocation of an elevator car A-D for the user 108 and the identification of the user 108 may be provided by any known methods as discussed above. The control unit 106 of the indication system 100 may obtain the elevator car allocation information and destination guidance information from the access control gate device, the elevator control system, and/or a database to which the elevator car allocation information and destination guidance information are stored. The detection of the at least one predefined gesture of the identified user is based on the monitoring data obtained from the at least one detection device 102. The control unit 106 may utilize machine vision in the detection of the at least one predefined gesture. The predefined gestures of the identified user 108 may e.g. comprise, but is not limited to, lower a look in front of feet, a wave of hand, a toss of head, or any other gesture of the user 108.
At a step 306, the control unit 104 controls the at least one indication device 104 to generate a visual indication 110 on a floor of the building in a vicinity of, i.e. close to, the identified user 108 in response to the detection of the at least one predefined gesture of the identified user 108. For example, the visual indication 110 may be generated on the floor in front of the feet of the user 108 as shown in the example
The visual indication 110 may be indicated during a predefined period of time. The predefined period of time may be such that the identified user 108 has time to see the visual indication, for example, but not limited to, the predefined period of time may be between 5 to 10 seconds. Alternatively, the visual indication 110 may be indicated until a detection of a predefined second gesture of the identified user 108. The predefined second gesture may be dependent on the previously detected predefined gesture and/or a counter gesture to the previously detected predefined gesture. According to an example, of the previously detected predefined gesture is lowering the look on the floor in front of his feet, the predefined second gesture of the identified user 108 may be a raising the look from the floor. Alternatively, if the previously detected predefined gesture is a wave of hand or a toss of head, the predefined second gesture of the identified user 108 may be a wave of hand or a toss of head into another direction.
According to an example embodiment of the invention, the monitoring may comprise tracking movements and gestures of the identified user 108 within the at least one monitoring area. This enables tracking the movement of the identified user 108 within the monitoring area and every time the identified user 108 performs the predefined gesture, the control unit 106 may control the at least one indication device 104 to generate the visual indication on the floor in a vicinity of the identified user 108, e.g. in front of the identified user 108, irrespective of the location of the identified user 108 as long as the identified user 108 resides within the monitored area. This enables that the indication of the elevator car allocation information and/or destination guidance information may follow the user 108 to the destination of the user 108.
According to an example embodiment of the invention, the method may further comprise detecting, by the control unit 106, based on the tracked movements of the identified user 108, if the identified user 108 exits the building. In response to the detection of the exit of the identified user 108, the control unit 106 may generate an instruction to the elevator control system to cancel all existing elevator car allocations for said identified user 108. This reduces amount of unnecessary elevator car allocations and thus improves the operation of the elevator system 120.
According to an example embodiment of the invention, in response to the cancelling the existing elevator car allocations for said identified user 108, the method may further comprise controlling, by the control unit 106, the at least one indication device 104 to generate the visual indication 110 on the floor of the building in a vicinity of the identified user 108, e.g. in front if the identified user 108, wherein the generated visual indication comprises an elevator car allocation cancel information.
The specific examples provided in the description given above should not be construed as limiting the applicability and/or the interpretation of the appended claims. Lists and groups of examples provided in the description given above are not exhaustive unless otherwise explicitly stated.
Number | Date | Country | |
---|---|---|---|
Parent | PCT/FI2020/050246 | Apr 2020 | US |
Child | 17941510 | US |