Multi-sensor platform for a building

Information

  • Patent Grant
  • 12038187
  • Patent Number
    12,038,187
  • Date Filed
    Monday, December 20, 2021
    3 years ago
  • Date Issued
    Tuesday, July 16, 2024
    7 months ago
  • CPC
    • F24F11/38
    • F24F11/39
    • F24F11/56
    • F24F11/89
  • Field of Search
    • US
    • 700 276000
    • CPC
    • F24F11/38
    • F24F11/39
    • F24F11/89
    • F24F11/56
  • International Classifications
    • F24F11/38
    • F24F11/39
    • F24F11/56
    • F24F11/89
    • Term Extension
      379
Abstract
A sensing assembly includes a baseboard and a daughterboard operatively coupled to the baseboard. The baseboard includes a microcontroller unit (MCU) mounted to the baseboard, the MCU executing a Real Time Operating System (RTOS) and embedded Artificial Intelligence (AI) code, and two or more sensors that are mounted to the baseboard and operatively coupled to the MCU. The daughterboard includes two or more sensors that are mounted to the daughterboard. The MCU is configured to receive an output signal from each of the two or more sensors mounted to the daughterboard and the two or more sensors mounted to the baseboard and to process two or more of the output signals using the embedded AI code to produce one or more output parameters. The baseboard includes communication circuitry for communicating one or more of the output parameters to a remote device such as a remote server.
Description
TECHNICAL FIELD

The present disclosure pertains generally to sensing assemblies and more particularly to multiple sensor sensing assemblies for a building.


BACKGROUND

Sensors are used to sense a variety of different conditions in a variety of different applications. In building management system (BMS) applications, for example, sensors may be used to sense environmental conditions such as temperature, humidity and various measures of indoor air quality. Alternatively, or in addition, sensors may be used to sense, for example, occupancy of a building space, security events occurring in or around a building space, and/or other abnormal events such as equipment failure occurring in a building space. A need remains for improved sensor assemblies that can be used to support the operation of building management systems.


SUMMARY

This disclosure relates generally to sensing assemblies and more particularly to multiple sensor sensing assemblies. In an example, a multiple sensor sensing assembly provides the opportunity to mix and match particular sensors for particular applications or use-cases, including some universal sensors on a baseboard and additional application specific sensors as desired on one or more daughterboards. In some cases, the multiple sensor sensing assembly may efficiently use artificial intelligence (AI) to locally synthesize or fuse outputs from various ones of the sensors to produce one or more output parameters that can be reported to a remote device such as a remote server.


A particular example may be found in a sensing assembly having a housing with mounting features for mounting the housing to a mounting surface of a building. This example sensing assembly includes a baseboard that is housed by the housing as well as a daughterboard that is housed by the housing and is operably coupled with the baseboard. The baseboard includes a microcontroller unit (MCU) that is mounted to the baseboard and two or more sensors that are mounted to the baseboard. The MCU includes a packaged integrated circuit die that has a central processing unit (CPU) and a non-volatile memory that is operably coupled to the CPU and that stores a Real Time Operating System (RTOS) and an embedded Artificial Intelligence (AI) code for execution by the CPU. The integrated circuit die also includes an I/O port. The two or more sensors mounted to the baseboard are operatively coupled to the I/O port of the MCU and include two or more of a temperature sensor, a humidity sensor, an ambient light sensor, and a microphone.


The daughterboard includes two or more sensors that are mounted to the daughterboard, including an IR sensor and a time of flight (TOF) sensor. The housing defines a window that exposes the IR sensor and the TOF sensor to a space in the building that is external of the housing. The MCU of the baseboard is configured to receive an output signal from each of the two or more sensors mounted to the daughterboard and the two or more sensors mounted to the baseboard, process two or more of the output signals using the embedded AI code to produce one or more output parameters and output the one or more output parameters via the I/O port of the MCU to the baseboard.


Another example may be found in a sensing assembly that includes a baseboard and a daughterboard that is operatively coupled to the baseboard. In this example, the baseboard includes a microcontroller unit (MCU) mounted to the baseboard, the MCU executing a Real Time Operating System (RTOS) and embedded Artificial Intelligence (AI) code, and two or more sensors that are mounted to the baseboard and operatively coupled to the MCU, the two or more sensors including two or more of a temperature sensor, a humidity sensor, an ambient light sensor, and a microphone. The daughterboard includes two or more sensors that are mounted to the daughterboard. The MCU of the baseboard is configured to receive an output signal from each of the two or more sensors mounted to the daughterboard and the two or more sensors mounted to the baseboard and to process two or more of the output signals using the embedded AI code to produce one or more output parameters. The baseboard includes communication circuitry for communicating one or more of the output parameters to a remote device such as a remote server.


Another example may be found in a method of assembling a sensing assembly. In this example, a housing is selected from a first housing and a second housing, wherein the first housing includes a sensor window that is orientated in a first orientation relative to a mounting surface, and the second housing includes a sensor window that is orientated in a second orientation relative to the mounting surface. A baseboard is installed in the selected housing. The baseboard includes a microcontroller unit (MCU) mounted to the baseboard, the MCU including a packaged integrated circuit die, wherein the integrated circuit die includes a central processing unit (CPU), an I/O port and a non-volatile memory operatively coupled to the CPU. The non-volatile memory stores a Real Time Operating System (RTOS) and an embedded Artificial Intelligence (AI) code for execution by the CPU. Two or more sensors are mounted to the baseboard and are operatively coupled to the I/O port of the MCU, the two or more sensors including two or more of a temperature sensor, a humidity sensor, an ambient light sensor, and a microphone.


A daughterboard is installed in the selected housing and operatively coupling the daughterboard to the baseboard, wherein the daughterboard includes two or more sensors mounted to the daughterboard. The first housing, when selected, supports the daughterboard in an orientation where at least one of the two or more sensors mounted to the daughterboard are orientated toward and aligned with the sensor window of the first housing. The second housing, when selected, supports the daughterboard in an orientation where at least one of the two or more sensors mounted to the daughterboard are orientated toward and aligned with the sensor window of the second housing.


The preceding summary is provided to facilitate an understanding of some of the features of the present disclosure and is not intended to be a full description. A full appreciation of the disclosure can be gained by taking the entire specification, claims, drawings, and abstract as a whole.





BRIEF DESCRIPTION OF THE DRAWINGS

The disclosure may be more completely understood in consideration of the following description of various illustrative embodiments of the disclosure in connection with the accompanying drawings, in which:



FIG. 1 is a schematic block diagram of an illustrative sensing assembly;



FIG. 2 is a schematic block diagram of an illustrative sensing assembly;



FIG. 3 is a schematic block diagram of an illustrative microcontroller unit (MCU) usable in the illustrative sensing assembly of FIG. 1 or the illustrative sensing assembly of FIG. 2;



FIG. 4A is a top perspective view of an illustrative housing usable in the illustrative sensing assembly of FIG. 1 or the illustrative sensing assembly of FIG. 2;



FIG. 4B is a bottom perspective view of the illustrative housing shown in FIG. 4A;



FIG. 5 is a flow diagram showing an illustrative method of assembling a sensing assembly such as the illustrative sensing assembly of FIG. 1 or the illustrative sensing assembly of FIG. 2;



FIG. 6 is a flow diagram showing an illustrative method of assembling a sensing assembly such as the illustrative sensing assembly of FIG. 1 or the illustrative sensing assembly of FIG. 2;



FIG. 7 is a schematic diagram showing an illustrative method of occupancy sensing;



FIG. 8 is a schematic diagram showing an illustrative method of balancing energy efficiency and indoor air quality; and



FIG. 9 is a schematic diagram showing an illustrative method of detecting inappropriate behavior.





While the disclosure is amenable to various modifications and alternative forms, specifics thereof have been shown by way of example in the drawings and will be described in detail. It should be understood, however, that the intention is not to limit aspects of the disclosure to the particular illustrative embodiments described. On the contrary, the intention is to cover all modifications, equivalents, and alternatives falling within the spirit and scope of the disclosure.


DESCRIPTION

The following description should be read with reference to the drawings wherein like reference numerals indicate like elements. The drawings, which are not necessarily to scale, are not intended to limit the scope of the disclosure. In some of the figures, elements not believed necessary to an understanding of relationships among illustrated components may have been omitted for clarity.


All numbers are herein assumed to be modified by the term “about”, unless the content clearly dictates otherwise. The recitation of numerical ranges by endpoints includes all numbers subsumed within that range (e.g., 1 to 5 includes 1, 1.5, 2, 2.75, 3, 3.80, 4, and 5).


As used in this specification and the appended claims, the singular forms “a”, “an”, and “the” include the plural referents unless the content clearly dictates otherwise. As used in this specification and the appended claims, the term “or” is generally employed in its sense including “and/or” unless the content clearly dictates otherwise.


It is noted that references in the specification to “an embodiment”, “some embodiments”, “other embodiments”, etc., indicate that the embodiment described may include a particular feature, structure, or characteristic, but every embodiment may not necessarily include the particular feature, structure, or characteristic. Moreover, such phrases are not necessarily referring to the same embodiment. Further, when a particular feature, structure, or characteristic is described in connection with an embodiment, it is contemplated that the feature, structure, or characteristic may be applied to other embodiments whether or not explicitly described unless clearly stated to the contrary.



FIG. 1 is a schematic block diagram of an illustrative sensing assembly 10. The illustrative sensing assembly 10 includes a baseboard 12 and a daughterboard 14 that is operatively coupled with the baseboard 12. While a single daughterboard 14 is shown, it will be appreciated that in some instances there may be two, three or more daughterboards 14 operatively coupled with the baseboard 12. In some instances, the baseboard 12 may be considered as being a motherboard. The baseboard 12 and the daughterboard 14 are both disposed within a housing 16 and are operatively coupled via a flexible ribbon connector 18 that extends between the baseboard 12 and the daughterboard 14. While not shown, the baseboard 12 and the daughterboard 14 may each include a connector port to which the flexible ribbon connector may be electrically and mechanically coupled in order to operatively couple the baseboard 12 and the daughterboard 14.


In the example shown, the baseboard 12 includes an MCU 20, which is a microcontroller unit. The baseboard 12 also includes two or more sensors 22, individually labeled as 22a and 22b, that are operatively coupled with the MCU 20. While two sensors 22 are shown, in some instances the baseboard 12 may include any number of sensors, including one sensor 22, three sensors 22, four sensors 22, or any greater number of sensors 22. At least some of the two or more sensors 22 may include MEMS sensors, for example. The two or more sensors 22 may include any variety of sensors, including but not limited to a temperature sensor, a humidity sensor, an ambient light sensor, an accelerometer, and a microphone. In some cases, the baseboard 12 may additionally or alternatively include sensors such as a vibration sensor or another surface mountable IMU (Inertia Measurement Unit) sensor. An IMU is an electronic device that measures and reports a body's specific force, angular rate, and sometimes the orientation of the body, using a combination of accelerometers, gyroscopes and optionally magnetometers. It will be appreciated that the two or more sensors 22, in combination with the MCU 20, will give the baseboard 12 a certain level of sensing ability and functionality. In some cases, the two or more sensors 22 on the baseboard may be considered to be universal sensors that have application across multiple use cases.


The daughterboard 14 may be considered as providing additional sensing ability and functionality to that provided by the baseboard 12. In some cases, the daughterboard 14 may include application specific sensors that are directed at one or more specific use-cases, but this is not required. In some cases, various different daughterboard 14 each having different sensor combinations may be made available, and a particular one of the different daughterboards may be selected and installed in the sensing assembly 10 depending on the specific use-case at hand.


In the example shown, the daughterboard 14 may include two or more sensors 24, individually labeled as 24a, 24b and 24c. While three sensors 24 are shown, in some instances the daughterboard 14 may include any number of sensors 24, including one sensor 24, two sensors 24, four sensors 24 or any greater number of sensors 24. At least some of the two or more sensors 24 may include MEMS sensors, for example. The two or more sensors 24 may include any variety of sensors including, for example, occupancy sensors and environmental sensors.


Examples of suitable occupancy sensors include but are not limited to IR sensors (such as passive infrared (PIR) sensors) and time of flight (TOF) sensors. Examples of suitable environmental sensors include but are not limited to a temperature sensor, a humidity sensor, a carbon dioxide (CO2) sensor, a carbon monoxide (CO) sensor, a total Volatile Organic Compound (VOC) sensor and a particulate matter (PM) sensor. In some cases, a carbon monoxide (CO) sensor may be used in a standalone daughterboard for use in areas that may be prone to accumulation of hazardous gases such as CO. An underground parking structure is an example of where a CO sensor may be beneficial.


In some instances, a first daughterboard 14 may be equipped as an occupancy card, and thus may include occupancy sensors (e.g. PIR, TOF, micro-LIDAR, mmWave Indoor Radar). A second daughterboard 14 may be equipped as an environmental card, and thus may include environmental sensors. A person assembling the sensing assembly 10 may use the first daughterboard 14 in order to give the sensing assembly 10 occupancy sensing capabilities. Alternatively, or in addition, a person assembling the sensing assembly 10 may use the second daughterboard 14 in order to give the sensing assembly 10 environmental sensing capabilities. If the housing 16 and the baseboard 12 will accommodate more than one daughterboard 14, the person assembling the sensing assembly 10 could include both the first daughterboard 14 and the second daughterboard 14 in order to give the sensing assembly 10 both occupancy sensing capability and environmental sensing capability. These are just examples.


In the example shown, the MCU 20 is configured to receive an output signal from each of the two or more sensors 24 of the daughterboard 14 and to receive an output signal from each of the sensors 22 of the baseboard 12. The MCU 20 is configured to process two or more of the output signals using embedded artificial intelligence (AI) code to produce one or more output parameters. The baseboard 12 includes communication circuitry 26 that is operably coupled with the MCU 20. The MCU 20 is able to communicate one or more of the output parameters to a remote device 28 via the communication circuitry 26. The remote device 28 may be a computer such as a desktop or a laptop, or perhaps a cloud-based server. The remote device 28 may be a mobile device such as a tablet or a smartphone, for example.


In some instances, the one or more output parameters may include one or more occupancy parameters. Examples of occupancy parameters include but are not limited to human presence, people count, people flow and people tracking. The one or more output parameters may include one or more environmental parameters. Examples of environmental parameters include but are not limited to noise, illuminance (light level) and Indoor Air Quality (IAQ). In some instances, the one or more output parameters may include one or more anomaly detection parameters. Examples of anomaly detection parameters may include but are not limited to an anomalous audio, an anomalous pressure, an anomalous temperature, an anomalous humidity, an anomalous ambient light, an anomalous vibration, an anomalous people presence, an anomalous people count, and an anomalous people flow.


In some cases, the MCU 20 is configured to identify one or more events and/or faults based at least in part on the one or more anomaly detection parameters. The embedded AI may be configured to learn to detect anomalies based at least in part on one or more of the processed output signals from the various sensors. In some cases, the embedded AI may be configured to learn over-time what is considered “anomalous” in the particular building to which the sensing assembly 10 is installed. In some cases, anomalous parameters may be communicated to the remote device 28 and not all sensor data, thereby reducing the amount of sensor data that is transmitted to the remote device 28. This can significantly reduce the bandwidth required for the communication channel to the remote device 28.



FIG. 2 is a schematic block diagram of an illustrative sensing assembly 30 that includes the same baseboard 12 as the sensing assembly 10, but includes a daughterboard 32 that has an IR sensor 34 and a TOF sensor 36. The sensing assembly 30 may include components discussed with respect to the sensing assembly 10 and the sensing assembly 10 may include components discussed with respect to the sensing assembly 30. In some cases, the IR sensor 34 may consume a low level of electrical power while the TOF sensor 36 may consume a relatively greater level of electrical power that could prevent the sensing assembly 30 from being battery powered with a reasonable lifetime if the TOF sensor 36 was always enabled, for example. In some cases, the IR sensor 34 may be configured to be powered on all of the time while the TOF sensor 36 is only powered on when the IR sensor 34 indicates movement or the possible presence of a person, for example.


In some cases, the IR sensor 34 may have a field of view (FOV) that overlaps a FOV of the TOF sensor 36. The TOF sensor 36 may have a FOV that is smaller than the FOV of the IR sensor 34, and thus when activated, can be considered to “zoom in” to a particular region of the FOV of the IR sensor 34. In some cases, the TOF sensor 36 has a power savings mode and a sensing mode, wherein after a period of no activity the MCU 20 may set the TOF sensor 36 to the power savings mode, and in response to the IR sensor detecting activity, the MCU 20 may set the TOF sensor 36 to the sensing mode.


In the example shown, the sensing assembly 30 may include a power storage device 38 that may be used to power the sensing assembly 30. The power storage device 38 may be a battery such as a rechargeable battery or a single-charge battery. The power storage device 38 may be a capacitor, for example. In some cases, the sensing assembly 30 may include a power harvesting device 40 that allows the sensing assembly 30 to capture at least some the power necessary to recharge the power storage device 38 and/or operate the sensing assembly 30 from its environment. As an example, the power harvesting device may be or otherwise include a solar cell that can generate electrical power from incident light. The sensing assembly 10 (FIG. 1) may include one or both of the power storage device 38 and the power harvesting device 40. In some cases, the sensing assembly 10 and/or sensing assembly 29 may be powered by line-power, with or without a backup power source. The sensing assembly 30 (FIG. 2) may include the communication circuitry 26. These are just examples.



FIG. 3 is a schematic block diagram showing the MCU 20. In some instances, the MCU 20 is a packaged integrated circuit die that includes a central processing unit (CPU) 50 and a non-volatile memory 52 that is operatively coupled to the CPU 50. The non-volatile memory 52 stores for execution by the CPU 50 a Real Time Operating System (RTOS) 54 and an embedded Artificial Intelligence (AI) code 56. The example MCU 20 also includes an I/O port 58. The I/O port 58 allows the MCU 20 to communicate with the sensors 22 that are mounted on the baseboard 12 as well as with the sensors 24 that are mounted on the daughterboard 14. The I/O port 58 may also allow the MCU 20 to communicate with the remote device 28 sometimes through a communication circuitry 26 on the baseboard 12.


In some instances, the baseboard 12 may be configured to receive an update or replacement to the embedded Artificial Intelligence (AI) code 56 via the communication circuitry 26 and to update or replace the embedded Artificial Intelligence (AI) code 56 that is stored in the non-volatile memory 52 with the updated or replacement embedded Artificial Intelligence (AI) code. The non-volatile memory 52 may further store one or more drivers for communicating with one or more of the sensors 24. The baseboard 12 may be configured to receive an update or replacement of one or more of the drivers via the communication circuitry 26 and update or replace one or more of the drivers stored in the non-volatile memory 52 with one or more of the updated or replacement drivers. The embedded Artificial Intelligence (AI) code 56 and/or the one or more drivers may be updated automatically or by an installer, depending on the sensors that are on the baseboard 12 and/or daughterboard 14 and the particular use-case of the sensing assembly 10 in the field. Such an update may be pushed to the sensing assembly 10 from, for example, a mobile device such as an installer's mobile phone, from a remote server such as a remote cloud server, and/or in any other suitable manner.



FIG. 4A is a top perspective view of an illustrative housing 16, and FIG. 4B is a bottom perspective view thereof. While the housing 16 is shown as being rectilinear, this is merely illustrative. The housing 16 may be circular, for example. In some cases, the housing 16 may be largely rectilinear, but may have rounded over corners and edges, for example. The housing 16 is sized to accommodate both the baseboard 12 and the daughterboard 14, 32 therein. The housing 16 may include one or more windows 60 formed within the housing 16. For example, the housing 16 may include a window 60a that is formed in a side wall 62 of the housing 16. In another example, the housing 16 may include a window 60b that is formed in a front surface 64 of the housing 16. It is contemplated that a particular housing may be selected for the sensing assembly dependent on the expected mounting orientation of the sensing assembly (e.g. on a wall surface, a ceiling surface or a door frame). In another example, the housing 16 may include a window 60a that is formed in a side wall 62 of the housing 16 and another window 60b that is formed in a front surface 64 of the housing 16. The size, shape and location of the window(s) may be dependent on the particular daughterboard(s) that is selected for inclusion in the housing 16. These are just examples.


It is contemplated that the windows 60 may or may not include an aperture or slot that extends through the material forming the housing 16. In some cases, the windows 60 may include a transparent or substantially transparent to whatever ranges of electromagnetic radiation are used by the sensors 24 of the housed baseboard and/or daughterboard. The windows 60 may be formed of a different material than the rest of the housing 16, for example, or may be open slots or apertures that extend through the housing wall to expose the underlying sensors directly to the environment outside of the housing 16.


The illustrative housing 16 also includes mounting features 66 that are disposed relative to a back surface 68 of the housing 16. The back surface 68 may be considered as the surface that is intended to contact a mounting surface of the building. The mounting surface may be a wall surface, a ceiling surface or even a door frame, for example. The mounting features 66 are generically shown, and may include any of a variety of different mounting technologies. In some cases, the mounting features 66 may be tapered holes that are meant to fit over a screw head and then slide down. Other mounting techniques are also contemplated.



FIG. 5 is a flow diagram showing an illustrative method 70 of assembling a sensing assembly (such as the sensing assembly 10 or the sensing assembly 30). A housing is selected from a first housing and a second housing, as indicated at block 72. The first housing includes a sensor window that is orientated in a first orientation relative to a mounting surface, as indicated at block 72a. The second housing includes a sensor window that is orientated in a second direction (different from the first direction) relative to the mounting surface, as indicated at block 72b.


In some cases, the first housing includes two opposing major surfaces (such as the front surface 64 and the back surface 68) with sidewalls (such as the side wall 62) extending between the two opposing major surfaces, and wherein the sensor window of the first housing extends along one of the two opposing major surfaces. The second housing may include two opposing major surfaces (such as the front surface 64 and the back surface 68) with sidewalls (such as the side wall 62) extending between the two opposing major surfaces, and wherein the sensor window of the second housing extends along one of the sidewalls.


A baseboard (such as the baseboard 12) is installed in the selected housing, as indicated at block 74. A daughterboard (such as the daughterboard 14, 32) is installed in the selected housing and is operatively coupled to the baseboard, wherein the daughterboard includes two or more sensors mounted to the daughterboard, as indicated at block 76. The first housing, when selected, supports the daughterboard in an orientation where at least one of the two or more sensors mounted to the daughterboard are orientated toward and aligned with the sensor window of the first housing, as indicated at block 76a. The second housing, when selected, supports the daughterboard in an orientation where at least one of the two or more sensors mounted to the daughterboard are orientated toward and aligned with the sensor window of the second housing, as indicated at block 76b.



FIG. 6 is a flow diagram showing an illustrative method 80 of assembling a sensing assembly (such as the sensing assembly 10 or the sensing assembly 30). A housing is selected from a first housing and a second housing, as indicated at block 82. The first housing and the second housing may be similar to those described with respect to FIG. 6. A baseboard (such as the baseboard 12) is installed in the selected housing, as indicated at block 84. The baseboard includes an MCU (such as the MCU 20) that is a packaged integrated circuit die, as indicated at block 86. The MCU includes a central processing unit (CPU), as indicated at block 86a. The MCU includes a non-volatile memory (such as the memory 52) that is operably coupled to the CPU (such as the CPU 50). The non-volatile memory stores for execution by the CPU a Real Time Operating System and an embedded Artificial Intelligence (AI) code, as indicated at block 86b. The MCU includes an I/O port (such as the I/O port 58), as indicated at block 86c.


The baseboard includes two or more sensors that are mounted to the baseboard and that are operatively coupled to the I/O port of the MCU. The two or more sensors include two or more of a temperature sensor, a humidity sensor, an ambient light sensor and a microphone, as indicated at block 88. A daughterboard is installed in the selected housing and is operatively coupled to the baseboard, the daughterboard including two or more sensors that are mounted to the daughterboard, as indicated at block 90.



FIG. 7 is a schematic diagram showing an illustrative method 100 of occupancy sensing using both an ultra-low power sensor 102 that is always powered on and a low power sensor 104 that is only active when needed. This may be considered as an example of the sensing assembly 10 or the sensing assembly 30. The ultra-low power sensor 102, which may include an infrared (IR) sensor 102a and an ambient light sensor 102b. The low-power sensor 104 may include a TOF sensor such as a micro-LIDAR sensor 104a. In some cases, the micro-LIDAR sensor 104a may include a people counting algorithm 104b. If the IR sensor 102a detects a person, as indicated at block 106, control passes to block 108 and the low power sensor 104 is turned on. If the IR sensor 102a does not detect a person, as indicated at block 110, a switch off timer is started, as indicated at block 112, and then the low power sensor 104 is eventually disabled or otherwise put into a sleep or other low power mode, as indicated at block 114.


If the ambient light sensor 102b detects an illumination level that meets or exceeds a threshold, as indicated at block 116, control passes to block 118 and the low power sensor 104 is activated. However, if the illumination level is below the threshold, control passes to block 112 and the switch off timer is started.



FIG. 8 is a schematic diagram showing an illustrative method 130 of balancing energy efficiency and indoor air quality. This may be considered as an example of the sensing assembly 10 or the sensing assembly 30. An IR sensor 132 detects people, as indicated at block 134. A MEMS microphone 136 detects sounds, and the embedded AI learns what the sound levels should be when a building space is occupied by people and when the building space is not occupied, as indicated at block 138. At block 140, sensor fusion provides for improved occupancy detection accuracy. Control passes both to block 142, which pertains to lighting control, and to block 144, which pertains to dynamic heating, ventilating and air conditioning (HVAC) control. A Micro-LIDAR/TOF sensor 146 counts people within a room or zone, as indicated at block 148. Environmental sensors provide an indication of the indoor environment and perhaps the air quality within the indoor environment, as indicated at block 152. Both block 148 and block 152 pass control to the dynamic HVAC control block 144. Dynamic HVAC control block 144 the controls the HVAC system based on the various inputs.



FIG. 9 is a schematic diagram showing an illustrative method of detecting inappropriate behavior. In this particular example, the inappropriate behavior centers around possible problems within a restroom in a building. A MEMS microphone 162 detects sounds within the space. If the sound exceeds a threshold, a determination may be made that there are more people in the space than are expected, as indicated at block 164. Obviously, the expected crowd within a bathroom is different than that expected in another type of space. The sounds detected by the MEMS microphone 162 are also analyzed for indications of problems, as indicated at block 166. This may include sounds of things being broken, or perhaps sounds of a person in distress (crying, yelling, etc.). At the same time, a Micro-LIDAR/TOF sensor 170 may count how many people are in the space, as indicated at block 172. This can serve as a back-up or confirmation of the crowd estimated using sound processing. The Micro-LIDAR/TOF sensor 170 can also detect a motionless person, which may be an indication of a drug overdose or other health problem. In each case, the information gleaned by processing sound information from the MEMS microphone 162 and the people detection from the Micro-LIDAR/TOF sensor 170 is passed to an alarm management block 168, which issues an alarm when appropriate.


Those skilled in the art will recognize that the present disclosure may be manifested in a variety of forms other than the specific embodiments described and contemplated herein. Accordingly, departure in form and detail may be made without departing from the scope and spirit of the present disclosure as described in the appended claims.

Claims
  • 1. A sensing assembly comprising: a housing with mounting features for mounting the housing to a mounting surface of a building;a baseboard housed by the housing, the baseboard including: a microcontroller unit (MCU) mounted to the baseboard, the MCU including a packaged integrated circuit die, wherein the integrated circuit die includes: a central processing unit (CPU);a non-volatile memory operatively coupled to the CPU, wherein the non-volatile memory stores for execution by the CPU: a Real Time Operating System (RTOS); an embedded Artificial Intelligence (AI) code;an I/O port;two or more sensors mounted to the baseboard and operatively coupled to the I/O port of the MCU, the two or more sensors including two or more of a temperature sensor, a humidity sensor, an ambient light sensor, and a microphone;a daughterboard housed by the housing and operatively coupled to the baseboard, the daughterboard including two or more sensors mounted to the daughterboard including: an IR sensor;a time of flight (TOF) sensor;the housing defining a window that exposes the IR sensor and the TOF sensor of the daughterboard to a space in the building that is external of the housing;the MCU of the baseboard is configured to: receive an output signal from each of the two or more sensors mounted to the daughterboard and the two or more sensors mounted to the baseboard;process two or more of the output signals using the embedded AI code to produce one or more output parameters; andoutput the one or more output parameters via the I/O port of the MCU to the baseboard.
  • 2. The sensing assembly of claim 1, wherein the one or more output parameters comprises one or more occupancy parameters.
  • 3. The sensing assembly of claim 2, wherein the one or more occupancy parameters comprises one or more of human presence, people count, people flow and people tracking.
  • 4. The sensing assembly of claim 1, wherein the one or more output parameters comprises one or more environmental parameters.
  • 5. The sensing assembly of claim 4, wherein the one or more environmental parameters comprises one or more of noise, illuminance and Indoor Air Quality (IAQ).
  • 6. The sensing assembly of claim 1, wherein the one or more output parameters comprises one or more anomaly detection parameters.
  • 7. The sensing assembly of claim 6, wherein the one or more the one or more anomaly detection parameters comprise one or more of an anomalous audio, an anomalous pressure, an anomalous temperature, an anomalous humidity, an anomalous ambient light, an anomalous vibration, an anomalous people presence, an anomalous people count, and an anomalous people flow.
  • 8. The sensing assembly of claim 7, wherein the MCU is configured to identify one or more events and/or faults based at least in part on the one or more anomaly detection parameters.
  • 9. The sensing assembly of claim 1, wherein the field of view (FOV) of the IR sensor overlaps the FOV of the TOF sensor and the FOV of the TOF sensor is smaller than the FOV of the IR sensor, and wherein the TOF sensor has a power savings mode and a sensing mode, wherein after a period of no activity the MCU sets the TOF sensor to the power savings mode, and in response to the IR sensor detecting activity, the MCU sets the TOF sensor to the sensing mode.
  • 10. The sensing assembly of claim 1, wherein the baseboard comprises communication circuitry for communicating one or more of the output parameters to a remote device.
  • 11. The sensing assembly of claim 10, wherein the communication circuitry supports wireless communication.
  • 12. The sensing assembly of claim 10, wherein the baseboard is configured to receive an update or replacement to the embedded Artificial Intelligence (AI) code via the communication circuitry and to update or replace the embedded Artificial Intelligence (AI) code stored in the non-volatile memory of the integrated circuit die of the MCU with the updated or replacement embedded Artificial Intelligence (AI) code.
  • 13. The sensing assembly of claim 10, wherein the non-volatile memory of the integrated circuit die of the MCU further stores one or more drivers for communicating with one or more of the sensors mounted to the daughterboard, and wherein the baseboard is configured to receive an update or replacement of one or more of the drivers via the communication circuitry and update or replace one or more of the drivers stored in the non-volatile memory of the integrated circuit die of the MCU with one or more of the updated or replacement drivers.
  • 14. A sensing assembly comprising: a baseboard including: a microcontroller unit (MCU) mounted to the baseboard, the MCU executing a Real Time Operating System (RTOS) and embedded Artificial Intelligence (AI) code;two or more sensors mounted to the baseboard and operatively coupled to the MCU, the two or more sensors including two or more of a temperature sensor, a humidity sensor, an ambient light sensor, and a microphone;a daughterboard operatively coupled to the baseboard, the daughterboard including two or more sensors mounted to the daughterboard;the MCU of the baseboard is configured to: receive an output signal from each of the two or more sensors mounted to the daughterboard and the two or more sensors mounted to the baseboard;process two or more of the output signals using the embedded AI to produce one or more output parameters; andthe baseboard includes communication circuitry for communicating one or more of the output parameters to a remote device.
  • 15. The sensing assembly of claim 14, wherein the daughterboard includes one or more of an IR sensor, a time of flight (TOF) sensor, a temperature sensor, a humidity sensor, a carbon dioxide (CO2) sensor, a carbon monoxide (CO) sensor, a total VOC sensor and a particulate matter (PM) sensor.
  • 16. The sensing assembly of claim 14, wherein the embedded AI is configured to learn to detect anomalies based at least in part on one or more of the processed output signals.
  • 17. The sensing assembly of claim 14, wherein the sensing assembly further comprises: a power storage device for powering the sensing assembly; anda power harvesting device for harvesting power from the ambient environment to recharge the power storage device.
  • 18. A method of assembling a sensing assembly comprising: selecting a housing from a first housing and a second housing, wherein the first housing includes a sensor window that is orientated in a first orientation relative to a mounting surface, and the second housing includes a sensor window that is orientated in a second orientation relative to the mounting surface;installing a baseboard in the selected housing, wherein the baseboard includes: a microcontroller unit (MCU) mounted to the baseboard, the MCU including a packaged integrated circuit die, wherein the integrated circuit die includes: a central processing unit (CPU);a non-volatile memory operatively coupled to the CPU, wherein the non-volatile memory stores for execution by the CPU: a Real Time Operating System (RTOS); an embedded Artificial Intelligence (AI) code;an I/O port;two or more sensors mounted to the baseboard and operatively coupled to the I/O port of the MCU, the two or more sensors including two or more of a temperature sensor, a humidity sensor, an ambient light sensor, and a microphone;installing a daughterboard in the selected housing and operatively coupling the daughterboard to the baseboard, wherein the daughterboard includes two or more sensors mounted to the daughterboard;the first housing, when selected, supporting the daughterboard in an orientation where at least one of the two or more sensors mounted to the daughterboard are orientated toward and aligned with the sensor window of the first housing; andthe second housing, when selected, supporting the daughterboard in an orientation where at least one of the two or more sensors mounted to the daughterboard are orientated toward and aligned with the sensor window of the second housing.
  • 19. The method of claim 18, wherein the first housing includes two opposing major surfaces with sidewalls extending between the two opposing major surfaces, and wherein the sensor window of the first housing extends along one of the two opposing major surfaces.
  • 20. The method of claim 19, wherein the second housing includes two opposing major surfaces with sidewalls extending between the two opposing major surfaces, and wherein the sensor window of the second housing extends along one of the sidewalls.
PRIORITY DATA

This application is a continuation of International Application No. PCT/CN2021/121137, filed Sep. 28, 2021, the contents of which are herein incorporated by reference in their entirety.

US Referenced Citations (455)
Number Name Date Kind
191512 Bennett et al. Jun 1877 A
4009647 Howorth Mar 1977 A
4375637 Desjardins Mar 1983 A
4918615 Suzuki et al. Apr 1990 A
4939922 Smalley et al. Jul 1990 A
5566084 Cmar Oct 1996 A
5727579 Chardack Mar 1998 A
5745126 Jain et al. Apr 1998 A
5751916 Kon et al. May 1998 A
5777598 Gowda et al. Jul 1998 A
5973662 Singers et al. Oct 1999 A
6065842 Fink May 2000 A
6139177 Venkatraman et al. Oct 2000 A
6144993 Fukunaga et al. Nov 2000 A
6157943 Meyer Dec 2000 A
6229429 Horon May 2001 B1
6238337 Kambhatla et al. May 2001 B1
6250601 Kolar et al. Jun 2001 B1
6334211 Kojima et al. Dec 2001 B1
6353853 Gravlin Mar 2002 B1
6369695 Horon Apr 2002 B2
6375038 Daansen et al. Apr 2002 B1
6429868 Dehner, Jr. et al. Aug 2002 B1
6473084 Phillips et al. Oct 2002 B1
6487457 Hull et al. Nov 2002 B1
6580950 Johnson et al. Jun 2003 B1
6598056 Hull et al. Jul 2003 B1
6619555 Rosen Sep 2003 B2
6704012 Lefave Mar 2004 B1
6712269 Watkins Mar 2004 B1
6720874 Fufido et al. Apr 2004 B2
6741915 Poth May 2004 B2
6796896 Laiti Sep 2004 B2
6801199 Wallman Oct 2004 B1
6816878 Zimmers et al. Nov 2004 B1
6876951 Skidmore et al. Apr 2005 B2
6882278 Winings et al. Apr 2005 B2
6904385 Budike, Jr. Jun 2005 B1
6907387 Reardon Jun 2005 B1
6911177 Deal Jun 2005 B2
6993403 Dadebo et al. Jan 2006 B1
6993417 Osann, Jr. Jan 2006 B2
7023440 Havekost et al. Apr 2006 B1
7031880 Seem et al. Apr 2006 B1
7062722 Carlin et al. Jun 2006 B1
7110843 Pagnano et al. Sep 2006 B2
7139685 Bascle et al. Nov 2006 B2
7164972 Imhof et al. Jan 2007 B2
7183899 Behnke Feb 2007 B2
7200639 Yoshida Apr 2007 B1
7222111 Budike, Jr. May 2007 B1
7222800 Wruck May 2007 B2
7257397 Shamoon et al. Aug 2007 B2
7280030 Monaco Oct 2007 B1
7292908 Borne et al. Nov 2007 B2
7295116 Kumar et al. Nov 2007 B2
7302313 Sharp et al. Nov 2007 B2
7308323 Kruk et al. Dec 2007 B2
7308388 Beverina et al. Dec 2007 B2
7313447 Hsiung et al. Dec 2007 B2
7346433 Budike, Jr. Mar 2008 B2
7356548 Culp et al. Apr 2008 B1
7379782 Cocco May 2008 B1
7383148 Ahmed Jun 2008 B2
7434742 Mueller et al. Oct 2008 B2
7447333 Masticola et al. Nov 2008 B1
7466224 Ward et al. Dec 2008 B2
7496472 Seem Feb 2009 B2
7512450 Ahmed Mar 2009 B2
7516490 Riordan et al. Apr 2009 B2
7548833 Ahmed Jun 2009 B2
7551092 Henry Jun 2009 B1
7552030 Guralnik et al. Jun 2009 B2
7557729 Hubbard et al. Jul 2009 B2
7567844 Thomas et al. Jul 2009 B2
7596473 Hansen et al. Sep 2009 B2
7610910 Ahmed Nov 2009 B2
7626507 LaCasse Dec 2009 B2
7664574 Imhof et al. Feb 2010 B2
7682464 Glenn et al. Mar 2010 B2
7688212 Farley Mar 2010 B2
7702421 Sullivan et al. Apr 2010 B2
7729882 Seem Jun 2010 B2
7755494 Melker et al. Jul 2010 B2
7761310 Rodgers Jul 2010 B2
7774227 Srivastava Aug 2010 B2
7797188 Srivastava Sep 2010 B2
7819136 Eddy Oct 2010 B1
7822806 Frank et al. Oct 2010 B2
7856370 Katta et al. Dec 2010 B2
7944358 Sorensen et al. May 2011 B2
7978083 Melker et al. Jul 2011 B2
7984384 Chaudhri et al. Jul 2011 B2
7986323 Kobayashi et al. Jul 2011 B2
8014970 Gulati et al. Sep 2011 B2
8024666 Thompson Sep 2011 B2
8051156 Sharma et al. Nov 2011 B1
8086047 Penke et al. Dec 2011 B2
8099178 Mairs et al. Jan 2012 B2
8151280 Sather et al. Apr 2012 B2
8176095 Murray et al. May 2012 B2
8218871 Angell et al. Jul 2012 B2
8219660 McCoy et al. Jul 2012 B2
8271941 Zhang et al. Sep 2012 B2
8294585 Barnhill Oct 2012 B2
8302020 Louch et al. Oct 2012 B2
8320634 Deutsch Nov 2012 B2
8334422 Gutsol et al. Dec 2012 B2
8344893 Drammeh Jan 2013 B1
8375118 Hao et al. Feb 2013 B2
8395515 Tokhtuev et al. Mar 2013 B2
8476590 Stratmann et al. Jul 2013 B2
8516016 Park et al. Aug 2013 B2
8558660 Nix et al. Oct 2013 B2
8639527 Rensvold et al. Jan 2014 B2
8698637 Raichman Apr 2014 B2
8718981 Bey et al. May 2014 B2
8816860 Ophardt et al. Aug 2014 B2
8869027 Louch et al. Oct 2014 B2
8904497 Hsieh Dec 2014 B2
8936944 Peltz et al. Jan 2015 B2
8947437 Garr et al. Feb 2015 B2
8950019 Loberger et al. Feb 2015 B2
9000926 Hollock et al. Apr 2015 B2
9030325 Taneff May 2015 B2
9098738 Bilet et al. Aug 2015 B2
9098993 Reed, Jr. Aug 2015 B2
9105071 Fletcher et al. Aug 2015 B2
9175356 Peltz et al. Nov 2015 B2
9240111 Scott et al. Jan 2016 B2
9280884 Schultz et al. Mar 2016 B1
9292972 Hailemariam et al. Mar 2016 B2
9320662 Hayes et al. Apr 2016 B2
9370600 DuPuis et al. Jun 2016 B1
9373242 Conrad et al. Jun 2016 B1
9396638 Wildman et al. Jul 2016 B2
9311807 Schultz et al. Aug 2016 B2
9406212 De Luca et al. Aug 2016 B2
9418535 Felch et al. Aug 2016 B1
9418536 Felch et al. Aug 2016 B1
9449219 Bilet et al. Sep 2016 B2
9477543 Henley et al. Oct 2016 B2
9497832 Verberkt et al. Nov 2016 B2
9513364 Hall et al. Dec 2016 B2
9526380 Hamilton et al. Dec 2016 B2
9526806 Park et al. Dec 2016 B2
9536415 De Luca et al. Jan 2017 B2
9538578 Stamakakis et al. Jan 2017 B1
9558648 Douglas Jan 2017 B2
9591267 Lipton et al. Mar 2017 B2
9613518 Dunn et al. Apr 2017 B2
9618224 Emmons et al. Apr 2017 B2
9618918 O'Keefe Apr 2017 B2
9640059 Hyland May 2017 B2
9672360 Barkan Jun 2017 B2
9685065 Diels Jun 2017 B2
9710700 Bilet et al. Jul 2017 B2
9715242 Pillai et al. Jul 2017 B2
9721452 Felch et al. Aug 2017 B2
9729341 Kaps et al. Aug 2017 B2
9729945 Schultz et al. Aug 2017 B2
9784464 Yamamoto et al. Oct 2017 B2
9843743 Lewis et al. Dec 2017 B2
9856634 Rodenbeck et al. Jan 2018 B2
9872088 Fadell et al. Jan 2018 B2
9875639 Bone et al. Jan 2018 B2
9911312 Wildman et al. Mar 2018 B2
9940819 Ferniany Apr 2018 B2
9956306 Brais et al. May 2018 B2
9986175 Frank et al. May 2018 B2
10087608 Dobizl et al. Oct 2018 B2
10171891 Stamatakis Jan 2019 B1
10223894 Raichman Mar 2019 B2
10228837 Hua et al. Mar 2019 B2
10235865 Thyroff Mar 2019 B2
10251610 Parthasarathy et al. Apr 2019 B2
10298411 Donlan et al. May 2019 B2
10303843 Bitran et al. May 2019 B2
10332382 Thyroff Jun 2019 B2
10382893 Wootton et al. Aug 2019 B1
10469590 Scanlin et al. Nov 2019 B2
10513416 Fang et al. Dec 2019 B2
10514817 Hua et al. Dec 2019 B2
10565844 Pourmohammad et al. Feb 2020 B2
10602474 Goldstein Mar 2020 B2
10607147 Raykov et al. Mar 2020 B2
10609304 Soper et al. Mar 2020 B2
10613504 Chowdhury Apr 2020 B2
10659289 Shif et al. May 2020 B2
10691081 Ray Jun 2020 B2
10705108 Moens et al. Jul 2020 B1
10708154 Pefkianakis et al. Jul 2020 B2
10721088 Saxena et al. Jul 2020 B2
10866003 Ajax Dec 2020 B2
10944830 Scanlin Mar 2021 B2
20020111698 Graziano et al. Aug 2002 A1
20020130868 Smith Sep 2002 A1
20030028269 Spriggs et al. Feb 2003 A1
20030030637 Grinstein et al. Feb 2003 A1
20030046862 Wolf et al. Mar 2003 A1
20030071814 Jou et al. Apr 2003 A1
20030078677 Hull et al. Apr 2003 A1
20030083957 Olefson May 2003 A1
20030103075 Rosselot Jun 2003 A1
20030171851 Brickfield et al. Sep 2003 A1
20030214400 Mizutani et al. Nov 2003 A1
20030233432 Davis et al. Dec 2003 A1
20040001009 Winings et al. Jan 2004 A1
20040030531 Miller et al. Feb 2004 A1
20040064260 Padmanabhan et al. Apr 2004 A1
20040143474 Haeberle et al. Jul 2004 A1
20040153437 Buchan Aug 2004 A1
20040168115 Bauernschmidt et al. Aug 2004 A1
20040233192 Hopper Nov 2004 A1
20040260411 Cannon Dec 2004 A1
20050010460 Mizoguchi et al. Jan 2005 A1
20050119767 Kiwimagi et al. Jun 2005 A1
20050143863 Ruane et al. Jun 2005 A1
20050267900 Ahmed et al. Dec 2005 A1
20060004841 Heikkonen et al. Jan 2006 A1
20060009862 Imhof et al. Jan 2006 A1
20060017547 Buckingham et al. Jan 2006 A1
20060020177 Seo et al. Jan 2006 A1
20060028471 Kincaid et al. Feb 2006 A1
20060029256 Miyoshi et al. Feb 2006 A1
20060058900 Johanson et al. Mar 2006 A1
20060067545 Lewis et al. Mar 2006 A1
20060067546 Lewis et al. Mar 2006 A1
20060077255 Cheng Apr 2006 A1
20060184326 McNally et al. Aug 2006 A1
20060231568 Lynn et al. Oct 2006 A1
20060265664 Simons et al. Nov 2006 A1
20060279630 Aggarwal et al. Dec 2006 A1
20070016955 Goldberg et al. Jan 2007 A1
20070055757 Mairs et al. Mar 2007 A1
20070055760 McCoy et al. Mar 2007 A1
20070061046 Mairs et al. Mar 2007 A1
20070067062 Mairs et al. Mar 2007 A1
20070088534 MacArthur et al. Apr 2007 A1
20070090951 Chan et al. Apr 2007 A1
20070091091 Gardiner et al. Apr 2007 A1
20070101433 Louch et al. May 2007 A1
20070114295 Jenkins May 2007 A1
20070120652 Behnke May 2007 A1
20070139208 Kates Jun 2007 A1
20070216682 Navratil et al. Sep 2007 A1
20070219645 Thomas et al. Sep 2007 A1
20070239484 Arond et al. Oct 2007 A1
20070268122 Kow et al. Nov 2007 A1
20080001763 Raja et al. Jan 2008 A1
20080027587 Nickerson et al. Jan 2008 A1
20080027885 Van Putten et al. Jan 2008 A1
20080036593 Rose-Pehrsson et al. Feb 2008 A1
20080062167 Boggs et al. Mar 2008 A1
20080067244 Marks Mar 2008 A1
20080099045 Glenn et al. May 2008 A1
20080103798 Domenikos et al. May 2008 A1
20080120396 Jayaram et al. May 2008 A1
20080144885 Zucherman et al. Jun 2008 A1
20080183424 Seem Jul 2008 A1
20080194009 Marentis Aug 2008 A1
20080198231 Ozdemir et al. Aug 2008 A1
20080209342 Taylor et al. Aug 2008 A1
20080222565 Taylor et al. Sep 2008 A1
20080224862 Cirker Sep 2008 A1
20080242945 Gugliotti et al. Oct 2008 A1
20080250800 Wetzel Oct 2008 A1
20080279420 Masticola et al. Nov 2008 A1
20080280275 Collopy Nov 2008 A1
20080303658 Melker et al. Dec 2008 A1
20080306985 Murray et al. Dec 2008 A1
20080320552 Kumar et al. Dec 2008 A1
20090001181 Siddaramanna et al. Jan 2009 A1
20090024944 Louch et al. Jan 2009 A1
20090065596 Seem et al. Mar 2009 A1
20090083120 Strichman et al. Mar 2009 A1
20090096791 Abshear et al. Apr 2009 A1
20090125337 Abri May 2009 A1
20090125825 Rye et al. May 2009 A1
20090144023 Seem Jun 2009 A1
20090157744 McConnell Jun 2009 A1
20090160673 Cirker Jun 2009 A1
20090237651 Arndt et al. Sep 2009 A1
20090322782 Kimchi et al. Dec 2009 A1
20100048167 Chow et al. Feb 2010 A1
20100058248 Park Mar 2010 A1
20100064001 Daily Mar 2010 A1
20100070089 Harrod et al. Mar 2010 A1
20100073162 Johnson et al. Mar 2010 A1
20100103019 Ammar Apr 2010 A1
20100123560 Nix et al. May 2010 A1
20100134296 Hwang Jun 2010 A1
20100156628 Ainsbury et al. Jun 2010 A1
20100156630 Ainsbury Jun 2010 A1
20100188228 Hyland Jul 2010 A1
20100223198 Noureldin et al. Sep 2010 A1
20100249955 Sitton Sep 2010 A1
20100286937 Hedley Nov 2010 A1
20100318200 Foslien et al. Dec 2010 A1
20100324962 Nesler et al. Dec 2010 A1
20110010654 Raymond et al. Jan 2011 A1
20110057799 Taneff Mar 2011 A1
20110077779 Fuller et al. Mar 2011 A1
20110083094 Aycock et al. Apr 2011 A1
20110087988 Ray et al. Apr 2011 A1
20110112854 Koch et al. May 2011 A1
20110126111 Gill et al. May 2011 A1
20110154426 Doser et al. Jun 2011 A1
20110161124 Lappinga et al. Jun 2011 A1
20110169646 Raichman Jul 2011 A1
20110184563 Foslien et al. Jul 2011 A1
20110202467 Hilber et al. Aug 2011 A1
20110273298 Snodgrass et al. Nov 2011 A1
20110291841 Hollock et al. Dec 2011 A1
20110295583 Hollock et al. Dec 2011 A1
20110298301 Wong et al. Dec 2011 A1
20110316703 Butler et al. Dec 2011 A1
20110320054 Brzezowski Dec 2011 A1
20120022700 Drees et al. Jan 2012 A1
20120039503 Chen et al. Feb 2012 A1
20120062382 Taneff Mar 2012 A1
20120066168 Fadell Mar 2012 A1
20120075464 Derenne et al. Mar 2012 A1
20120109988 Li et al. May 2012 A1
20120112883 Wallace et al. May 2012 A1
20120112906 Borke et al. May 2012 A1
20120131217 Delorme et al. May 2012 A1
20120158185 El-Mankabady et al. Jun 2012 A1
20120216243 Gill et al. Aug 2012 A1
20120224057 Gill et al. Sep 2012 A1
20120259466 Ray et al. Oct 2012 A1
20120262472 Garr et al. Oct 2012 A1
20120272146 D'souza et al. Oct 2012 A1
20120276517 Banaszuk et al. Nov 2012 A1
20120291068 Khushoo et al. Nov 2012 A1
20120303652 Tseng Nov 2012 A1
20120310418 Harrod et al. Dec 2012 A1
20130007316 Moon et al. Jan 2013 A1
20130055132 Foslien Feb 2013 A1
20130060794 Puttabasappa et al. Mar 2013 A1
20130082842 Balazs et al. Apr 2013 A1
20130086152 Hersche et al. Apr 2013 A1
20130091631 Hayes et al. Apr 2013 A1
20130110295 Zheng et al. May 2013 A1
20130169681 Rasane et al. Jul 2013 A1
20130184880 McMahon Jul 2013 A1
20130187775 Marsden et al. Jul 2013 A1
20130204570 Mendelson et al. Aug 2013 A1
20130229276 Hunter Sep 2013 A1
20130268293 Knudson et al. Oct 2013 A1
20130289774 Day et al. Oct 2013 A1
20140032157 Khiani Jan 2014 A1
20140040998 Hsieh Feb 2014 A1
20140046490 Foslien et al. Feb 2014 A1
20140046722 Rosenbloom et al. Feb 2014 A1
20140058539 Park Feb 2014 A1
20140079282 Marcheselli Mar 2014 A1
20140167917 Wallace et al. Jun 2014 A2
20140207291 Golden Jul 2014 A1
20140266693 Moulton Sep 2014 A1
20140292518 Wildman et al. Oct 2014 A1
20140307076 Deutsch Oct 2014 A1
20140309757 Le Sant et al. Oct 2014 A1
20140316582 Berg-Sonne et al. Oct 2014 A1
20140320289 Raichman Oct 2014 A1
20140342724 Hill et al. Nov 2014 A1
20150022316 Dixon et al. Jan 2015 A1
20150025329 Amarasingham et al. Jan 2015 A1
20150032264 Emmons et al. Jan 2015 A1
20150056909 Chien Feb 2015 A1
20150070174 Douglas Mar 2015 A1
20150077258 Nelson et al. Mar 2015 A1
20150113462 Chen et al. Apr 2015 A1
20150153918 Chen Jun 2015 A1
20150156567 Oliver et al. Jun 2015 A1
20150161874 Thyroff et al. Jun 2015 A1
20150167995 Fadell et al. Jun 2015 A1
20150168949 Hua et al. Jun 2015 A1
20150194043 Dunn et al. Jul 2015 A1
20150198707 Al-Alusi Jul 2015 A1
20150212717 Nair Jul 2015 A1
20150213222 Amarasingham et al. Jul 2015 A1
20150213379 Nair et al. Jul 2015 A1
20150216369 Hamilton et al. Aug 2015 A1
20150253748 Brun et al. Sep 2015 A1
20150281287 Gill et al. Oct 2015 A1
20150310312 Mongeon Oct 2015 A1
20150369505 Malve Dec 2015 A1
20160027278 McIntosh et al. Jan 2016 A1
20160061476 Schultz et al. Mar 2016 A1
20160061477 Schultz et al. Mar 2016 A1
20160061794 Schultz et al. Mar 2016 A1
20160061795 Schultz et al. Mar 2016 A1
20160063833 Schultz et al. Mar 2016 A1
20160066067 Schultz et al. Mar 2016 A1
20160110833 Fix Apr 2016 A1
20160116181 Aultman et al. Apr 2016 A1
20160139067 Grace May 2016 A1
20160178225 Vallikannu Jun 2016 A1
20160253897 Wildman et al. Sep 2016 A1
20160255516 Hill et al. Sep 2016 A1
20160298864 Ekolind et al. Oct 2016 A1
20160306934 Sperry et al. Oct 2016 A1
20160314683 Felch et al. Oct 2016 A1
20160328948 Ferniany Nov 2016 A1
20160335731 Hall Nov 2016 A1
20160367925 Blackley Dec 2016 A1
20160371619 Foster Dec 2016 A1
20170024986 Austin Jan 2017 A1
20170193792 Bermudez Rodriguez et al. Jul 2017 A1
20170256155 Sengstaken, Jr. Sep 2017 A1
20170280949 Wildman et al. Oct 2017 A1
20170294106 Thyroff Oct 2017 A1
20170301220 Jarrell Oct 2017 A1
20170323376 Glaser et al. Nov 2017 A1
20170363721 Yang et al. Dec 2017 A1
20170365024 Koch et al. Dec 2017 A1
20180004178 Haines Jan 2018 A1
20180005382 Wang et al. Jan 2018 A1
20180016773 Chandler et al. Jan 2018 A1
20180053397 Kaur Feb 2018 A1
20180106897 Shouldice et al. Apr 2018 A1
20180113897 Donlan Apr 2018 A1
20180120412 Kim May 2018 A1
20180151054 Pi May 2018 A1
20180218591 Easter Aug 2018 A1
20180292098 Bandyopadhyay et al. Oct 2018 A1
20180293038 Meruva et al. Oct 2018 A1
20180301014 Worral et al. Oct 2018 A1
20180313695 Shim et al. Nov 2018 A1
20180365957 Wright et al. Dec 2018 A1
20190012607 Holliday Jan 2019 A1
20190035250 Hunter et al. Jan 2019 A1
20190049139 Ko Feb 2019 A1
20190051138 Easter Feb 2019 A1
20190139395 Rogachev et al. May 2019 A1
20190209719 Andersen et al. Jul 2019 A1
20190227156 Santra et al. Jul 2019 A1
20190295386 Roberts Sep 2019 A1
20190323897 Wang et al. Oct 2019 A1
20200009280 Kupa et al. Jan 2020 A1
20200074836 Kolavennu et al. Mar 2020 A1
20200090089 Aston et al. Mar 2020 A1
20200146557 Cheung et al. May 2020 A1
20200162503 Shurtleff et al. May 2020 A1
20200177485 Shurtleff et al. Jun 2020 A1
20200200420 Nayak et al. Jun 2020 A1
20200226363 Holliday et al. Jul 2020 A1
20200327315 Mullins Oct 2020 A1
20200364999 Mullins Nov 2020 A1
20210027882 Wellig Jan 2021 A1
20210150380 Verteletskyi et al. May 2021 A1
20210208555 Palanivel et al. Jul 2021 A1
20210396861 Wellig Dec 2021 A1
20220021667 Wellig et al. Jan 2022 A1
Foreign Referenced Citations (47)
Number Date Country
2387100 Nov 2003 CA
2538139 Mar 2005 CA
103110410 May 2013 CN
103970977 Aug 2014 CN
105116848 Dec 2015 CN
108961714 Dec 2018 CN
110009245 Jul 2019 CN
110084928 Aug 2019 CN
110456343 Nov 2019 CN
209785093 Dec 2019 CN
110827457 Feb 2020 CN
1669912 Jun 2006 EP
2310981 Apr 2011 EP
3723456 Oct 2020 EP
369935 Oct 2015 IN
7085166 Mar 1995 JP
11024735 Jan 1999 JP
11317936 Nov 1999 JP
2001356813 Dec 2001 JP
2005242531 Sep 2005 JP
2005311563 Nov 2005 JP
10195017 Dec 2011 KR
1172747 Aug 2012 KR
101445367 Oct 2014 KR
1499081 Mar 2015 KR
9621264 Nov 1996 WO
2004029518 Apr 2004 WO
2005045715 May 2005 WO
2008152433 Dec 2008 WO
2008157755 Dec 2008 WO
2009012319 Jan 2009 WO
2009079648 Jun 2009 WO
2010106474 Sep 2010 WO
2011025085 Mar 2011 WO
2011043732 Apr 2011 WO
2011057173 May 2011 WO
2011123743 Oct 2011 WO
2013062725 May 2013 WO
2013178819 Dec 2013 WO
2014009291 Jan 2014 WO
2014098861 Jun 2014 WO
2014135517 Sep 2014 WO
2016123536 Aug 2016 WO
2017057274 Apr 2017 WO
2018206934 Nov 2018 WO
2019046580 Mar 2019 WO
2020024553 Feb 2020 WO
Non-Patent Literature Citations (144)
Entry
“Energy Manager User Guide,” Release 3.2, Honeywell, 180 pages, 2008.
“Fuzzy Logic Toolbox 2.1, Design and Stimulate Fuzzy Logic Systems,” The MathWorks, 2 pages, May 2004.
“Junk Charts, Recycling Chartjunk as junk art,” 3 pages, Oct. 2, 2006.
“Model Predictive Control Toolbox 2, Develop Internal Model-Based Controllers for Constrained Multivariable Processes,” The MathWorks, 4 pages, Mar. 2005.
Honeywell, “Product Guide 2004,” XP-002472407, 127 pages, 2004.
“Statistics Toolbox, for Use with Matlab,” User's Guide Version2, The MathWorks, 408 pages, Jan. 1999.
“Vykon Energy Suite Student Guide,” Tridium Inc., 307 pages, Mar. 3, 2006.
“Web Based Energy Information Systems for Energy Management and Demand Response in Commercial Buildings,” California Energy Commission, 80 pages, Oct. 2003.
Andover Controls, Network News, vol. 2, No. 2, 8 pages, 1997.
Andover Controls World, 4 pages, Spring 1997.
Bell, Michael B. et al., “Early Event Detection-Results from A Prototype Implementation,” AICHE Spring National Meeting, 15 pages, Apr. 2005.
CADGRAPHICS, “The CADGRAPHICS User's Guide,” 198 pages, 2003.
Carrier Comfort Network CCN Web, “Web Browser User Interface to the Carrier Comfort Network,” 2 pages, 2002.
Carrier Comfort Network CCN Web, Overview and Configuration Manual, 134 pages, Apr. 2006.
Carrier Comfort Network CCN Web, Product Data, 2 pages, Apr. 2006.
Carrier, “i-Vu Powerful and Intuitive Front End for Building Control,” 2 pages, Aug. 2005.
Carrier, “i-Vu Web-Based Integrated Control System,” 3 pages, 2005.
Carrier, Demo Screen Shots, 15 pages, prior to Aug. 27, 2007.
Carrier, i-Vu CCN 4.0, Owner's Guide, 20 pages, Jul. 2007.
Carrier, i-Vu CCN, 7 pages, 2007.
Chen, Tony. F., “Rank Revealing QR Factorizations,” Linear Algebra and It's Applications, vol. 88-89, p. 67-82, Apr. 1987.
Circon, “i-Browse Web-Based Monitoring and Control for Facility Management,” 2 pages, prior to Aug. 27, 2007.
Published Australian Application 2009904740, 28 pages, Application Filed on Sep. 29, 2009.
Echelon, “Energy Control Solutions with the i.Lon SmartServer,” 4 pages, 2007.
Echelon, “i.Lon 100e3 Internet Server Models 72101R-300, 72101R-308, 72102R-300, 72103-R300 . . . ” 5 pages, copyright 2002-2007.
Echelon, “i.Lon 100e3 Internet Server New Features,” 15 pages, Sep. 2006.
Echelon, “i.Lon SmartServer,” 5 pages, 2007.
Honeywell News Release, “Honeywell's New Sysnet Facilities Integration System for Boiler Plant and Combustion Safety Processes,” 4 pages, Dec. 15, 1995.
Honeywell, “Excel Building Supervisor-Integrated R7044 and FS90 Ver. 2.0,” Operator Manual, 70 pages, Apr. 1995.
Honeywell, “Introduction of the S7350A Honeywell WebPAD Information Appliance,” Home and Building Control Bulletin, 2 pages, Aug. 29, 2000; Picture of WebPad Device with touch screen, 1 Page; and screen shots of WebPad Device, 4 pages.
Honeywell, Excel 15B W7760B Building Manager Release 2.02.00, Installation Instructions, 28 pages, Dec. 2004.
Honeywell, The RapidZone Solution, Excel 5000 Open System, Application Guide, 52 pages, Jan. 2004.
http://pueblo.lbl.gov/˜olken . . . , “Remote Building Monitoring and Operations Home Page,” 5 pages, prior to Aug. 27, 2007.
http://www.commercial.carrier.com/commercial/hvac/productdescription . . . , “Carrier: i-Vu CCN,” 1 page, printed Mar. 11, 2008.
http://www.commercial.carrier.com/commercial/hvac/productdescription . . . , “Carrier: 33CSCCNWEB-01 CCN Web Internet Connection to the Carrier Comfort Network,” 1 page, printed Mar. 11, 2008.
http://www.docs.hvacpartners.com/idc/groups/public/documents/techlit/gs-controls-ivuccn.rtf, “Products,” 5 pages, printed Jul. 3, 2007.
http://www.lightstat.com/products/istat.asp, Lightstat Incorporated, “Internet Programmable Communicating Thermostats,” 1 page, printed Mar. 13, 2007.
http://www.sharpsystems.com/products/pc_notebooks/actius/rd/3d/, “Actius RD3D Desktop Replacement Notebook with Industry-Breakthrough 3D Screen,” Sharp, 1 page, printed Jun. 16, 2005.
http://www2.sims.berkeley.edu/courses/is213/s06/projects/lightson;final.html, “Lights On A Wireless Lighting Control System,” 11 pages, printed Mar. 22, 2007.
i.Lon 100e3 Internet Server, 1 page, prior to Aug. 27, 2007.
i.Lon, SmartServer, 2 pages, prior to Aug. 27, 2007.
i-stat, Demo Screen Shots, 9 pages, printed Mar. 13, 2007.
i-stat, The Internet Programmable Thermostat, 2 pages, prior to Aug. 27, 2007.
Jeffrey Ball, “Green Goal of ‘Carbon Neutrality’ Hits Limit,” TheWall Street Journal, 7 pages, Dec. 30, 2008.
Johnson Controls, Network Integration Engine (NIE) 3 pages, Nov. 9, 2007.
Johnson Controls, Network Integration Engine (NIE), Product Bulletin, pp. 1-11, Jan. 30, 2008.
Kourti, Theodora, “Process Analysis and Abnormal Situation Detection: From Theory to Practice,” IEEE Control Systems Magazine, p. 10-25, Oct. 2002.
Mathew, Paul A., “Action-Oriented Benchmarking, Using CEUS Date to Identify and Prioritize Efficiency Opportunities in California Commercial Buildings,” 26 pages, Jun. 2007.
Morrison, Don et al., “The Early Event Detection Toolkit,” Honeywell Process Solutions, 14 pages, Jan. 2006.
Narang, “WEBARC: Control and Monitoring of Building Systems Over the Web,” 53 pages, May 1999.
Iuliana Bocicor et al. “Wireless Sensor Network based System for the Prevention of Hospital Acquired Infections”, Arxiv.org, Cornell University Ithaca, NY 14853, May 2, 2017, XP080947042, (Abstract).
Shhedi Zaid Ali et al., “Traditional and ICT Solutions for Preventing the Hospital Acquired Infection”, 2015 20th International Conference on Control Systems and Computer Science, IEEE, May 27, 2015, pp. 867-873, XP033188038.
Extended European Search Report, EP application No. 20151295.1, pp. 13, May 26, 2020.
U.S. Appl. No. 14/109,496, filed Dec. 17, 2013.
www.geappliances.com/home-energy-manager/about-energy-monitors.htm, “Energy Monitor, Home Energy Monitors, GE Nucleus,” 2 pages, printed Jan. 15, 2013.
www.luciddesigngroup.com/network/apps.php#/homepage, “Lucid Design Group—Building Dashboard Network—Apps, ” 7 pages, Jan. 15, 2013.
Preuveneers et al., “Intelligent Widgets for Intuitive Interaction and Coordination in Smart Home Environments,” IEEE Eighth International Conference on Intelligent Environments, pp. 157-164, 2012.
Wu et al., “A Web 2.0 Based Scientific Application Framework,” 7 pages, prior to Jul. 24, 2014.
“4.0 Today's Activities, The Home Dashboard,” CRBM info@hand website, 46 pages, prior to Apr. 25, 2013.
“Free Facilities Dashboards,” eSight Energy Website, 2 pages, prior to Apr. 25, 2013.
Alerton Building Controls, Gallery Prints, 7 pages, Dec. 19, 2013.
Carter, “Industrial Energy Management Dashboards Require a Toolkit,” Cross Automation, 11 pages, Nov. 4, 2013.
U.S. Appl. No. 14/169,071, filed Jan. 30, 2014.
U.S. Appl. No. 14/169,083, filed Jan. 30, 2014.
U.S. Appl. No. 14/461,188, filed Aug. 15, 2014.
U.S. Appl. No. 14/482,607, filed Sep. 10, 2014.
e-homecontrols.com, “e-Home Controls Website,” link to actual website no longer works, 1 page, prior to Dec. 19, 2013.
http://www.ccbac.com, “C&C (/)—Omniboard,” 5 pages, Dec. 19, 2013.
http://www.domcontroller.com/en/, “DomController Home Automation Software—Control Anything from Anywhere,” 11 pages, printed Jan. 6, 2015.
http://www.novar.com/ems-bas/opus-building-automation-system, “Novar OPUS BAS,” 1 page, prior to Feb. 13, 2013.
Instituto Superior Tecnico, “A 3D Interactive Environment for Automated Building Control,” Master's Dissertation, 120 pages, Nov. 2012.
Panduit Corp., “Enable a Building Automation with Panduit Enterprise Solutions,” 4 pages, Nov. 2012.
“WEBs-AX Web-Enabled Building Solutions,” sales brochure, Honeywell International Inc., Mar. 2009.
“Attune Advisory Services,” press release, Honeywell International Inc., Mar. 20, 2012.
enteliWEB product from Delta Controls, web pages retrieved on May 9, 2013 from http://deltacontrols.com/products/facilities-management/supervisory-software et seq. by the Internet Archive at web.archive.org.
“BACnet Protocol Implementation Conformance Statement” for enteliWEB, Delta Controls, Jul. 17, 2013.
Castle, “7 Software Platforms that Make Building Energy Management Easy,” http://greentechadvocates.com/2012/11/28/7-software-platforms-that-make-building-energy-managment-easy/, Nov. 28, 2012.
enteliWEB catalog sheet, Delta Controls, Inc., 2012.
enteliWEB catalog sheet, Delta Controls., 2010.
“Intelligent Building Management Systems in Miami,” Advanced Control Corp., Mar. 7, 2013.
“The Ohio State University,” BACnet International Journal, vol. 5, p. 4, Jan. 2013.
Bobker et al., “Operational Effectiveness in Use of BAS,” Proceedings of the 13th International Conference for Enhanced Building Operations, Oct. 8, 2013.
Castelo, “A 3D Interactive Environment for Automated Building Control,” Elsevier, Nov. 8, 2012.
“Creston Special Report: How Intelligent building management solutions are reducing operational costs,” Creston, 2012.
“Building Automation Software Solutions,” Iconics, 2013.
Lacey, “The Top 10 Software Vendors Connecting Smart Buildings to the Smart Grid,” http://www.greentechmedia.com/articles/read/the-top-10-companies-in-enterprise-smart-grid, Jul. 18, 2013.
“NiagraAX Product Model Overview,” Tridium, Inc., 2005.
“An Overview of NiagraAX: A comprehensive software platform designed to create smart device applications,” Tridium, Inc., 2005.
“Phoenix Controls Portal,” Phoenix Controls, Inc., 2013.
Quirk, “A Brief History of BIM,” Arch Daily, Dec. 7, 2012.
Samad et al., “Leveraging the Web: A Universal Framework for Building Automation,” Proceedings of the 2007 American Control Conference, Jul. 11, 2007.
Sinha et al., “9 Key attributes of energy dashboards and analytics tools,” https://www.greenbiz.com/blog/2013/08/28/9-key-attributes-energy-dashboards-and=analytics-tools, Aug. 28, 2013.
Sinopoli, “Dashboards For Buildings,” http://www/automatedbuildings.com/news/dec10/articles/sinopoli/101119034404sinopoli.html, Dec. 2010.
Sinopoli, “Modeling Building Automation and Control Systems,” http://www.automatedbuildings.com/news/jun13/articles/sinopoli/130521122303sinopoli.html, Jun. 2013.
Zito, “What is Tridium Part 1,” http://blog.buildingautomationmonthly.com/what-is-tridium/, May 12, 2013.
Zito, “What is Tridium Part 2,” http://blog.buildingautomationmonthly.com/tridium-part-2/, Sep. 10, 2013.
Search Report and Written Opinion from related International PCT Application No. PCT/US2018/025189 dated Jul. 17, 2018 (12 pages).
“Data analytics and smart buildings increase comfort and energy efficiency”, https://www.microsoft.com/itshowcase/Article/Content/845/Data-analytics-and-smart-buildings-increase-comfort-and-energy-efficiency, Dec. 19, 2016, 8 pages.
Donnelly, “Building Energy Management: Using Data as a Tool”, http://www.buildingefficiencyinitiative.org/sites/ default/files/legacy/InstituteBE/media/Library/Resources/Existing-Building-Retrofits/Using-Building-Data-as-a-Tool.pdf, Oct. 2012, 9 pages.
“ASHRAE Dashboard Research Project,” 29 pages, Aug. 28, 2008.
Olken et al., “Object Lessons Learned from a Distributed System for Remote Building Monitoring and Operation,” ACM SIGPLAN Notices, vol. 33, No. 10, pp. 284-295, Oct. 1998.
Proliphix, Inc., “Proliphix IP Devices: HTTP API,” 28 pages, Jan. 23, 2006.
Proliphix, Inc., Remote Management User Guide, 12 pages, prior to Aug. 27, 2007.
Richard Rogan et al., “Smart and Final Food Stores: A Case Study in Web Based Energy Information and Collection,” Web Based Energy Information and Control Systems: Case Studies and Application, Chapter 6, p. 59-64, 2005.
Sharp, “Actius AL3DU 3D LC Display High Performance 3D Visualization,” 2 pages, prior to Mar. 17, 2006.
So et al., “Building Automation on the Information Superhighway,” ASHRAE (American Society of Heating Refrigerating, and Air Conditioning) Transactions, vol. 104, Part 2, pp. 176-191, 1998.
So et al., “Building Automation Systems on the Internet,” Facilities vol. 15, No. 5/6, pp. 125-133, May/Jun. 1997.
Talon, “Raptor Controller,” 6 pages, Oct. 2003.
Talon, “Workstation Software,” 4 pages, Nov. 2002.
Trane, “System Programming, Tracer Summit Version 14, BMTW-SVP01D-EN,” 623 pages, 2002.
Lucid Design Group, Inc., “Building Dashboard,” 2 pages, Printed May 30, 2013.
“America's Largest Managed Security Services Provider Launches Comprehensive, Integrated Covid-19 Safety Program for Office Buildings and Suites,” KastleSafeSpaces, 5 pages, May 11, 2020.
“Biometric Door Reader With Body Temperature Detection,” Kintronics, 9 pages, accessed May 21, 2020.
“Body Surface Temperature Screening with Alarm Function TVS-200IS/TVS-5O0IS,” Nippon Avionics Co., 3 pages, accessed May 21, 2020.
“BriefCam announces video analytics innovation for contact tracing, physical distancing, occupancy management and face mask detection,” BriefCam LTD, 11 pages, Jun. 5, 2020.
“Thermal Imaging SmartPhone Can Be used For Temperature Screening of People,” CAT, 3 pages, accessed Jul. 13, 2020.
“Contact Tracing Now Available on Identiv's Hirsch Velocity Access Control Platform,” Identiv, 5 pages, May 21, 2020.
Silva et al., “Cough localization for the detection of respiratory diseases in pig houses,” ScienceDirect, 7 pages, May 28, 2008.
Oey et al., “Evaluation of Isolation Compliance Using Real Time Video In Critical Care,” North Shore University Hospital, 1 page, Oct. 9, 2015.
“Facial Attendace System With Temperature Screening Now In India,” IANS, 5 pages, Mar. 19, 2020.
“Plan to Re-Open,” EHIGH, 16 pages, accessed Jun. 13, 2020.
“How Smarter AI-Powered Cameras Can Mitigate the Spread of Wuhan Novel,” AnyConnect, 22 pages, 2020.
“How to fight COVID-19 with machine learning,” DataRevenue, 20 pages, accessed May 25, 2020.
“INNCONTROL 5,” Honeywell, 2 pages, Aug. 8, 2018.
“IP Door Access Control,” Kintronics, 21 pages, 2014.
“Kogniz Al Health Response Platform,” Kogniz, 9 pages, accessed May 21, 2020.
“Machine Learning Could Check If You're Social Distancing Properly at Work,” MIT Technology Review, 7 pages, Apr. 17, 2020.
Punn et al., “Monitoring COVID-19 social distancing with person detection and tracking via fine-tuned YOLO v3 and Deepsort techniques,” 10 pages, May 6, 2020.
Burt, “NEC launches dual face biometric and fever detection system for access control,” Biometric Update, 4 pages, May 8, 2020.
“Remote temperature monitoring,” AXIS Communication, 10 pages, 2014.
“FebriEye-Al Based Thermal Temperature Screening System,” vehant, 1 page, 2020.
“See The World In A New Way Hikvision Thermal Cameras,” Hikvision, 12 pages, 2017.
Allain, “Trying out the iPhone Infrared Camera: The FLIR One,” Wired, 15 pages, 2014.
Dasgupta, “Your voice may be able to tell you if you have Covid,” Hindustan Times, 4 pages, Apr. 16, 2020.
Ganguty, “Gurugram-based startup Staqu has modified AI-powered JARVIS to battle coronavirus,” Yourstory, 7 pages, Mar. 31, 2020.
“White Paper Usage of D6T-44L/D6T-8L Thermal Sensor,” OMRON Corporation, 18 pages, 2013.
“MEMS: Thermal Sensors-D6T,” OMRON C Electronic Components LLC, 7 pages, Nov. 2012.
Extended European Search Report EP Application No. 20217467.8-1205, May 28, 2021 (6 pages).
Zhang et al.; “Real-Time Human Motion Behavior Detection via CNN Using mmWave Radar,” IEEE Sensors Council, 4 pages, Feb. 2019.
“A Sensor Above the Rest,” VAYYAR, 6 pages, accessed Oct. 7, 2020.
Thakur, “Infrared Sensors for Autonomous Vehicles,” Intech Open, 17 p. 2018.
“Presence Sensing”, 1 page, accessed Jul. 21, 2021.
Trane, “Creating Input/Output Objects,” 196 pages, retrieved Jul. 10, 2020.
Trane, “Using the Graphing Control Editor,” 181 pages, retrieved Jul. 10, 2020.
Related Publications (1)
Number Date Country
20230101344 A1 Mar 2023 US
Continuations (1)
Number Date Country
Parent PCT/CN2021/121137 Sep 2021 WO
Child 17556445 US