This application relates generally to a building management system of a building. This application relates more particularly to systems for managing, processing, and visualizing data for the building.
A building management system may aggregate and store building data received from building equipment and/or other data sources. The building data can be stored in a database. The building management system can include a building system that operates analytic and/or control algorithms against the data of the database to control the building equipment. However, the development and/or deployment of the analytic and/or control algorithms may be time consuming and require a significant amount of software development. Furthermore, the analytic and/or control algorithms may lack flexibility to adapt to changing circumstances in the building. In some cases, the output data of the analytic and/or control algorithms may be hard for a user to conceptualize and relate to the physical components of the building for which the information is generated.
One implementation of the present disclosure is a building system including one or more storage devices storing instructions thereon that, when executed by one or more processors, cause the one or more processors to ingest information comprising at least one of occupancy information or energy usage information associated with a building. The instructions further cause the one or more processors to generate a space usage recommendation based on the information. The instructions further cause the one or more processors to cause a graphical model of the building to include a representation of the information and the space usage recommendation. The instructions further cause the one or more processors to cause a display device of a user device to display the graphical model within a user interface.
Another implementation of the present disclosure is a building system including one or more storage devices storing instructions thereon that, when executed by one or more processors, cause the one or more processors to ingest information comprising at least one of occupancy information or energy usage information associated with a building. The instructions further cause the one or more processors to generate a building recommendation based on the information. The instructions further cause the one or more processors to cause a graphical model of the building to include the building recommendation. The instructions further cause the one or more processors to cause a display device of a user device to display the graphical model within a user interface. The instructions further cause the one or more processors to receive a request via the user interface to perform an action associated with the building recommendation. The instructions further cause the one or more processors to transmit a command to perform the action to a device associated with the building recommendation.
Another implementation of the present disclosure is a method including ingesting, by one or more processors of a building system, information comprising at least one of occupancy information or energy usage information associated with a building. The method further includes generating, by the one or more processors, a building recommendation based on the information. The method further includes causing, by the one or more processors, a graphical model of the building to include a representation of the information and the building recommendation. The method further includes causing, by the one or more processors, a display device of a user device to display the graphical model within a user interface.
Various objects, aspects, features, and advantages of the disclosure will become more apparent and better understood by referring to the detailed description taken in conjunction with the accompanying drawings, in which like reference characters identify corresponding elements throughout. In the drawings, like reference numbers generally indicate identical, functionally similar, and/or structurally similar elements.
Referring generally to the FIGURES, systems and methods for generating three dimensional graphical models (e.g., building models) with intelligent visualization are shown, according to various exemplary embodiments. For example, the systems and methods described herein may pull in or ingest various information, such as a plurality of digital twins (e.g., graph projections associated with virtually represented assets), a variety of externally accessed information relating to one or more virtually represented assets, and/or various other information relating to, associated with, or otherwise pertaining to a graphical model to be generated and displayed to a user.
In some instances, a digital twin can be a virtual representation of a building and/or an entity of the building (e.g., space, piece of equipment, occupant, etc.). Furthermore, the digital twin can represent a service performed in a building, e.g., facility management, clean air optimization, energy prediction, equipment maintenance, etc. In some instances, the systems and methods described herein allow for the cross-correlation of information received or ingested from one or more external sources or systems (e.g., via one or more external access application programming interface (APIs) or software development kit (SDK) components) by using one or more device or asset identification numbers to determine a location of a corresponding virtual asset (e.g., associated with an ingested digital twin) within the graphical model. The cross-correlated information may then be visually represented within the graphical model by displaying the cross-correlated information near the corresponding virtual asset or by utilizing the cross-correlated information to alter a visual representation of the virtual asset itself (e.g., creating a heat map at a cross-correlated location or space within the graphical model, highlighting the corresponding virtual asset within the graphical model, etc.).
In some embodiments, each digital twin can include an information data store and a connector. The information data store can store the information describing the entity that the digital twin operates for (e.g., attributes of the entity, measurements associated with the entity, control points or commands of the entity, etc.). In some embodiments, the data store can be a graph including various nodes and edges. The connector can be a software component that provides telemetry from the entity (e.g., physical device) to the information store. In some embodiments, the systems and methods described herein are configured to allow for various cross-correlated information received from or ingested from the one or more external sources or systems to be pushed to the corresponding digital twin associated with the virtual asset and used to update one or more pieces of stored information of the digital twin.
In some embodiments, the systems and methods described herein can cause the graphical model to render in a user interface of a user device and allow a user to view the model, view information associated with the components of the model, and/or navigate throughout the model. In some embodiments, a user can provide commands and/or inputs via the user device within the rendered graphical model to request information from and/or push data to one or more of the digital twins and/or one or more external sources or systems associated with one or more virtual assets. In some instances, the commands and/or inputs may further trigger one or more actions by one or more physical assets (e.g., increasing the set point temperature of an air conditioning unit) corresponding to one or more virtual assets interacted with by the user within the graphical model.
Referring now to
The building data platform 100 includes applications 110. The applications 110 can be various applications that operate to manage the building subsystems 122. The applications 110 can be remote or on-premises applications (or a hybrid of both) that run on various computing systems. The applications 110 can include an alarm application 168 configured to manage alarms for the building subsystems 122. The applications 110 include an assurance application 170 that implements assurance services for the building subsystems 122. In some embodiments, the applications 110 include an energy application 172 configured to manage the energy usage of the building subsystems 122. The applications 110 include a security application 174 configured to manage security systems of the building.
In some embodiments, the applications 110 and/or the cloud platform 106 interacts with a user device 176. In some embodiments, a component or an entire application of the applications 110 runs on the user device 176. The user device 176 may be a laptop computer, a desktop computer, a smartphone, a tablet, and/or any other device with an input interface (e.g., touch screen, mouse, keyboard, etc.) and an output interface (e.g., a speaker, a display, etc.).
The applications 110, the twin manager 108, the cloud platform 106, and the edge platform 102 can be implemented on one or more computing systems, e.g., on processors and/or memory devices. For example, the edge platform 102 includes processor(s) 118 and memories 120, the cloud platform 106 includes processor(s) 124 and memories 126, the applications 110 include processor(s) 164 and memories 166, and the twin manager 108 includes processor(s) 148 and memories 150.
The processors can be general purpose or specific purpose processors, an application specific integrated circuit (ASIC), one or more field programmable gate arrays (FPGAs), a group of processing components, or other suitable processing components. The processors may be configured to execute computer code and/or instructions stored in the memories or received from other computer readable media (e.g., CDROM, network storage, a remote server, etc.).
The memories can include one or more devices (e.g., memory units, memory devices, storage devices, etc.) for storing data and/or computer code for completing and/or facilitating the various processes described in the present disclosure. The memories can include random access memory (RAM), read-only memory (ROM), hard drive storage, temporary storage, non-volatile memory, flash memory, optical memory, or any other suitable memory for storing software objects and/or computer instructions. The memories can include database components, object code components, script components, or any other type of information structure for supporting the various activities and information structures described in the present disclosure. The memories can be communicably connected to the processors and can include computer code for executing (e.g., by the processors) one or more processes described herein.
The edge platform 102 can be configured to provide connection to the building subsystems 122. The edge platform 102 can receive messages from the building subsystems 122 and/or deliver messages to the building subsystems 122. The edge platform 102 includes one or multiple gateways, e.g., the gateways 112-116. The gateways 112-116 can act as a gateway between the cloud platform 106 and the building subsystems 122. The gateways 112-116 can be or function similar to the gateways described in U.S. patent application Ser. No. 17/127,303, filed Dec. 18, 2020, the entirety of which is incorporated by reference herein. In some embodiments, the applications 110 can be deployed on the edge platform 102. In this regard, lower latency in management of the building subsystems 122 can be realized.
The edge platform 102 can be connected to the cloud platform 106 via a network 104. The network 104 can communicatively couple the devices and systems of building data platform 100. In some embodiments, the network 104 is at least one of and/or a combination of a Wi-Fi network, a wired Ethernet network, a ZigBee network, a Bluetooth network, and/or any other wireless network. The network 104 may be a local area network or a wide area network (e.g., the Internet, a building WAN, etc.) and may use a variety of communications protocols (e.g., BACnet, IP, LON, etc.). The network 104 may include routers, modems, servers, cell towers, satellites, and/or network switches. The network 104 may be a combination of wired and wireless networks.
The cloud platform 106 can be configured to facilitate communication and routing of messages between the applications 110, the twin manager 108, the edge platform 102, and/or any other system. The cloud platform 106 can include a platform manager 128, a messaging manager 140, a command processor 136, and an enrichment manager 138. In some embodiments, the cloud platform 106 can facilitate messaging between the building data platform 100 via the network 104.
The messaging manager 140 can be configured to operate as a transport service that controls communication with the building subsystems 122 and/or any other system, e.g., managing commands to devices (C2D), commands to connectors (C2C) for external systems, commands from the device to the cloud (D2C), and/or notifications. The messaging manager 140 can receive different types of data from the applications 110, the twin manager 108, and/or the edge platform 102. The messaging manager 140 can receive change on value data 142, e.g., data that indicates that a value of a point has changed. The messaging manager 140 can receive time series data 144, e.g., a time correlated series of data entries each associated with a particular time stamp. Furthermore, the messaging manager 140 can receive command data 146. All of the messages handled by the cloud platform 106 can be handled as an event, e.g., the data 142-146 can each be packaged as an event with a data value occurring at a particular time (e.g., a temperature measurement made at a particular time).
The cloud platform 106 includes a command processor 136. The command processor 136 can be configured to receive commands to perform an action from the applications 110, the building subsystems 122, the user device 176, etc. The command processor 136 can manage the commands, determine whether the commanding system is authorized to perform the particular commands, and communicate the commands to the commanded system, e.g., the building subsystems 122 and/or the applications 110. The commands could be a command to change an operational setting that control environmental conditions of a building, a command to run analytics, etc.
The cloud platform 106 includes an enrichment manager 138. The enrichment manager 138 can be configured to enrich the events received by the messaging manager 140. The enrichment manager 138 can be configured to add contextual information to the events. The enrichment manager 138 can communicate with the twin manager 108 to retrieve the contextual information. In some embodiments, the contextual information is an indication of information related to the event. For example, if the event is a time series temperature measurement of a thermostat, contextual information such as the location of the thermostat (e.g., what room), the equipment controlled by the thermostat (e.g., what VAV), etc. can be added to the event. In this regard, when a consuming application, e.g., one of the applications 110 receives the event, the consuming application can operate based on the data of the event, the temperature measurement, and also the contextual information of the event.
The enrichment manager 138 can solve a problem that when a device produces a significant amount of information, the information may contain simple data without context. An example might include the data generated when a user scans a badge at a badge scanner of the building subsystems 122. This physical event can generate an output event including such information as “DeviceBadgeScannerID,” “BadgeID,” and/or “Date/Time.” However, if a system sends this data to a consuming application, e.g., Consumer A and a Consumer B, each customer may need to call the building data platform knowledge service to query information with queries such as, “What space, build, floor is that badge scanner in?” or “What user is associated with that badge?”
By performing enrichment on the data feed, a system can be able to perform inferences on the data. A result of the enrichment may be transformation of the message “DeviceBadgeScannerId, BadgeId, Date/Time,” to “Region, Building, Floor, Asset, DeviceId, BadgeId, UserName, EmployeeId, Date/Time Scanned.” This can be a significant optimization, as a system can reduce the number of calls by 1/n, where n is the number of consumers of this data feed.
By using this enrichment, a system can also have the ability to filter out undesired events. If there are 100 building in a campus that receive 100,000 events per building each hour, but only 1 building is actually commissioned, only 1/10 of the events are enriched. By looking at what events are enriched and what events are not enriched, a system can do traffic shaping of forwarding of these events to reduce the cost of forwarding events that no consuming application wants or reads.
An example of an event received by the enrichment manager 138 may be:
An example of an enriched event generated by the enrichment manager 138 may be:
By receiving enriched events, an application of the applications 110 can be able to populate and/or filter what events are associated with what areas. Furthermore, user interface generating applications can generate user interfaces that include the contextual information based on the enriched events.
The cloud platform 106 includes a platform manager 128. The platform manager 128 can be configured to manage the users and/or subscriptions of the cloud platform 106. For example, what subscribing building, user, and/or tenant utilizes the cloud platform 106. The platform manager 128 includes a provisioning service 130 configured to provision the cloud platform 106, the edge platform 102, and the twin manager 108. The platform manager 128 includes a subscription service 132 configured to manage a subscription of the building, user, and/or tenant while the entitlement service 134 can track entitlements of the buildings, users, and/or tenants.
The twin manager 108 can be configured to manage and maintain a digital twin. The digital twin can be a digital representation of the physical environment, e.g., a building. The twin manager 108 can include a change feed generator 152, a schema and ontology 154, a graph projection manager 156, a policy manager 158, an entity, relationship, and event database 160, and a graph projection database 162.
The graph projection manager 156 can be configured to construct graph projections and store the graph projections in the graph projection database 162. Example of graph projections are shown in
In some embodiment, the graph projection manager 156 generates a graph projection for a particular user, application, subscription, and/or system. In this regard, the graph projection can be generated based on policies for the particular user, application, and/or system in addition to an ontology specific for that user, application, and/or system. In this regard, an entity could request a graph projection and the graph projection manager 156 can be configured to generate the graph projection for the entity based on policies and an ontology specific to the entity. The policies can indicate what entities, relationships, and/or events the entity has access to. The ontology can indicate what types of relationships between entities the requesting entity expects to see, e.g., floors within a building, devices within a floor, etc. Another requesting entity may have an ontology to see devices within a building and applications for the devices within the graph.
The graph projections generated by the graph projection manager 156 and stored in the graph projection database 162 can be a knowledge graph and is an integration point. For example, the graph projections can represent floor plans and systems associated with each floor. Furthermore, the graph projections can include events, e.g., telemetry data of the building subsystems 122. The graph projections can show application services as nodes and API calls between the services as edges in the graph. The graph projections can illustrate the capabilities of spaces, users, and/or devices. The graph projections can include indications of the building subsystems 122, e.g., thermostats, cameras, air handling units, variable air volume (VAV) systems, cooling towers, pumps, chillers, valves, dampers, lighting, light sensors, fire and safety devices, access control devices, parking sensors, Wifi devices, audio/visual systems, etc. The graph projection database 162 can store graph projections that keep up a current state of a building.
The graph projections of the graph projection database 162 can be digital twins of a building. Digital twins can be digital replicas of physical entities (e.g., locations, spaces, equipment, assets, etc.) that enable an in-depth analysis of data of the physical entities and provide the potential to monitor systems to mitigate risks, manage issues, and utilize simulations to test future solutions. Digital twins can play an important role in helping technicians find the root cause of issues and solve problems faster, in supporting safety and security protocols, and in supporting building managers in more efficient use of energy and other facilities resources. Digital twins can be used to enable and unify security systems, employee experience, facilities management, sustainability, etc.
In some embodiments the enrichment manager 138 can use a graph projection of the graph projection database 162 to enrich events. In some embodiments, the enrichment manager 138 can identify nodes and relationships that are associated with, and are pertinent to, the device that generated the event. For example, the enrichment manager 138 could identify a thermostat generating a temperature measurement event within the graph. The enrichment manager 138 can identify relationships between the thermostat and spaces, e.g., a zone that the thermostat is located in. The enrichment manager 138 can add an indication of the zone to the event.
Furthermore, the command processor 136 can be configured to utilize the graph projections to command the building subsystems 122. The command processor 136 can identify a policy for a commanding entity within the graph projection to determine whether the commanding entity has the ability to make the command. For example, the command processor 136, before allowing a user to make a command, may determine, based on the graph projection database 162, that the user has a policy to be able to make the command.
In some embodiments, the policies can be conditional based policies. For example, the building data platform 100 can apply one or more conditional rules to determine whether a particular system has the ability to perform an action. In some embodiments, the rules analyze a behavioral based biometric. For example, a behavioral based biometric can indicate normal behavior and/or normal behavior rules for a system. In some embodiments, when the building data platform 100 determines, based on the one or more conditional rules, that an action requested by a system does not match a normal behavior, the building data platform 100 can deny the system the ability to perform the action and/or request approval from a higher level system.
For example, a behavior rule could indicate that a user has access to log into a system with a particular IP address between 8 A.M. through 5 P.M. However, if the user logs in to the system at 7 P.M., the building data platform 100 may contact an administrator to determine whether to give the user permission to log in.
The change feed generator 152 can be configured to generate a feed of events that indicate changes to the digital twin, e.g., to the graph. The change feed generator 152 can track changes to the entities, relationships, and/or events of the graph. For example, the change feed generator 152 can detect an addition, deletion, and/or modification of a node or edge of the graph, e.g., changing the entities, relationships, and/or events within the database 160. In response to detecting a change to the graph, the change feed generator 152 can generate an event summarizing the change. The event can indicate what nodes and/or edges have changed and how the nodes and edges have changed. The events can be posted to a topic by the change feed generator 152.
The change feed generator 152 can implement a change feed of a knowledge graph. The building data platform 100 can implement a subscription to changes in the knowledge graph. When the change feed generator 152 posts events in the change feed, subscribing systems or applications can receive the change feed event. By generating a record of all changes that have happened, a system can stage data in different ways, and then replay the data back in whatever order the system wishes. This can include running the changes sequentially one by one and/or by jumping from one major change to the next. For example, to generate a graph at a particular time, all change feed events up to the particular time can be used to construct the graph.
The change feed can track the changes in each node in the graph and the relationships related to them, in some embodiments. If a user wants to subscribe to these changes and the user has proper access, the user can simply submit a web API call to have sequential notifications of each change that happens in the graph. A user and/or system can replay the changes one by one to reinstitute the graph at any given time slice. Even though the messages are “thin” and only include notification of change and the reference “id/seq id,” the change feed can keep a copy of every state of each node and/or relationship so that a user and/or system can retrieve those past states at any time for each node. Furthermore, a consumer of the change feed could also create dynamic “views” allowing different “snapshots” in time of what the graph looks like from a particular context. While the twin manager 108 may contain the history and the current state of the graph based upon schema evaluation, a consumer can retain a copy of that data, and thereby create dynamic views using the change feed.
The schema and ontology 154 can define the message schema and graph ontology of the twin manager 108. The message schema can define what format messages received by the messaging manager 140 should have, e.g., what parameters, what formats, etc. The ontology can define graph projections, e.g., the ontology that a user wishes to view. For example, various systems, applications, and/or users can be associated with a graph ontology. Accordingly, when the graph projection manager 156 generates a graph projection for a user, system, or subscription, the graph projection manager 156 can generate a graph projection according to the ontology specific to the user. For example, the ontology can define what types of entities are related in what order in a graph, for example, for the ontology for a subscription of “Customer A,” the graph projection manager 156 can create relationships for a graph projection based on the rule:
RegionBuildingFloorSpaceAsset
For the ontology of a subscription of “Customer B,” the graph projection manager 156 can create relationships based on the rule:
BuildingFloorAsset
The policy manager 158 can be configured to respond to requests from other applications and/or systems for policies. The policy manager 158 can consult a graph projection to determine what permissions different applications, users, and/or devices have. The graph projection can indicate various permissions that different types of entities have and the policy manager 158 can search the graph projection to identify the permissions of a particular entity. The policy manager 158 can facilitate fine grain access control with user permissions. The policy manager 158 can apply permissions across a graph, e.g., if “user can view all data associated with floor 1” then they see all subsystem data for that floor, e.g., surveillance cameras, heating, ventilation, and/or air conditioning (“HVAC”) devices, fire detection and response devices, etc.
The twin manager 108 includes a query manager 165 and a twin function manager 167. The query manger 164 can be configured to handle queries received from a requesting system, e.g., the user device 176, the applications 110, and/or any other system. The query manager 165 can receive queries that include query parameters and context. The query manager 165 can query the graph projection database 162 with the query parameters to retrieve a result. The query manager 165 can then cause an event processor, e.g., a twin function, to operate based on the result and the context. In some embodiments, the query manager 165 can select the twin function based on the context and/or perform operates based on the context. In some embodiments, the query manager 165 is configured to perform a variety of differing operations. For example, in some instances, the query manager 165 is configured to perform any of the operations performed by the query manager described in U.S. patent application Ser. No. 17/537,046, filed Nov. 29, 2021, the entirety of which is incorporated by reference herein.
The twin function manager 167 can be configured to manage the execution of twin functions. The twin function manager 167 can receive an indication of a context query that identifies a particular data element and/or pattern in the graph projection database 162. Responsive to the particular data element and/or pattern occurring in the graph projection database 162 (e.g., based on a new data event added to the graph projection database 162 and/or change to nodes or edges of the graph projection database 162), the twin function manager 167 can cause a particular twin function to execute. The twin function can be executed based on an event, context, and/or rules. The event can be data that the twin function executes against. The context can be information that provides a contextual description of the data, e.g., what device the event is associated with, what control point should be updated based on the event, etc. The twin function manager 167 can be configured to perform a variety of differing operations. For example, in some instances, the twin function manager 167 is configured to perform any of the operations of the twin function manager described in U.S. patent application Ser. No. 17/537,046, referenced above.
Referring now to
The graph projection 200 includes a device hub 202 which may represent a software service that facilitates the communication of data and commands between the cloud platform 106 and a device of the building subsystems 122, e.g., door actuator 214. The device hub 202 is related to a connector 204, an external system 206, and a digital asset “Door Actuator” 208 by edge 250, edge 252, and edge 254.
The cloud platform 106 can be configured to identify the device hub 202, the connector 204, the external system 206 related to the door actuator 214 by searching the graph projection 200 and identifying the edges 250-254 and edge 258. The graph projection 200 includes a digital representation of the “Door Actuator,” node 208. The digital asset “Door Actuator” 208 includes a “DeviceNameSpace” represented by node 207 and related to the digital asset “Door Actuator” 208 by the “Property of Object” edge 256.
The “Door Actuator” 214 has points and time series. The “Door Actuator” 214 is related to “Point A” 216 by a “has_a” edge 260. The “Door Actuator” 214 is related to “Point B” 218 by a “has_A” edge 259. Furthermore, time series associated with the points A and B are represented by nodes “TS” 220 and “TS” 222. The time series are related to the points A and B by “has_a” edge 264 and “has_a” edge 262. The time series “TS” 220 has particular samples, sample 210 and 212 each related to “TS” 220 with edges 268 and 266 respectively. Each sample includes a time and a value. Each sample may be an event received from the door actuator that the cloud platform 106 ingests into the entity, relationship, and event database 160, e.g., ingests into the graph projection 200.
The graph projection 200 includes a building 234 representing a physical building. The building includes a floor represented by floor 232 related to the building 234 by the “has_a” edge from the building 234 to the floor 232. The floor has a space indicated by the edge “has_a” 270 between the floor 232 and the space 230. The space has particular capabilities, e.g., is a room that can be booked for a meeting, conference, private study time, etc. Furthermore, the booking can be canceled. The capabilities for the floor 232 are represented by capabilities 228 related to space 230 by edge 280. The capabilities 228 are related to two different commands, command “book room” 224 and command “cancel booking” 226 related to capabilities 228 by edge 284 and edge 282 respectively.
If the cloud platform 106 receives a command to book the space represented by the node, space 230, the cloud platform 106 can search the graph projection 200 for the capabilities for the 228 related to the space 230 to determine whether the cloud platform 106 can book the room.
In some embodiments, the cloud platform 106 could receive a request to book a room in a particular building, e.g., the building 234. The cloud platform 106 could search the graph projection 200 to identify spaces that have the capabilities to be booked, e.g., identify the space 230 based on the capabilities 228 related to the space 230. The cloud platform 106 can reply to the request with an indication of the space and allow the requesting entity to book the space 230.
The graph projection 200 includes a policy 236 for the floor 232. The policy 236 is related set for the floor 232 based on a “To Floor” edge 274 between the policy 236 and the floor 232. The policy 236 is related to different roles for the floor 232, read events 238 via edge 276 and send command 240 via edge 278. The policy 236 is set for the entity 203 based on has edge 251 between the entity 203 and the policy 236.
The twin manager 108 can identify policies for particular entities, e.g., users, software applications, systems, devices, etc. based on the policy 236. For example, if the cloud platform 106 receives a command to book the space 230. The cloud platform 106 can communicate with the twin manager 108 to verify that the entity requesting to book the space 230 has a policy to book the space. The twin manager 108 can identify the entity requesting to book the space as the entity 203 by searching the graph projection 200. Furthermore, the twin manager 108 can further identify the edge has 251 between the entity 203 and the policy 236 and the edge 1178 between the policy 236 and the command 240.
Furthermore, the twin manager 108 can identify that the entity 203 has the ability to command the space 230 based on the edge 274 between the policy 236 and the floor 232 and the edge 270 between the floor 232 and the space 230. In response to identifying the entity 203 has the ability to book the space 230, the twin manager 108 can provide an indication to the cloud platform 106.
Furthermore, if the entity 203 makes a request to read events for the space 230, e.g., the sample 210 and the sample 212, the twin manager 108 can identify the edge has 251 between the entity 203 and the policy 236, the edge 276 between the policy 236 and the read events 238, the edge 274 between the policy 236 and the floor 232, the “has_a” edge 270 between the floor 232 and the space 230, the edge 271 between the space 230 and the door actuator 214, the edge 260 between the door actuator 214 and the point A 216, the “has_a” edge 264 between the point A 216 and the TS 220, and the edges 268 and 266 between the TS 220 and the samples 210 and 212 respectively.
Additional examples of potential graph projections can be found in U.S. patent application Ser. No. 17/537,046, referenced above. However, it will be appreciated that a variety of differing graph projections may be implemented, as desired for a given application or scenario. As such, the example graph projections provided herein are provided as examples, and are in no way meant to be limiting.
Referring now to
A digital twin (or a shadow) may be a computing entity that describes a physical thing (e.g., a building, spaces of a building, devices of a building, people of the building, equipment of the building, etc.) through modeling the physical thing through a set of attributes that define the physical thing. A digital twin can refer to a digital replica of physical assets (a physical device twin) and can be extended to store processes, people, places, systems that can be used for various purposes. The digital twin can include both the ingestion of information and actions learned and executed through artificial intelligence agents.
In
The twin manager 108 stores the graph 329 which may be a graph data structure including various nodes and edges interrelating the nodes. The graph 329 may be the same as, or similar to, the graph projections described herein with reference to
The floor node 322 is related to the zone node 318 by the “has” edge 340 indicating that the floor represented by the floor node 322 has another zone represented by the zone node 318. The floor node 322 is related to another zone node 324 via a “has” edge 342 representing that the floor represented by the floor node 322 has a third zone represented by the zone node 324.
The graph 329 includes an AHU node 314 representing an AHU of the building represented by the building node 326. The AHU node 314 is related by a “supplies” edge 330 to the VAV node 312 to represent that the AHU represented by the AHU node 314 supplies air to the VAV represented by the VAV node 312. The AHU node 314 is related by a “supplies” edge 336 to the VAV node 320 to represent that the AHU represented by the AHU node 314 supplies air to the VAV represented by the VAV node 320. The AHU node 314 is related by a “supplies” edge 332 to the VAV node 316 to represent that the AHU represented by the AHU node 314 supplies air to the VAV represented by the VAV node 316.
The VAV node 316 is related to the zone node 318 via the “serves” edge 334 to represent that the VAV represented by the VAV node 316 serves (e.g., heats or cools) the zone represented by the zone node 318. The VAV node 320 is related to the zone node 324 via the “serves” edge 338 to represent that the VAV represented by the VAV node 320 serves (e.g., heats or cools) the zone represented by the zone node 324. The VAV node 312 is related to the zone node 310 via the “serves” edge 328 to represent that the VAV represented by the VAV node 312 serves (e.g., heats or cools) the zone represented by the zone node 310.
Furthermore, the graph 329 includes an edge 333 related to a time series node 364. The time series node 364 can be information stored within the graph 329 and/or can be information stored outside the graph 329 in a different database (e.g., a time series database). In some embodiments, the time series node 364 stores time series data (or any other type of data) for a data point of the VAV represented by the VAV node 316. The data of the time series node 364 can be aggregated and/or collected telemetry data of the time series node 364.
Furthermore, the graph 329 includes an edge 337 related to a time series node 366. The time series node 366 can be information stored within the graph 329 and/or can be information stored outside the graph 329 in a different database (e.g., a time series database). In some embodiments, the time series node 366 stores time series data (or any other type of data) for a data point of the VAV represented by the VAV node 316. The data of the time series node 364 can be inferred information, e.g., data inferred by one of the artificial intelligence agents 370 and written into the time series node 364 by the artificial intelligence agent 370. In some embodiments, the time series 364 and/or 366 are stored in the graph 329 but are stored as references to time series data stored in a time series database.
The twin manager 108 includes various software components. For example, the twin manager 108 includes a device management component 348 for managing devices of a building. The twin manager 108 includes a tenant management component 350 for managing various tenant subscriptions. The twin manager 108 includes an event routing component 352 for routing various events. The twin manager 108 includes an authentication and access component 354 for performing user and/or system authentication and grating the user and/or system access to various spaces, pieces of software, devices, etc. The twin manager 108 includes a commanding component 356 allowing a software application and/or user to send commands to physical devices. The twin manager 108 includes an entitlement component 358 that analyzes the entitlements of a user and/or system and grants the user and/or system abilities based on the entitlements. The twin manager 108 includes a telemetry component 360 that can receive telemetry data from physical systems and/or devices and ingest the telemetry data into the graph 329. For example, the telemetry data can come from thermostats, cameras, air handling units, variable air volume (VAV) systems, cooling towers, pumps, chillers, valves, dampers, lighting, light sensors, fire and safety devices, access control devices, parking sensors, Wi-fi devices, audio/visual systems, or any other devices within the building. Furthermore, the twin manager 108 includes an integrations component 362 allowing the twin manager 108 to integrate with other applications.
The twin manager 108 includes a gateway 306 and a twin connector 308. The gateway 306 can be configured to integrate with other systems and the twin connector 308 can be configured to allow the gateway 306 to integrate with the twin manager 108. The gateway 306 and/or the twin connector 308 can receive an entitlement request 302 and/or an inference request 304. The entitlement request 302 can be a request received from a system and/or a user requesting that an AI agent action be taken by the AI agent 370. The entitlement request 302 can be checked against entitlements for the system and/or user to verify that the action requested by the system and/or user is allowed for the user and/or system. The inference request 304 can be a request that the AI agent 370 generates an inference, e.g., a projection of information, a prediction of a future data measurement, an extrapolated data value, etc.
The cloud platform 106 is shown to receive a manual entitlement request 386. The request 386 can be received from a system, application, and/or user device (e.g., from the applications 110, the building subsystems 122, and/or the user device 176). The manual entitlement request 386 may be a request for the AI agent 370 to perform an action, e.g., an action that the requesting system and/or user has an entitlement for. The cloud platform 106 can receive the manual entitlement request 386 and check the manual entitlement request 386 against an entitlement database 384 storing a set of entitlements to verify that the requesting system has access to the user and/or system. The cloud platform 106, responsive to the manual entitlement request 386 being approved, can create a job for the AI agent 370 to perform. The created job can be added to a job request topic 380 of a set of topics 378.
The job request topic 380 can be fed to AI agents 370. For example, the topics 380 can be fanned out to various AI agents 370 based on the AI agent that each of the topics 380 pertains to (e.g., based on an identifier that identifies an agent and is included in each job of the topic 380). The AI agents 370 include a service client 372, a connector 374, and a model 376. The model 376 can be loaded into the AI agent 370 from a set of AI models stored in the AI model storage 368. The AI model storage 368 can store models for making energy load predictions for a building, weather forecasting models for predicting a weather forecast, action/decision models to take certain actions responsive to certain conditions being met, an occupancy model for predicting occupancy of a space and/or a building, etc. The models of the AI model storage 368 can be neural networks (e.g., convolutional neural networks, recurrent neural networks, deep learning networks, etc.), decision trees, support vector machines, and/or any other type of artificial intelligence, machine learning, and/or deep learning category. In some embodiments, the models are rule based triggers and actions that include various parameters for setting a condition and defining an action.
The AI agent 370 can include triggers 395 and actions 397. The triggers 395 can be conditional rules that, when met, cause one or more of the actions 397. The triggers 395 can be executed based on information stored in the graph 329 and/or data received from the building subsystems 122. The actions 397 can be executed to determine commands, actions, and/or outputs. The output of the actions 397 can be stored in the graph 329 and/or communicated to the building subsystems 122.
The AI agent 370 can include a service client 372 that causes an instance of an AI agent to run. The instance can be hosted by the artificial intelligence service client 388. The client 388 can cause a client instance 392 to run and communicate with the AI agent 370 via a gateway 390. The client instance 392 can include a service application 394 that interfaces with a core algorithm 398 via a functional interface 396. The core algorithm 398 can run the model 376, e.g., train the model 376 and/or use the model 376 to make inferences and/or predictions.
In some embodiments, the core algorithm 398 can be configured to perform learning based on the graph 329. In some embodiments, the core algorithm 398 can read and/or analyze the nodes and relationships of the graph 329 to make decisions. In some embodiments, the core algorithm 398 can be configured to use telemetry data (e.g., the time series data 364) from the graph 329 to make inferences on and/or perform model learning. In some embodiments, the result of the inferences can be the time series 366. In some embodiments, the time series 364 is an input into the model 376 that predicts the time series 366.
In some embodiments, the core algorithm 398 can generate the time series 366 as an inference for a data point, e.g., a prediction of values for the data point at future times. The time series 364 may be actual data for the data point. In this regard, the core algorithm 398 can learn and train by comparing the inferred data values against the true data values. In this regard, the model 376 can be trained by the core algorithm 398 to improve the inferences made by the model 376.
In some embodiments, the system 300 is configured to execute one or more artificial intelligence agents to infer and/or predict information based on information obtained or otherwise retrieved from the graph 329. For example, in some instances, the system 300 may include a variety of different AI agents associated with and configured to analyze information pertaining to any of the various nodes within the graph 329. In some instances, the AI agents may analyze not only the nodes they pertain to, but also a variety of connectors and various triggers associated with those AI agents. For example, in some instances AI agents may be utilized to infer and/or predict information pertaining to the corresponding nodes, and to subsequently trigger various actions within the system 300. In some embodiment, the AI agents may trigger various actions according to associated trigger rules and action rules. The trigger rules and action rules can be logical statements and/or conditions that include parameter values and/or create associated output actions. In some instances, these trigger rules and actions rule may be defined by a user of the system 300. In some other instances, the AI agents may learn, create, or otherwise generate the trigger rules and actions rules based on various desired outcomes (e.g., reduce or minimize energy usage, improve or maximize air circulation, etc.). Example AI agents, triggers, actions, and trigger/rule learning processes are described in U.S. patent application Ser. No. 17/537,046, referenced above.
Referring now to
The system 400 includes a schema infusing tool 404. The schema infusing tool can infuse a particular schema, the schema 402, into various systems, services, and/or equipment in order to integrate the data of the various systems, services, and/or equipment into the building data platform 100. The schema 402 may be the BRICK schema, in some embodiments. In some embodiment, the schema 402 may be a schema that uses portions and/or all of the BRICK schema but also includes unique class, relationship types, and/or unique schema rules. The schema infusing tool 404 can infuse the schema 402 into systems such as systems that manage and/or produce building information model (BIM) data 418, building automation system (BAS) systems that produce BAS data 420, and/or access control and video surveillance (ACVS) systems that produce ACVS data 422. In some embodiments, the BIM data 418 can be generated by BIM automation utilities 2501.
The BIM data 418 can include data such as Revit data 424 (e.g., Navisworks data), industrial foundation class (IFC) data 426, gbxml data 428, and/or CoBie data 430. The BAS data 420 can include Modelica data 432 (e.g., Control Description Language (CDL) data), Project Haystack data 434, BACnet data 436, Metasys data 438, and/or EasyIO data 440. All of this data can utilize the schema 402 and/or be capable of being mapped into the schema 402.
The BAS data 420 and/or the ACVS data 422 may include time series data 408. The time series data 408 can include trends of data points over time, e.g., a time correlated set of data values each corresponding to time stamps. The time series data can be a time series of data measurements, e.g., temperature measurements, pressure measurements, etc. Furthermore, the time series data can be a time series of inferred and/or predicted information, e.g., an inferred temperature value, an inferred energy load, a predicted weather forecast, identities of individuals granted access to a facility over time, etc. The time series data 408 can further indicate command and/or control data, e.g., the damper position of a VAV over time, the set point of a thermostat over time, etc.
The system 400 includes a schema mapping toolchain 412. The schema mapping toolchain 412 can map the data of the metadata sources 406 into data of the schema 402, e.g., the data in schema 414. The data in schema 414 may be in a schema that can be integrated by an integration toolchain 416 with the building data platform 100 (e.g. ingested into the databases, graphs, and/or knowledge bases of the building data platform 100) and/or provided to the AI services and applications 410 for execution).
The AI services and applications 410 include building control 442, analytics 444, micro-grid management 446, and various other applications 448. The building control 442 can include various control applications that may utilize AI, ML, and/or any other software technique for managing control of a building. The building control 442 can include auto sequence of operation, optimal supervisory controls, etc. The analytics 444 include clean air optimization (CAO) applications, energy prediction model (EPM) applications, and/or any other type of analytics.
Referring now to
The system 500 includes various tools for converting the metadata sources 406 into the data in schema 414. Various mapping tools 502-512 can map data from an existing schema into the schema 402. For example, the mapping tools 502-512 can utilize a dictionary that provides mapping rules and syntax substitutions. In some embodiments, that data sources can have the schema 402 activated, e.g., schema enable 518-522. If the schema 402 is enabled for a Metasys data source, an easy IO data source, or an ACVS data sources, the output data by said systems can be in the schema 402. Examples of schema mapping techniques can be found in U.S. patent application Ser. No. 16/663,623 filed Oct. 25, 2019, U.S. patent application Ser. No. 16/885,968 filed May 28, 2020, and U.S. patent application Ser. No. 16/885,959 filed May 28, 2020, the entireties of which are incorporated by reference herein.
For the EasyIO data 440, the EasyIO controller objects could be tagged with classes of the schema 402. For the Revit data 424, the metadata of a REVIT model could be converted into the schema 402, e.g., into a resource description format (RDF). For the Metasys data 438, Metasys SCT data could be converted into RDF. An OpenRefine aided mapping tool 514 and/or a natural language aided mapping tool 516 could perform the schema translation for the BACnet data 436.
The schema data output by the tools 502-522 can be provided to a reconciliation tool 530. The reconciliation tool 530 can be configured to merge complementary or duplicate information and/or resolve any conflicts in the data received from the tools 502-522. The result of the reconciliation performed by the reconciliation tool 530 can be the data in schema 414 which can be ingested into the building data platform 100 by the ingestion tool 532. The ingestion tool 532 can generate and/or update one or more graphs managed and/or stored by the twin manager 108. For example, the graph could be any of the graphs described with reference to
The system 500 includes agents that perform operations on behalf of the AI services and applications 410. For example, as shown in the system 500, the analytics 444 are related to various agents, a CAO AI agent 524, an EPM AI agent 526, and various other AI agents 528. The agents 524-528 can receive data from the building data platform 100, e.g., the data that the ingestion tool 532 ingests into the building data platform 100, and generate analytics data for the analytics 444.
Referring now to
The system 600 includes a client 602. The client 602 can integrate with the knowledge graph 614 and also with a graphical building model 604 that can be rendered on a screen of the user device 176. For example, the knowledge graph 614 could be any of the graphs described with reference to
The client 602 can retrieve information from the knowledge graph 614, e.g., an inference generated by the CAO AI agent 524, a prediction made by the EPM AI agent 526, operational data stored in the knowledge graph 614, and/or any other relevant information. The client 602 can ingest the values of the retrieved information into the graphical building model 604 which can be displayed on the user device 176. In some embodiments, when a particular visual component is being displayed on the user device 176 for the virtual model 604, e.g., a building, the corresponding information for the building can be displayed in the interface, e.g., inferences, predictions, and/or operational data.
For example, the client 602 could identify a node of the building in the knowledge graph 614, e.g., a building node, such as building node 234. The client 602 could identify information linked to the building node via edges, e.g., an energy prediction node related to the building node via an edge. The client 602 can cause the energy prediction associated with the building node to be displayed in the graphical building model 604.
In some embodiments, a user can provide input through the graphical building model 604. The input may be a manual action that a user provides via the user device 176. The manual action can be ingested into the knowledge graph 614 and stored as a node within the knowledge graph 614. In some embodiments, the manual action can trigger one of the agents 524-526 causing the agent to generate an inference and/or prediction which is ingested into the knowledge graph 614 and presented for user review in the model 604.
In some embodiments, the knowledge graph 614 includes data for the inferences and/or predictions that the agents 524 and 526 generate. For example, the knowledge graph 614 can store information such as the size of a building, the number of floors of the building, the equipment of each floor of the building, the square footage of each floor, square footage of each zone, ceiling heights, etc. The data can be stored as nodes in the knowledge graph 614 representing the physical characteristics of the building. In some embodiments, the CAO AI agent generates inferences and/or the EPM AI agent 526 makes the predictions based on the characteristic data of the building and/or physical areas of the building.
For example, the CAO AI agent 524 can operate on behalf of a CAO AI service 616. Similarly, the EPM AI agent 526 can operate on behalf of an EPM AI service 618. Furthermore a service bus 620 can interface with the agent 524 and/or the agent 526. A user can interface with the agents 524-526 via the user device 176. The user can provide an entitlement request, e.g., a request that the user is entitled to make and can be verified by an AI agent manager 622. The AI agent manager 622 can send an AI job request based on a schedule to the service bus 620 based on the entitlement request. The service bus 620 can communicate the AI job request to the appropriate agent and/or communicate results for the AI job back to the user device 176.
In some embodiments, the CAO AI agent 524 can provide a request for generating an inference to the CAO AI service 616. The request can include data read from the knowledge graph 614, in some embodiments.
The CAO AI agent 524 includes a client 624, a schema translator 626, and a CAO client 628. The client 624 can be configured to interface with the knowledge graph 614, e.g., read data out of the knowledge graph 614. The client 624 can further ingest inferences back into the knowledge graph 614. For example, the client 624 could identify time series nodes related to one or more nodes of the knowledge graph 614, e.g., time series nodes related to an AHU node via one or more edges. The client 624 can then ingest the inference made by the CAO AI agent 524 into the knowledge graph 614, e.g., add a CAO inference or update the CAO inference within the knowledge graph 614.
The client 624 can provide data it reads from the knowledge graph 614 to a schema translator 626 that may translate the data into a specific format in a specific schema that is appropriate for consumption by the CAO client 628 and/or the CAO AI service 616. The CAO client 628 can run one or more algorithms, software components, machine learning models, etc. to generate the inference and provide the inference to the client 624. In some embodiments, the client 624 can interface with the EPM AI service 618 and provide the translated data to the EPM AI service 618 for generating an inference. The inference can be returned by the EPM AI service 618 to the CAO client 628.
The EPM AI agent 526 can operate in a similar manner to the CAO AI agent 524, in some embodiments. The client 630 can retrieve data from the knowledge graph 614 and provide the data to the schema translator 632. The schema translator 632 can translate the data into a readable format by the CAO AI service 616 and can provide the data to the EPM client 634. The EPM client 634 can provide the data along with a prediction request to the CAO AI service 616. The CAO AI service 616 can generate the prediction and provide the prediction to the EPM client 634. The EPM client 634 can provide the prediction to the client 630 and the client 630 can ingest the prediction into the knowledge graph 614.
In some embodiments, the agents 524-526 combined with the knowledge graph 614 can create a digital twin. In some embodiments, the agents 524-526 are implemented for a specific node of the knowledge graph 614, e.g., on behalf of some and/or all of the entities of the knowledge graph 614. In some embodiments, the digital twin includes trigger and/or actions as also described in U.S. patent applicant Ser. No. 17/537,046, referenced above. In this regard, the agents can trigger based on information of the knowledge graph 614 (e.g., building ingested data and/or manual commands provide via the model 604) and generate inferences and/or predictions with data of the knowledge graph 614 responsive to being triggered. The resulting inferences and/or predictions can be ingested into the knowledge graph 614. The inferences and/or predictions can be displayed within the model 604.
In some embodiments, the animations of the model 604 can be based on the inferences and/or predictions of the agents 524-526. In some embodiments, charts or graphs can be included within the model 604, e.g., charting or graphing time series values of the inferences and/or predictions. For example, if an inference is an inference of a flow rate of a fluid (e.g., water, air, refrigerant, etc.) through a conduit, the speed at which arrows moving through the virtual conduit can be controlled based on the inferred flow rate inferred by an agent. Similarly, if the model 604 provides a heat map indicating occupancy, e.g., red indicating high occupancy, blue indicating medium occupancy, and green indicating low occupancy, an agent could infer an occupancy level for each space of the building and the color coding for the heat map of the model 604 could be based on the inference made by the agent.
In some embodiments, the graphical building model 604 can be a three dimensional or two dimensional graphical building. The graphical building model 604 can be a building information model (BIM), in some embodiments. The BIM can be generated and viewed based on the knowledge graph 614. An example of rendering graph data and/or BIM data in a user interface is described in greater detail in U.S. patent application Ser. No. 17/136,752 filed Dec. 29, 2020, U.S. patent application Ser. No. 17/136,768 filed Dec. 29, 2020, and U.S. patent application Ser. No. 17/136,785 filed December 29, 2020, the entirety of which is incorporated by reference herein.
In some embodiments, the graphical building model 604 includes one or multiple three dimensional building elements 606. The three dimensional building elements 606 can form a building when combined, e.g., can form a building model of a building or a campus model of a campus. The building elements 606 can include floors of a building, spaces of a building, equipment of a building, etc. Furthermore, each three dimensional building element 606 can be linked to certain data inferences 608, predictions 610, and/or operational data 612. The data 608-612 can be retrieved from the knowledge graph 614 for display in an interface via the user device 176.
Referring now to
In some embodiments, the systems and methods described herein (e.g., the system 700 and the associated methods performed by the system 700) may be configured to ingest data from and/or output data to digital twins of a building and associated entities. In some embodiments, the systems and methods may additionally or alternatively be configured to ingest data from and/or output data to data sources/systems other than digital twins.
As shown, the system 700 includes a viewer rendering component 702 configured to communicate with a twin manager 704, one or more external access components 706 (which may also be referred to as “plug-in packs”), and various platform manager components 708 to obtain building information (or any other type of environment information) and generate the rendering of the virtual building for display on a viewer interface 710 for viewing by an end user. For example, in some embodiments, the user views the viewer interface 710 via the user device 176. It should be appreciated that the various components of the system 700 may be accessible from or otherwise stored, managed, operated, or supported by any combination of the various components of the building data platform 100 (e.g., the edge platform 102, the cloud platform 106, the twin manager 108, the applications 110, and/or any other system accessible via the network 104).
For example, in some instances, the user selects to view a rendering of a virtual building via the viewer interface 710. In some instances, the viewer interface 710 is configured to provide one or more potential virtual buildings from which the user may select to view a rendering within the viewer interface 710. For example, as shown in
With reference again to
The beckon applications 712 may additionally pull in external information from one or more external sources or computing systems via the external access components 706 to be implemented, overlaid, or otherwise incorporated within the display of the rendering of the graphical model (e.g., the virtual building). For example, in some instances, the external access components 706 may be one or more external access application programming interface (API) and/or software development kit (SDK) components. In some instances, the external access components 706 may pull external information from one or more external third-party applications associated with vendors, maintenance companies, third-party service providers (e.g., HVAC service providers, internal air quality service providers, occupancy data service providers, security service providers, fire suppression and prevention service providers, etc.), and/or other entities associated with the building being virtually rendered. In some instances, the external access components 706 may pull external information from one or more external third-party applications associated with various other entities (e.g., weather service applications, traffic monitoring applications, etc.) that may be pertinent to the virtual building being rendered. In some instances, the beckon applications 712 may further push information to the various third-party applications via the external access components 706 based on one or more inputs from the user via the viewer interface (e.g., movement of a virtual entity, a command to a given device, etc.). In some instances, the information pushed to the various third-party applications may be defined via a subscription service application, an entitlement service, and/or any other application associated with controlling the flow of information into and out of the viewer interface 710 provided to the user.
In some embodiments, the external access components 706 are configured to provide a mapping or list of commands to receive and/or request data from and/or push data to the one or more external applications or systems. In some embodiments, the external access components 706 are additionally configured to receive, request, or push information about the format and content of the data. In some embodiments, this information about the format and content of the data may include information allowing the system 700 to correlate disparate formats of multiple external systems to a format of the viewer rendering component 702 (e.g., to be displayed within the viewer interface 710).
In some instances, the beckon applications 712 further communicate with one or more of the platform manager components 708. For example, the platform manager components 708 may include a digital key service application to fetch corresponding entitlements associated with entities attempting to access or view the virtual rendering of the building. For example, in some instances, a particular entitlement may be accessed using a digital key service (e.g., a digital credential and a corresponding validation application) to ensure that the user attempting to access or view the rendering of the virtual building is entitled to so. Further, the entitlements for a given user may give the user access to varying levels of information to be displayed within or overlaid on the rendering of the virtual building, in the same or a similar manner to that described above, with reference to the entitlement service 134 of
In some instances, the platform manager components 708 may include a tenant service application configured to define the various entitlements associated with the entities (e.g., similar to the subscription service 132). In some instances, the platform manager components 708 may include applications similar to or the same as any of the provisioning service 130, the subscription service 132, and/or the entitlement service 134 described herein.
The rendering application 714 may be configured to ingest the various information fetched by the beckon applications 712 (e.g., a REVIT or NEVUS work file, associated graph projection information, various externally obtained information from third parties, etc.) and use the various information to render the graphical model (e.g., the virtual building) within the viewer interface 710. In some instances, the rendering application 714 may incorporate both the virtual representation of the various entities associated with the building (e.g., the building layout, devices within the buildings) and information pertaining to the various entities associated with the building (e.g., event information, alarm information, inferences about the entities, predictions about the entities, etc.), as discussed below, with reference to
In some instances, the rendering application 714 may receive or ingest the external information from the one or more external sources or systems via the external access components 706. In these instances, the rendering application 714 may then cross-correlate one or more device or asset identification numbers associated with the received or ingested external information with one or more device or asset identification numbers received from the twin manager 704 (e.g., associated with one or more rendered virtual assets within a virtual building) to determine a location of the corresponding virtual asset within the graphical model (e.g., within the virtual building). The rendering application 714 may then cause the graphical model (e.g., the virtual building) to include a representation of the external information associated with the virtual asset. For example, in some instances, the rendering application 714 may overlay the external information received from the one or more external sources or systems pertaining to the virtual asset within the viewer interface 710. In some instances, the rendering application 714 may modify the virtual asset within the viewer interface 710 based on the external information (e.g., a heat map having various colors based on the external information at various locations within the virtual building, highlighting one or more assets based on the external information, etc.).
Referring generally to
Referring now to
In some instances, one of the selectable user interface buttons may be clicked by the user to display an asset list window 904. Within the asset list window 904, the user is able to select from a list of virtual assets (e.g., entities) within the virtual building 902 (e.g., an asset list including all of the virtual equipment assets within the virtual building 902). In some instances, each entity displayed within the asset list window 904 includes a name of the entity (e.g., a device ID) and an accompanying entity icon. In some instances, the entities may include mechanical entities, electrical entities, plumbing entities, air distribution entities, or any other entities used within a given building. Upon selection of an entity, the user may be provided with various asset details pertaining to the entity and/or navigated to the entity within the rendering of the virtual building 902, as will be discussed below with reference to
Referring now to
Referring now to
In some instances, as discussed above, upon selecting a given entity within the asset list window 904, the user may be navigated to the entity within the rendering of the virtual building 902. In some instances, the user may select to be navigated to the entity within the rendering of the virtual building using a navigation icon presented within the asset details window 910. For example,
With reference again to
In some instances, the entity icons for the modelled assets and the unmodelled assets may be shown in different colors (e.g., the entity icons for the modelled assets may be gray and the entity icons for the unmodelled assets may be red). In some other instances, the entity icons for the modelled assets may be representative of the asset that the entity icon is associated with, while the entity icons for the unmodelled assets may be null icons (e.g., a circle with an X through it). For example, the selected entity 912 shown in
In some instances, the user is allowed to move the location of the virtual unmodelled asset 914 to a desired location within the virtual building 902. In these instances, the beckon applications 712 discussed above may communicate this location change of the unmodelled asset to the twin manager 704 to be incorporated into the corresponding graph projection associated with the unmodelled asset. That is, the user may be allowed to manipulate the position of an unmodelled asset within the viewer interface 710 and have that change communicated to and effectuated within the twin manager 704. In some instances, the user may be similarly allowed to manipulate one or more types of modelled assets within the viewer interface 710 and have those changes communicated to and effectuated within the twin manager 704 in a similar manner.
In some instances, the user is allowed to add a modelled asset or an unmodelled asset (e.g., the unmodelled asset 914) to the virtual building from a list of potential modeled assets and potential unmodelled assets. For example, in some instances, the viewer interface 710 may allow for the user to click on a particular wall (or any other selectable area) and choose have a modelled or unmodelled version of a device (e.g., depending on whether the device has an associated modelled asset) installed on that wall. In some instances, the system 700 (e.g., via one or more agents) may be configured to automatically position the added modelled or unmodelled asset on the wall (or within any other selectable area) based on a standard positioning scheme (e.g., a safety, regulatory, or normative rule for similar devices). For example, if the user is adding a light switch (e.g., a modelled or unmodelled asset representing a light switch) within a virtual room, the light switch may be automatically placed on a selected wall at a standard height and distance from a nearby door frame. Similarly, a user may add a camera (e.g., a modelled or unmodelled asset representing a camera) within a virtual room, and the camera may automatically be placed at a standard position on the ceiling (e.g., a standard distance from a corner of the room).
Referring now to
In some instances, to create this link, new assets may be manually ingested into the graph projections of the twin manager 704 via the viewer interface 710. For example, the viewer interface 710 may allow for the user to manually create associations (e.g., via one or more asset ingestion APIs, AI agents, and/or applications) between new virtual assets added to the virtual building 902 and new physical assets installed within the physical building.
In some instances, some new assets may belong to one or more BACnet protocols, and may thus be ingested into the graph projections of the twin manager 704 as connector components. To create connections with these connector components or to control the connector components, the BIM assets are ingested into the graph projections of the twin manager 704 and a relationship is created between the BIM assets and associated bit connector components. Again, this ingestion may be performed manually or, in some instances, automatically using the viewer interface 710 via one or more asset ingestion APIs, AI agents, and/or applications.
In any case, the link (e.g., the edge) connecting the modelled asset to the corresponding physical asset may allow for the user (e.g., assuming the user has the proper entitlements) to control the functioning of the asset within the physical building via interaction with the virtual building 902 within the viewer interface 710 (e.g., on the user device 176). For example, commands from the user input into the viewer interface 710 may be communicated back to the twin manager 704 to update the graph projection (e.g., a device status, a device set point), which may then be ultimately communicated to the control circuit of the physical asset (e.g., via the edge platform 102, the network 104, the cloud platform 106) to control the functionality of the physical asset.
As shown in
In some instances, the command and control component 916 may receive a command from the user regarding a virtual asset associated with a physical asset in a physical building, and the command may be communicated from the viewer interface 710 to the twin manager 704. From twin manager 704, the command may be communicated to a cloud platform (e.g., the cloud platform 106). From the cloud platform (e.g., the cloud platform 106, the command may be communicated to an edge platform (e.g., the edge platform 102), which may ultimately provide the command to the physical asset. It should be appreciated that, in other instances, the flow process for communicating the command received by the command and control component 916 to the physical asset may be different. Further, in some instances, changes to various device settings may be reflected within the viewer interface 710, the twin manager 704, and also within one or more metadata sources 406 (e.g., Metasys data 438), which may be linked together via one or more graph projections or other associations.
Referring now to
Referring now to
In some instances, the various alarm indicators 1612 may be selectable by the user to display corresponding alarm overlays 1614 that are overlaid onto the virtual building 902 proximate the selected alarm indicators 1612. As shown, the alarm overlays 1614 may include a device name (e.g., associated with a device having a fault), a fault description (e.g., a description of the fault), a fault duration (e.g., how long the fault has been occurring), and/or an error code (e.g., an identifiable code associated with the type of fault occurring with the device). In some instances, the alarm overlays 1614 may further include a link or button 1616 configured to allow for the user to have various assets associated with the alarm highlighted within the virtual building 902. In some instances, the various information and functionality provided via the alarm indicators 1612 and/or the alarm overlays 1614 may be customizable by the user. For example, in some instances, certain alarm indicators 1612 may be customizable by the user to be displayed in a variety of colors (e.g., red for high-priority alarms and green for low-priority alarms). In some instances, the alarm overlays 1614 may be customizable by the user to include varying levels or types of information, as desired for a given alarm, alarm type, alarm priority, etc.
Referring now to
Referring now to
In some instances, the heat map overlays 1806 provide a visual representation (e.g., different colors over an area) of a temperature distribution, an airflow or ventilation distribution, an indoor air quality distribution (e.g., CO2 levels, humidity, PM2.5 levels), a camera coverage distribution, an occupancy distribution, a lighting distribution, an energy usage distribution, an energy efficiency distribution, or any other pertinent type of distribution within the floor view of the virtual building, as desired for a given application and by fetching data from corresponding physical sensors within the physical building. For example, in some instances, high temperature areas may be overlaid with a red color and low temperature areas may be overlaid with a green or blue color. Between the high temperature areas and the low temperature areas may be a gradient color scheme indicating temperature drop off from the high-temperature area to the low-temperature area within the floor view, thereby creating the corresponding heat map overlay. In some instances, certain colors within the heat map may be indicative that a given sensor level is above or below an acceptable threshold (e.g., a temperature threshold, an air quality threshold, an energy consumption threshold). In some instances, this threshold may be set by a user via one or more options provided within the viewer interface 710. In some instances, the viewer interface 710 may allow the user to select the color scheme for a given heat map.
It will be appreciated that a variety of different types of heat map overlays may be utilized in a variety of configurations or color schemes to depict a variety of distribution types, as desired for a given application. In some instances, the heat maps shown may be selectively shown at various times throughout a given day, week, month, quarter, or year. For example, in some instances, the user may use a time slider on the heat map page 1802 to selectively view different heat maps (e.g., temperature, indoor air quality, occupancy, energy, etc.) overlaid onto a selected area representing various distributions at different times.
As an illustrative example, in some instances, a user may utilize a temperature or energy consumption heat map to identify various hot or cold areas within a given area. The user may then use the information gleaned from the heat map to make various layout, design, or device set point changes within the given area or throughout the building. Further, in some instances, the user may view a variety of heat maps pertaining to different distributions (e.g., utilizing various sensor and/or device data) to identify or correlate how various distributions interrelate (e.g., how a high temperature area may be correlated with a low energy efficiency area, how a lighting distribution may be correlated with an occupancy distribution, etc.).
In some instances, similar to the heat map overlay 1806, the viewer interface 710 may provide a lighting or camera coverage overlay configured to show a light or camera coverage distribution within a room, floor, or other area. For example, the viewer interface 710 may indicate a path of light clearance or camera visibility coming from a particular light or camera within a selected space. In these instances, the light or camera distribution may be viewed within a given area and the user may determine whether additional lights and/or cameras may be necessary.
Referring now to
In some instances, the various alarm indications 1906 may be filtered based on the floor that the alarms are associated with, a type of each of the alarms (e.g., power failure, an open door fault), a criticality of the alarms (e.g., high, medium, low), or any other relevant filtering criteria. Further, in some instances, upon clicking on a particular one of the alarm indications 1906, the user is allowed to obtain additional information regarding the alarm, such as the device(s) associated with the alarm, the type of alarm, time series data associated with the alarm (e.g., when the alarm began), the criticality of the alarm, or any other relevant data. In some instances, upon clicking on a particular one of the alarm indications 1906, the user is allowed to interact with the alarm (e.g., acknowledge a fault, mute a fault, trigger a standard operating procedure, etc.).
Referring now to
Referring now to
For example, in some instances, the viewer interface 710, when utilized within the enterprise manager user interface 2002, is configured to fetch twin data from the twin manager 704 via the beckon service 712. In some instances, the twin manager is in communication with a bridge component 2010 configured to allow for the user of the viewer interface 710 to perform control and command functions via interaction with the viewer interface 710, as discussed above with respect to
Once the appropriate information has been fetched by the viewer interface 710, the viewer interface 710 may then communicate the appropriate information with the rendering application 714 to create any of the various views and/or pages discussed above, with reference to
Referring now to
In some instances, the widgets 2102 include a building viewer widget including the viewer interface 710 described above. The building viewer widget including the viewer interface 710 is configured to allow the user of the enterprise manager page 2100 to select and view the virtual building 902 (or any other selected virtual building), and to perform any of the various functionality with respect to the virtual building 902 as discussed above, with reference to
Referring now to
As discussed above, the systems and methods described herein can be utilized to generate and present three-dimensional and/or two-dimensional renderings of virtual buildings or other virtual environments. In some embodiments, the three-dimensional and/or two-dimensional renderings may be enhanced or otherwise overlaid with various information pertaining to various assets within the virtual building or other virtual environment corresponding to physical assets within a corresponding physical building or other physical environment. Accordingly, it will be understood that the systems and methods described herein may be utilized in a variety of contexts to enhance or improve a user's understanding of and/or interaction with a virtual rendering of a corresponding physical environment. For example, in some instances, the systems and methods described herein may be utilized to generate a three-dimensional rendering of a building having or that will have various occupancy monitoring and/or guidance systems and devices and/or energy usage monitoring and/or guidance systems and devices installed therein.
In some instances, the systems described herein (e.g., any of systems 100, 300, 400, 500, 600, 700, 2000) may be configured to ingest (e.g., via beckon applications 712) various occupancy and/or energy usage information associated with physical assets within the building, correlate that information with various virtual assets within a graphical model, and provide different overlays, recommendations, etc. regarding various functions of the building.
For example, in some instances, the systems described herein may be configured to provide an occupancy heat map overlay (e.g., similar to the heat map overlays 1806) for overlaying onto a graphical model of the building (e.g., within the viewer interface 710). These occupancy heat map overlays may allow for a user to quickly discern where occupants are currently (or have historically been) located within the building. In some instances, a user interface (e.g., the viewer interface 710) may be configured to allow the user to selectively view the occupancy information for the plurality of spaces for different periods of time using a temporal slider.
In some instances, the systems described herein may be configured to generate one or more space usage recommendations. For example, in some instances, the one or more space usage recommendations may be generated based on occupancy information and/or energy usage information pertaining to various assets within the building. In some instances, a user interface (e.g., the viewer interface 710) may be configured to display the one or more space usage recommendations to a user, thereby allowing the user to act on the one or more space usage recommendations. In some instances, the user interface may further allow for the user to selectively control various assets and/or schedules associated with the building via the user interface. For example, in some instances, the user may selectively control, adjust, or otherwise modify a space usage schedule associated with one or more spaces within the building based on the one or more space usage recommendations. In some instances, the systems described herein may be configured to automatically control the various assets and/or schedules associated with the building (e.g., according to one or more predetermined rules set by a user of the system). For example, in some instances, the systems described herein may automatically alter or modify a space usage schedule based on the one or more generated space usage recommendations.
In some instances, the systems described herein may be configured to determine a location of one of an occupant or an asset based on ingested asset information or the occupancy information. For example, in some instances, an occupant's location may be determined using various occupancy information, such as, for example, access control device information indicating that the occupant has entered a specific area, a camera feed of a location within the building, whether the user is connected to a particular Wi-Fi or other network source, etc. In some instances, an asset's location may be determined using various asset location information. For example, in some instances, a mobile asset (e.g., an automatically navigated vacuum, a robotic task handler) may have a geolocation tag, a radio frequency identification (RFID) tag, or other identifying tag embedded within the mobile asset that may be sensed in various locations within the building. In some instances, when determining the location of an occupant or an asset within the building, the systems described herein may be configured to determine a more general location (e.g., on the second floor of the building) or a more specific location (e.g., two- dimensional or three-dimensional coordinates within the building). In some instances, determining a more specific location may include determining a height of the occupant or asset within the building (e.g., as measured from a ground level or with respect to a floor baseline level within the building).
In some instances, the systems described herein may be configured to generate and display (e.g., within the viewer interface 710) or otherwise represent a real-time flow of occupants into, out of, and/or within the building. In some instances, the real-time flow may be depicted or represented in two dimensions or three dimensions within the graphical model. In some instances, the systems described herein may be further configured to detect an emergency event occurring within the building (e.g., via the alarm application 168 or one or more beckon applications 712). In some instances, the systems described herein may further be configured to generate (e.g., via one of the AI agents 370) one or more emergency instructions based on a detected emergency event and the real-time flow of occupants into, out of, and within the building.
In some instances, the emergency instructions may comprise navigational directions for exiting the building that may be displayed to the user via a user interface (e.g., the viewer interface 710). For example, in the case of a fire, the navigational directions may be configured to navigate the user out of the building while avoiding a detected fire (e.g., using received temperature sensor date from various locations within the building). In some instances, the emergency instructions may comprise one or more shelter-in-place instructions, for example, in the case of an active shooter or otherwise dangerous person being detected within the building. In some instances, the emergency instructions may comprise one or more instructions for locating a source of an emergency event (e.g., to be utilized by one or more first responders). For example, in the event of a fire, the instructions may provide a location of the fire within the building to be used by a fire fighting team. In the case of an active shooter or otherwise dangerous individual entering the building, the instructions may provide a real-time location of the active shooter or otherwise dangerous individual (e.g., determined based on one or more camera feeds) to be used by law enforcement.
In some instances, in addition to or as an alternative to the navigational directions discussed above, the systems described herein may be configured to control (e.g., via the command processor 136, the building control 442, or various other control components) one or more auditory or visual devices (e.g., a public address (PA) system) within the building to provide one or more auditory or visual instructions to occupants based on the emergency event and/or the real-time flow of occupants into, out of, and within the building. For example, in some instances, based on the navigational directions and a detected occupancy within various areas within the building, the various auditory or visual devices within the building may provide location-specific auditory and/or visual instructions to occupants.
In some instances, the systems described herein may ingest (e.g., via beckon applications 712) occupancy information including one or more camera feeds of one or more areas within the building. In these cases, the systems described herein may determine (e.g., via one of the AI agents 370) whether one or more areas within the building are occupied based on the camera feed. For example, in some instances, one or more of the AI agents 370 may be configured to determine whether various areas within the building are occupied based on the camera feed using one or more model-based predictions. In some instances, the one or more model-based predictions are configured to identify one or more human-related objects (e.g., jackets, purses, backpacks, coffee mugs, etc.) arranged within a work space. If there are human-related objects within a given work space, the model-based predictions may be further configured to determine that the work space is occupied.
In some instances, the systems described herein may further be configured to identify (e.g., via one of the AI agents 370) one or more hidden areas within the building. The term “hidden areas” in this respect is used to refer to an area that is not visible within any camera feed within the building. For example, the AI agents 370 may be configured to identify the hidden areas within the building based on the one or more camera feeds and the graphical model. That is, the AI agents 370 may compare the various camera data received via the camera feeds to the graphical model to determine whether any portions of the graphical model (and thus the building) are not shown within a camera feed. In some instances, a user interface (e.g., the viewer interface 710) is configured to indicate the hidden areas to the user within the graphical model. Additionally, in some instances, the AI agents 370 may be configured to generate one or more camera installation suggestions based on the hidden areas where one or more cameras may be installed to cover the identified hidden areas. In some instances, these camera installation suggestions may be displayed to the user via a user interface (e.g., the viewer interface 710).
In some instances, the systems described herein may ingest (e.g., via beckon applications 712) energy usage information including light distribution information pertaining to or otherwise associated with one or more light generation devices within the building and ambient light information pertaining to or otherwise associated with an amount of ambient light entering one or more areas within the building. In some instances, the systems described herein may determine (e.g., via one of the AI agents 370) that one or more areas within the building have an overall lighting level above a lighting level threshold based on the light distribution information and the ambient light information. Accordingly, in these instances, the systems described herein may be configured to automatically adjust (e.g., via the command processor 136, the building control 442, or various other control components) a light level setting associated with the one or more light generation devices based on the overall lighting level being above the lighting level threshold.
In some instances, the systems described herein may ingest (e.g., via beckon applications 712) occupancy information including a counted number of media access control (MAC) addresses connected to a given Wi-Fi network or other network switch within a building, a counted number of access device activations for entering and exiting the building (or a given area within the building) over a given time period, and/or one or more camera feeds associated with one or more entrances or exits of the building. Accordingly, in some instances, one or more agents (e.g., AI agents 370) may be configured to calculate an occupancy within the building or within an area within the building based on any of these factors. In some instances, the aforementioned occupancy information may further be combined to provide a more accurate occupancy calculation. In some instances, the aforementioned occupancy information may further allow for the occupancy to be calculated without directly using an occupancy sensor (e.g., the occupancy may indirectly calculated).
In some instances, the systems described herein may further ingest (e.g., via beckon applications 712) various indoor air quality (IAQ) information from one or more air quality detection devices. In some instances, one or more agents (e.g., AI agents 370) may be configured to generate one or more building recommendations based on a combination of the calculated occupancy and the IAQ information. For example, in some instances, the AI agent 370 may determine that an occupied space has a poor air quality. In this instance, the AI agent 370 may generate a recommendation that some or all of the occupants move to an area having better air quality or better air circulation.
In some instances, the systems described herein may ingest (e.g., via beckon applications 712) various occupancy information during a post-construction phase of a building. In these instances, a user interface (e.g., the viewer interface 710) may be configured to depict one or more representations of how different spaces within the building are being utilized. For example, in some instances, an architect may use the representations of how different spaces within the building are being utilized (e.g., which spaces are being used as collaboration spaces, where queues or lines are forming for a sales booth, etc.) when considering how to redesign a given space (or how to design a future building). Further, in some instances, one or more agents (e.g., AI agents 370) may be configured to generate one or more building redesign recommendations based on the ingested occupancy information. In some instances, the one or more building redesign recommendations may be displayed to the user (e.g., via the viewer interface 710). For example, in some instances, the AI agent 370 may determine that a particular area is utilized more often than other areas within the building based on the occupancy information. Accordingly, the AI agent 370 may identifying one or more distinguishing factors pertaining to the utilized area (e.g., a shape of the room, an arrangement of furniture, a location of various air ducts) and generate the one or more building redesign recommendations as recommendations for making other areas more similar to the utilized area.
In some instances, the systems described herein may ingest (e.g., via beckon applications 712) occupancy information including historical occupancy information associated with one or more prior buildings that are similar to a new building that is in a pre-construction phase or a construction phase. In these instances, one or more agents (e.g., AI agents 370) may generate predicted occupancy information for the new building based on the historical occupancy information. In these instances, the one or more agents may further generate one or more design recommendations during the pre-construction or construction phase of the new building based on the predicted occupancy information. In some instances, the predicted occupancy information and/or the one or more design recommendations may be presented visually to a user (e.g., an architect) via a user interface (e.g.,, via the viewer interface 710). For example, in some instances, the user may be presented with one or more simulated flows of occupants into, out of, and/or within the new building, which may be overlaid onto the graphical model of the new building within the viewer interface.
As discussed above, in some instances, the systems described herein may ingest (e.g., via beckon applications 712) building information modeling (BIM) data for a BIM model from a digital twin (e.g., via the twin manager 704) and generate the graphical model based on the BIM data (e.g., utilizing the rendering application 714). In some instances, one or more of the beckon applications 712 may be configured to detect a modification to the BIM data for the BIM model within either the graphical model (e.g., received via the viewer interface 710) or the digital twin (e.g., received via the twin manager 704). In these instances, the one or more beckon applications 712 may be configured to, based on this detected modification to the BIM data, automatically update the BIM data for the BIM model within the other of the graphical model (e.g., via communication with the rendering application 714) or the digital twin (e.g., via communication with the twin manager 704). For example, in some instances, the modification to the BIM data may comprise adding user notes pertaining to one or more assets within the building, adding device servicing information pertaining to one or more devices within the building, adding one or more virtual assets corresponding to one or more physical assets that have been added to the building, and/or any other potential BIM data modification.
In some instances, the one or more beckon applications 712 may be configured to automatically update the BIM data within the other of the graphical model or the digital twin in real-time or nearly real-time. In some instances, the one or more beckon applications 712 may further be configured to automatically update BIM data within both the graphical model and the digital twin based on various information ingested from various external computing systems. For example, in some instances, one or more external control and/or tracking systems may be configured to receive and automatically associate various information with various assets within the building and to automatically push BIM data modifications (e.g., via the one or more beckon applications 712) to both of the graphical model and the digital twin. Further, in some instances, the one or more beckon applications 712 may be configured to push information associated with any BIM data modifications to associated external systems associated with assets affected by the BIM data modification(s).
In some instances, the systems described herein may further ingest (e.g., via beckon applications 712) fault detection information, temperature information, smoke detection information, viral detection information, particulate detection information, and/or airflow information. In some instances, one or more agents (e.g., AI agents 370) may be configured to predict a future fault associated with an area within the building based on fault detection information for another area within the building and any of the temperature information, the smoke detection information, the viral detection information, the particulate detection information, and/or the airflow information. For example, in some instances, the AI agents 370 may determine that a fire, smoke, a virally contaminated air pocket, and/or an air pocket with poor air quality (e.g., a high particulate count) is moving or is likely to move within the building (e.g., based on the airflow information and/or other sensor data), which may then cause the predicted future fault. In some instances, a user interface (e.g., the viewer interface 710) may display a corresponding future fault indicator representing the predicted future fault within or overlaid onto the graphical model.
In some instances, the systems described herein may generate overlay representations of flow paths associated with any of occupants, temperature changes, smoke, airborne viral droplets, airborne particulate, and/or airflow within the building. For example, in some instances, the overlay representations may be generated by the rendering application 714 based on one or more of the occupancy information, the temperature information, the smoke detection information, the viral detection information, the particulate detection information, and/or the airflow information. In these instances, a user interface (e.g., the viewer interface 710) may include any of the overlay representations displayed on top of or within the graphical model. In some instances, the overlay representations may be representative of current building conditions. In some instances, the overlay representations may be representative of predicted future building conditions. For example, in some instances, one or more agents (e.g., AI agents 370) may be configured to predict future flow paths associated with any of the occupants, the temperature changes, the smoke, the airborne viral droplets, the airborne particulate, and/or the airflow within the building, which may then be displayed to the user view the viewer interface 710.
In some instances, one or more agents (e.g., AI agents 370) may be configured to determine, based on the flow path of the airflow within the building, one or more installation locations of one or more smoke detectors, viral detectors, particular detectors, or indoor air quality sensors. For example, in some instances, if an area has a higher average airflow associated with it, it may be more suitable for detecting smoke, viral loads, particular, and/or air quality generally within the building than another area having a lower average airflow. As such, the AI agents 370 may determine the installation locations based on areas having a high level of airflow. In some instances, the user interface (e.g., the viewer interface 710) may display indications to the user corresponding to the one or more installation locations within the graphical model.
In some instances, one or more agents (e.g., AI agents 370) may be configured to recommend one or more actions to be taken within one or more first areas within the building based on one or more of fault detection information for a second area in the building, temperature information for the second area in the building, smoke detection information for the second area in the building, viral detection information for the second area in the building, particulate detection information for the second area in the building, and/or airflow information for the second area in the building. For example, if a particular area within the building has a fault indicating that the temperature is too high, and an adjacent area within the building has a lower temperature, but is closed off to the first area, the AI agent 370 may recommend opening one or more doors between the first area and the second area or routing a portion of any conditioned air from the second area to the first area. In some instances, the AI agents may consider occupancy information when recommending the one or more actions as well. For example, in some instances, if an area has a high occupancy and is adjacent to an area with a low occupancy, the AI agent 370 may recommend that various occupants spread to the area with the low occupancy. In some of these instances, the AI agent 370 may further control (e.g., via the command processor 136, the building control 442, or various other control components) various devices within the low occupancy area in preparation for the occupants moving there based on the recommended actions.
In some instances, the systems described herein may ingest (e.g., via beckon applications 712) both occupancy information and energy usage information. In some instances, one or more agents (e.g., the AI agent 370) may determine, based on the occupancy information and the energy usage information, that one or more spaces (e.g., work spaces) are consuming energy, but lack occupants. For example, this may happen if an energy-consuming device is accidentally left plugged in at a workstation by an employee. In these instances, the user may be presented (e.g., via the viewer interface 710) with an indication of the one or more spaces consuming energy that lack occupants within the graphical model. For example, in some instances, these spaces may be highlighted or otherwise color-coded for quick identification by the user. In some instances, this determination and indication of the one or more spaces within the graphical model may be performed in real-time or nearly real-time based on telemetry data received from various devices within the building.
Further, in some instances, the user may control (e.g., via the viewer interface 710) one or more energy-consuming devices within the identified spaces to reduce the unnecessary energy consumption. For example, if a particular device has been left plugged in or otherwise on and consuming energy at an identified workspace, the user may be able to identify which device or devices are still consuming energy and to specifically turn those devices off within the viewer interface 710. In some instances, turning the devices off may comprise cutting off power to a particular power outlet at the workstation. In some instances, the user may further be provided (e.g., via the viewer interface) with a real-time energy usage metric (e.g., calculated by one of the AI agents 370). For example, the real-time energy usage metric may provide the user with a metric (e.g., a score) for how well the energy consumed within the building is being utilized by occupants. As an example, if several occupants are in the only room within a building consuming energy, this scenario may receive a “great” score (e.g., “100/100”). Conversely, as another example, if the building has zero occupants and is continuing to consume a high amount of energy, this scenario may receive a “poor” score (e.g., “0/100”). In some instances, various rules may be set by authorized users (e.g., building sustainability managers) to calculate the real-time energy usage metric, as desired for a given application.
In some instances, the one or more agents (e.g., the AI agent 370) may detect one or more faults within the building based on ingested information. For example, in some instances, the faults may pertain to an excess water consumption, a high zone temperature, excess energy consumption, etc. In some instances, the one or more agents (e.g., the AI agent) may further determine a view selection rule based on the detected fault and automatically navigate the user to an intelligent fault visualization (e.g., within the graphical model via the viewer interface 710) of an area associated with the fault based on the view selection rule.
In some instances, in the case of a high zone temperature occurring within an area associated with the fault, the intelligent fault visualization displayed to the user (e.g., via the viewer interface 710) may isolate various air-side equipment associated with the fault such that the other equipment is hidden from view. In some other instances, the intelligent fault visualization may isolate a floor on which the fault is occurring, isolate various other types of equipment associated with the fault, highlight equipment served by or serving a piece of equipment on which the fault is occurring, include a heat map overlay (e.g., heat map overlay 1806) indicating one or more sensor readings, or include any other indications or views, as desired for a given fault.
In some instances, one or more view selection rules may be based on a location of a user device within the building. For example, if a fault pertains to a high energy consumption, instead of showing the equipment affected by the fault, if the system (e.g., one or the AI agents 370) determines that a user device on which the user is accessing the graphical model is within the building, the user may be presented (e.g., via the viewer interface 710) with directions to the equipment affected by the fault. Similarly, in some instances, one or more view selection rules may be determined based on a role of the user. For example, the user may be any of a manager, a service technician, an emergency responder, etc. Accordingly, in some instances, the view selection rules may depend on the user accessing the viewer interface 710. For example, in some instances, a manager may be provided with a view showing an overall building effect. In some instances, a service technician may be provided with a view showing a specific area within the building and highlighting the affected equipment to allow for the service technician to quickly identify and address underlying issues causing the fault. In some instances, an emergency responder may be provided with directions to a given area where the fault is occurring (e.g., an area where a smoke detected has detected a likely fire).
In some instances, the systems described herein may ingest (e.g., via the beckon applications 712) environmental information. For example, in some instances, the environmental information may include temperature information, humidity information, particulate information (e.g., carbon particulate information), or any other suitable environmental information. As such, in some instances, the user may be presented (e.g., within the graphical model via the viewer interface 710) with a representation of an environmental status of an area within the building. In some instances, the representation of the environmental status may be based on historical environmental information. In these instances, the representation may be a temporal representation of the environmental status of an area within the building over a given time period. For example, in some instances, the temporal representation may be provided via an animated overlay that is overlaid onto the graphical model (e.g., accessible via a “play button” within the viewer interface 710).
In some instances, one or more agents (e.g., AI agents 370) may predict, based on the environmental information, future environmental information. For example, in some instances, an AI agent 370 may utilize one or more prediction models to predict various future environmental information. Accordingly, in these instances, the time period of the temporal representation discussed above may comprises a future period of time based on the predicted future environmental information. Further, in some instances, the one or more agents (e.g., AI agents 370) may determine one or more recommended actions based on the predicted future environmental information and present an indication (e.g., via a pop-up window) of the one or more recommended actions to the user (e.g., via the viewer interface 710).
In some instances, one or more agents (e.g., AI agents 370) may predict, based on the environmental information, an environmental impact for a given space redesign. For example, in some instances, an AI agent 370 may be able to predict future environmental information for a current space design and simulated future environmental information for one or more space redesigns. For example, a “space redesign” may include a rearrangement of one or more pieces of equipment or other assets, additional of one or more pieces of equipment or other assets, and/or removal of one or more pieces of equipment or other assets. In some instances, the user may additionally be able to view a temporal representation of the simulated future environmental information (e.g., via a similar “play button” within the viewer interface 710) for various simulated space redesigns. Further, in some instances, an AI agent 370 may predict, based on a comparison between the simulated future environmental information and the predicted future environmental information for the current space design, various environmental impact information for a given space redesign. In some instances, one or more indications of the environmental impact information may be presented to the user (e.g., via the viewer interface 710).
In some instances, the predicted future environmental information predicted by the one or more agents (e.g., AI agents 370) may include predicted a travel path of smoke or fire within the building. In some instances, the user may be presented (e.g., via the viewer interface 710) with an animated overlay depicting the predicted travel path of the smoke or fire within the building overlaid onto the graphical model. In some instances, the one or more agents (e.g., AI agents 370) may further identify one or more occupied areas (e.g., based on ingested occupancy information) within the predicted travel path. In these instances, the one or more agents may further identify an exit path for occupants within the occupied area to exit the building based on the predicted travel path. For example, the exit path may be a path within the building that avoids or otherwise circumnavigates the predicted travel path of the smoke or fire. In some instances, the user may be presented (e.g., via the viewer interface 710) with an indication of the exit path overlaid onto the graphical model. In some instances, the predicted travel path and/or the exit path may be updated in real-time or nearly real-time based on additional ingested information.
The construction and arrangement of the systems and methods as shown in the various exemplary embodiments are illustrative only. Although only a few embodiments have been described in detail in this disclosure, many modifications are possible (e.g., variations in sizes, dimensions, structures, shapes and proportions of the various elements, values of parameters, mounting arrangements, use of materials, colors, orientations, etc.). For example, the position of elements may be reversed or otherwise varied and the nature or number of discrete elements or positions may be altered or varied. Accordingly, all such modifications are intended to be included within the scope of the present disclosure. The order or sequence of any process or method steps may be varied or re-sequenced according to alternative embodiments. Other substitutions, modifications, changes, and omissions may be made in the design, operating conditions and arrangement of the exemplary embodiments without departing from the scope of the present disclosure.
The present disclosure contemplates methods, systems and program products on any machine-readable media for accomplishing various operations. The embodiments of the present disclosure may be implemented using existing computer processors, or by a special purpose computer processor for an appropriate system, incorporated for this or another purpose, or by a hardwired system. Embodiments within the scope of the present disclosure include program products comprising machine-readable media for carrying or having machine-executable instructions or data structures stored thereon. Such machine-readable media can be any available media that can be accessed by a general purpose or special purpose computer or other machine with a processor. By way of example, such machine-readable media can comprise RAM, ROM, EPROM, EEPROM, CD-ROM or other optical disk storage, magnetic disk storage or other magnetic storage devices, or any other medium which can be used to carry or store desired program code in the form of machine-executable instructions or data structures and which can be accessed by a general purpose or special purpose computer or other machine with a processor. When information is transferred or provided over a network or another communications connection (either hardwired, wireless, or a combination of hardwired or wireless) to a machine, the machine properly views the connection as a machine-readable medium. Thus, any such connection is properly termed a machine-readable medium. Combinations of the above are also included within the scope of machine-readable media. Machine-executable instructions include, for example, instructions and data which cause a general purpose computer, special purpose computer, or special purpose processing machines to perform a certain function or group of functions.
Although the figures show a specific order of method steps, the order of the steps may differ from what is depicted. Also two or more steps may be performed concurrently or with partial concurrence. Such variation will depend on the software and hardware systems chosen and on designer choice. All such variations are within the scope of the disclosure. Likewise, software implementations could be accomplished with standard programming techniques with rule based logic and other logic to accomplish the various connection steps, processing steps, comparison steps and decision steps.
In various implementations, the steps and operations described herein may be performed on one processor or in a combination of two or more processors. For example, in some implementations, the various operations could be performed in a central server or set of central servers configured to receive data from one or more devices (e.g., edge computing devices/controllers) and perform the operations. In some implementations, the operations may be performed by one or more local controllers or computing devices (e.g., edge devices), such as controllers dedicated to and/or located within a particular building or portion of a building. In some implementations, the operations may be performed by a combination of one or more central or offsite computing devices/servers and one or more local controllers/computing devices. All such implementations are contemplated within the scope of the present disclosure. Further, unless otherwise indicated, when the present disclosure refers to one or more computer-readable storage media and/or one or more controllers, such computer-readable storage media and/or one or more controllers may be implemented as one or more central servers, one or more local controllers or computing devices (e.g., edge devices), any combination thereof, or any other combination of storage media and/or controllers regardless of the location of such devices.
This application claims the benefit of and priority to U.S. Provisional Patent Application No. 63/416,884, filed Oct. 17, 2022, which is incorporated herein by reference in its entirety.
Number | Date | Country | |
---|---|---|---|
63416884 | Oct 2022 | US |