This application relates generally to a building system of a building. This application relates more particularly to systems for managing and processing data of the building system.
A building may aggregate and store building data received from building equipment and/or other data sources. The building data can be stored in a database. The building can include a building system that operates analytic and/or control algorithms against the data of the database to control the building equipment. However, the development and/or deployment of the analytic and/or control algorithms may be time consuming and require a significant amount of software development. Furthermore, the analytic and/or control algorithms may lack flexibility to adapt to changing circumstances in the building. In some cases, the output data of the analytic and/or control algorithms may be hard for a user to conceptualize and relate to the physical components of the building for which the information is generated for.
One implementation of the present disclosure is a building system including one or more storage devices storing instructions thereon that, when executed by one or more processors, cause the one or more processors to receive an indication to execute an artificial intelligence (AI) agent. The instructions cause the one or more processors to execute the AI agent based on a digital twin to generate an inference or a prediction, the digital twin including representations of entities of a building and relationships between the entities of the building, cause a graphical model of the building including graphical representations of the entities to include information based on the inference or the prediction, and cause a display device of a user device to display the graphical model.
In some embodiments, the graphical model of the building is a three dimensional building model of the building.
In some embodiments, the graphical model is a building information model (BIM).
In some embodiments, the building are at least one of a space, a device, or a person.
In some embodiments, the digital twin includes a graph data structure including a nodes representing the entities of a building and a edges between the nodes representing the relationships between the entities of the building. In some embodiments, the inference or the prediction is for an entity of the entities, wherein a first node of the nodes represents the entity. In some embodiments, the instructions cause the one or more processors to identify one or more second nodes storing or identifying data for the entity, the one or more second nodes related to the first node by one or more edges of the edges, retrieve the data for the entity from the one or more second nodes or based on the one or more second nodes, and execute the AI agent based on the data for the entity to generate the inference or the prediction.
In some embodiments, the digital twin includes a graph data structure including a nodes representing the entities of a building and a edges between the nodes representing the relationships between the entities of the building. In some embodiments, the inference or the prediction is for an entity of the entities, wherein a first node of the nodes represents the entity. In some embodiments, the instructions cause the one or more processors to identify a second node of the nodes for storing the inference or the prediction by identifying one or more edges of the edges between the first node and the second node and cause the second node of the nodes to store the inference or the prediction.
In some embodiments, the instructions cause the one or more processors to receive an input from the user device in the graphical model and execute the AI agent based on the digital twin to generate the inference or the prediction in response to receiving the input.
In some embodiments, the instructions cause the one or more processors to receive an interaction with a graphical representation of an entity of the graphical representations of the entities from the user device in the graphical model, cause the graphical model of the building to include one or more inferences or one or more predictions associated with the entity, and cause the display device of the user device to display the graphical model.
In some embodiments, the inference or the prediction is for an entity of the entities. In some embodiments, a graphical representation of the entity is included within the graphical model of the building. In some embodiments, the instructions cause the one or more processors to cause the graphical model of the building including the graphical representations of the entities to animate the graphical representation of the entity based on the inference or the prediction.
In some embodiments, the AI agent is at least one of a clean air optimization (CAO) agent configured to generate one or more operating parameters for an air handler unit of the building or an energy prediction model AI agent configured to predict an energy usage of a component of the building.
In some embodiments, the inference is at least one inferred operating value that causes one or more pieces of equipment of the building to achieve an operating goal. In some embodiments, the prediction is at least one predicted value of a parameter at a future time.
In some embodiments, at least one of the inference or the prediction are a timeseries of time correlated data values.
In some embodiments, the instructions cause the one or more processors to cause the one or more pieces of equipment of the building to operate based on the inferred operating value to achieve the operating goal.
Another implementation of the present disclosure is a method including receiving, by one or more processing circuits, an indication to execute an artificial intelligence (AI) agent and executing, by the one or more processing circuits, the AI agent based on the digital twin to generate an inference or a prediction, the digital twin including representations of entities of a building and relationships between the entities of the building. The method includes causing, by the one or more processing circuits, a graphical model of the building including graphical representations of the entities to include information based on the inference or the prediction and causing, by the one or more processing circuits, a display device of a user device to display the graphical model.
In some embodiments, the graphical model of the building is a three dimensional building model of the building.
In some embodiments, wherein the digital twin includes a graph data structure including a plurality of nodes representing the entities of a building and a plurality of edges between the plurality of nodes representing the relationships between the entities of the building. In some embodiments, the inference or the prediction is for an entity of the entities, wherein a first node of the nodes represents the entity. In some embodiments, the method further includes identifying, by the one or more processing circuits, one or more second nodes storing or identifying data for the entity, the one or more second nodes related to the first node by one or more edges of the edges, retrieving, by the one or more processing circuits, the data for the entity from the one or more second nodes or based on the one or more second nodes, and executing, by the one or more processing circuits, the AI agent based on the data for the entity to generate the inference or the prediction.
In some embodiments, the digital twin includes a graph data structure including nodes representing the entities of a building and edges between the nodes representing the relationships between the entities of the building. In some embodiments, the inference or the prediction is for an entity of the entities, wherein a first node of the nodes represents the entity. In some embodiments, the method further includes identifying, by the one or more processing circuits, a second node of the nodes for storing the inference or the prediction by identifying one or more edges of the edges between the first node and the second node and causing, by the one or more processing circuits, the second node of the nodes to store the inference or the prediction.
In some embodiments, the method further includes receiving, by the one or more processing circuits, an interaction with a graphical representation of an entity of the graphical representations of the entities from the user device in the graphical model, causing, by the one or more processing circuits, the graphical model of the building to include one or more inferences or one or more predictions associated with the entity, and causing, by the one or more processing circuits, the display device of the user device to display the graphical model.
In some embodiments, the inference or the prediction is for an entity of the entities. In some embodiments, a graphical representation of the entity is included within the graphical model of the building. In some embodiments, the instructions cause the one or more processors to cause the graphical model of the building including the graphical representations of the entities to animate the graphical representation of the entity based on the inference or the prediction.
Another implementation of the present disclosure is a building system including one or more processing circuits configured to receive an indication to execute an artificial intelligence (AI) agent. The one or more processing circuits are configured to execute the AI agent a digital twin to generate an inference or a prediction, the digital twin including representations of entities of a building and relationships between the entities of the building, cause a graphical model of the building including graphical representations of the entities to include information based on the inference or the prediction, and cause a display device of a user device to display the graphical model.
In some embodiments, the instructions cause the one or more processors to generate an action based on the inference or the prediction, wherein the action is at least one of a setting adjustment for a piece of equipment or a ticket generation for a work order to perform maintenance on the piece of equipment, receive input via the display device from the user device to perform the action, and perform the action responsive to receiving the input.
Various objects, aspects, features, and advantages of the disclosure will become more apparent and better understood by referring to the detailed description taken in conjunction with the accompanying drawings, in which like reference characters identify corresponding elements throughout. In the drawings, like reference numbers generally indicate identical, functionally similar, and/or structurally similar elements.
Overview
Referring generally to the FIGURES, systems and methods for digital twins of a building that integrate inferences and predictions with three dimensional building models are shown, according to various exemplary embodiments. A digital twin can be a virtual representation of a building and/or an entity of the building (e.g., space, piece of equipment, occupant, etc.). Furthermore, the digital twin can represent a service performed in a building, e.g., facility management, clean air optimization, energy prediction, equipment maintenance, etc.
In some embodiments, the digital twin can include an information data store and a connector. The information data store can store the information describing the entity that the digital twin operates for (e.g., attributes of the entity, measurements associated with the entity, control points or commands of the entity, etc.). In some embodiments, the data store can be a graph including various nodes and edges. The connector can be a software component the provides telemetry from the entity (e.g., physical device) to the information store. Furthermore, the digital twin can include artificial intelligence (AI), e.g., an AI agent. The AI can be one or more machine learning algorithms and/or models that operate based on information of the information data store and outputs information. The AI agent can run against a common data model, e.g., BRICK, and can be easily implemented in various different buildings, e.g., against various different building models. Running against BRICK can allow for the AI agent to be plug-and-play and reduce AI design and/or deployment time.
In some embodiments, the AI agent for the digital twin can call an AI service to determine inferences and/or predict future data values. In some embodiments, the predictions are potential future states. In some embodiments, the predictions predict a timeseries of a data point into the future. The predictions could be predicted indoor temperature for an hour, inferred future air quality from 15 minute air quality readings, etc. In some embodiments, the digital twin can store predicted and/or inferred information in a graph data store as a node in the graph data store related to an entity that the digital twin represents or otherwise operates for. In some embodiments, the digital twin, or other digital twins, can operate against the predicted and/or inferred data, e.g., operate to construct and implement control algorithms for operating equipment of a building based on predicted future data points of the building.
Furthermore, the AI agent can include various triggers and/or actions, conditions that define when and how command and control occurs for an entity. The triggers and actions can be rule based conditional and operational statements that are associated with a specific digital twin, e.g., are stored and executed by an AI agent of the digital twin. In some embodiments, the building system can identify actions and/or triggers (or parameters for the actions and/or triggers) through machine learning algorithms. In some embodiments, the building system can evaluate the conditions/context of the graph and determine and/or modify the triggers and actions of a digital twin.
In some embodiments, the building system can implement and manage a three dimensional (or two dimensional) building model. The building system can cause the model to render in a user interface of a user device and allow a user to view the model, view information associated with the components of the model, and/or navigate throughout the model. The model can include information for various physical components, e.g., data for VAVs, data for thermostats, data for sensors, data for spaces, etc. In some embodiments, the information for the various physical components is generated based on a digital twin for those physical components, e.g., inferences or predictions generated for the physical components by one or more AI agents. The inferred and/or predicted information of the digital twin can be presented to the user in the three dimensional model.
Furthermore, in some embodiments, a user can provide commands and/or inputs into the digital twin via the three dimensional model. Various buttons and/or other input elements can be provided within the three dimensional model that a user can interact with via their user device. The inputs can be ingested into a graph of the digital twin and various agents of the digital twin can operate based on the ingested input. For example, a command or analysis request for data could be made via the three dimensional model. The command or analysis request could cause an agent to run to operate the building equipment based on the command and/or derive data for presentation in the three dimensional model.
Referring now to
The building data platform 100 includes applications 110. The applications 110 can be various applications that operate to manage the building subsystems 122. The applications 110 can be remote or on-premises applications (or a hybrid of both) that run on various computing systems. The applications 110 can include an alarm application 168 configured to manage alarms for the building subsystems 122. The applications 110 include an assurance application 170 that implements assurance services for the building subsystems 122. In some embodiments, the applications 110 include an energy application 172 configured to manage the energy usage of the building subsystems 122. The applications 110 include a security application 174 configured to manage security systems of the building.
In some embodiments, the applications 110 and/or the cloud platform 106 interacts with a user device 176. In some embodiments, a component or an entire application of the applications 110 runs on the user device 176. The user device 176 may be a laptop computer, a desktop computer, a smartphone, a tablet, and/or any other device with an input interface (e.g., touch screen, mouse, keyboard, etc.) and an output interface (e.g., a speaker, a display, etc.).
The applications 110, the twin manager 108, the cloud platform 106, and the edge platform 102 can be implemented on one or more computing systems, e.g., on processors and/or memory devices. For example, the edge platform 102 includes processor(s) 118 and memories 120, the cloud platform 106 includes processor(s) 124 and memories 126, the applications 110 include processor(s) 164 and memories 166, and the twin manager 108 includes processor(s) 148 and memories 150.
The processors can be a general purpose or specific purpose processors, an application specific integrated circuit (ASIC), one or more field programmable gate arrays (FPGAs), a group of processing components, or other suitable processing components. The processors may be configured to execute computer code and/or instructions stored in the memories or received from other computer readable media (e.g., CDROM, network storage, a remote server, etc.).
The memories can include one or more devices (e.g., memory units, memory devices, storage devices, etc.) for storing data and/or computer code for completing and/or facilitating the various processes described in the present disclosure. The memories can include random access memory (RAM), read-only memory (ROM), hard drive storage, temporary storage, non-volatile memory, flash memory, optical memory, or any other suitable memory for storing software objects and/or computer instructions. The memories can include database components, object code components, script components, or any other type of information structure for supporting the various activities and information structures described in the present disclosure. The memories can be communicably connected to the processors and can include computer code for executing (e.g., by the processors) one or more processes described herein.
The edge platform 102 can be configured to provide connection to the building subsystems 122. The edge platform 102 can receive messages from the building subsystems 122 and/or deliver messages to the building subsystems 122. The edge platform 102 includes one or multiple gateways, e.g., the gateways 112-116. The gateways 112-116 can act as a gateway between the cloud platform 106 and the building subsystems 122. The gateways 112-116 can be the gateways described in U.S. Provisional Patent Application No. 62/951,897 filed Dec. 20, 2019, the entirety of which is incorporated by reference herein. In some embodiments, the applications 110 can be deployed on the edge platform 102. In this regard, lower latency in management of the building subsystems 122 can be realized.
The edge platform 102 can be connected to the cloud platform 106 via a network 104. The network 104 can communicatively couple the devices and systems of building data platform 100. In some embodiments, the network 104 is at least one of and/or a combination of a Wi-Fi network, a wired Ethernet network, a ZigBee network, a Bluetooth network, and/or any other wireless network. The network 104 may be a local area network or a wide area network (e.g., the Internet, a building WAN, etc.) and may use a variety of communications protocols (e.g., BACnet, IP, LON, etc.). The network 104 may include routers, modems, servers, cell towers, satellites, and/or network switches. The network 104 may be a combination of wired and wireless networks.
The cloud platform 106 can be configured to facilitate communication and routing of messages between the applications 110, the twin manager 108, the edge platform 102, and/or any other system. The cloud platform 106 can include a platform manager 128, a messaging manager 140, a command processor 136, and an enrichment manager 138. In some embodiments, the cloud platform 106 can facilitate messaging between the building data platform 100 via the network 104.
The messaging manager 140 can be configured to operate as a transport service that controls communication with the building subsystems 122 and/or any other system, e.g., managing commands to devices (C2D), commands to connectors (C2C) for external systems, commands from the device to the cloud (D2C), and/or notifications. The messaging manager 140 can receive different types of data from the applications 110, the twin manager 108, and/or the edge platform 102. The messaging manager 140 can receive change on value data 142, e.g., data that indicates that a value of a point has changed. The messaging manager 140 can receive timeseries data 144, e.g., a time correlated series of data entries each associated with a particular time stamp. Furthermore, the messaging manager 140 can receive command data 146. All of the messages handled by the cloud platform 106 can be handled as an event, e.g., the data 142-146 can each be packaged as an event with a data value occurring at a particular time (e.g., a temperature measurement made at a particular time).
The cloud platform 106 includes a command processor 136. The command processor 136 can be configured to receive commands to perform an action from the applications 110, the building subsystems 122, the user device 176, etc. The command processor 136 can manage the commands, determine whether the commanding system is authorized to perform the particular commands, and communicate the commands to the commanded system, e.g., the building subsystems 122 and/or the applications 110. The commands could be a command to change an operational setting that control environmental conditions of a building, a command to run analytics, etc.
The cloud platform 106 includes an enrichment manager 138. The enrichment manager 138 can be configured to enrich the events received by the messaging manager 140. The enrichment manager 138 can be configured to add contextual information to the events. The enrichment manager 138 can communicate with the twin manager 108 to retrieve the contextual information. In some embodiments, the contextual information is an indication of information related to the event. For example, if the event is a timeseries temperature measurement of a thermostat, contextual information such as the location of the thermostat (e.g., what room), the equipment controlled by the thermostat (e.g., what VAV), etc. can be added to the event. In this regard, when a consuming application, e.g., one of the applications 110 receives the event, the consuming application can operate based on the data of the event, the temperature measurement, and also the contextual information of the event.
The enrichment manager 138 can solve a problem that when a device produces a significant amount of information, the information may contain simple data without context. An example might include the data generated when a user scans a badge at a badge scanner of the building subsystems 122. This physical event can generate an output event including such information as “DeviceBadgeScannerID,” “BadgeID,” and/or “Date/Time.” However, if a system sends this data to a consuming application, e.g., Consumer A and a Consumer B, each customer may need to call the building data platform knowledge service to query information with queries such as, “What space, build, floor is that badge scanner in?” or “What user is associated with that badge?”
By performing enrichment on the data feed, a system can be able to perform inferences on the data. A result of the enrichment may be transformation of the message “DeviceBadgeScannerId, BadgeId, Date/Time,” to “Region, Building, Floor, Asset, DeviceId, BadgeId, UserName, EmployeeId, Date/Time Scanned.” This can be a significant optimization, as a system can reduce the number of calls by 1/n, where n is the number of consumers of this data feed.
By using this enrichment, a system can also have the ability to filter out undesired events. If there are 100 building in a campus that receive 100,000 events per building each hour, but only 1 building is actually commissioned, only 1/10 of the events are enriched. By looking at what events are enriched and what events are not enriched, a system can do traffic shaping of forwarding of these events to reduce the cost of forwarding events that no consuming application wants or reads.
An example of an event received by the enrichment manager 138 may be:
An example of an enriched event generated by the enrichment manager 138 may be:
By receiving enriched events, an application of the applications 110 can be able to populate and/or filter what events are associated with what areas. Furthermore, user interface generating applications can generate user interfaces that include the contextual information based on the enriched events.
The cloud platform 106 includes a platform manager 128. The platform manager 128 can be configured to manage the users and/or subscriptions of the cloud platform 106. For example, what subscribing building, user, and/or tenant utilizes the cloud platform 106. The platform manager 128 includes a provisioning service 130 configured to provision the cloud platform 106, the edge platform 102, and the twin manager 108. The platform manager 128 includes a subscription service 132 configured to manage a subscription of the building, user, and/or tenant while the entitlement service 134 can track entitlements of the buildings, users, and/or tenants.
The twin manager 108 can be configured to manage and maintain a digital twin. The digital twin can be a digital representation of the physical environment, e.g., a building. The twin manager 108 can include a change feed generator 152, a schema and ontology 154, a projection manager 156, a policy manager 158, an entity, relationship, and event database 160, and a graph projection database 162.
The graph projection manager 156 can be configured to construct graph projections and store the graph projections in the graph projection database 162. Examples of graph projections are shown in
In some embodiment, the graph projection manager 156 generates a graph projection for a particular user, application, subscription, and/or system. In this regard, the graph projection can be generated based on policies for the particular user, application, and/or system in addition to an ontology specific for that user, application, and/or system. In this regard, an entity could request a graph projection and the graph projection manager 156 can be configured to generate the graph projection for the entity based on policies and an ontology specific to the entity. The policies can indicate what entities, relationships, and/or events the entity has access to. The ontology can indicate what types of relationships between entities the requesting entity expects to see, e.g., floors within a building, devices within a floor, etc. Another requesting entity may have an ontology to see devices within a building and applications for the devices within the graph.
The graph projections generated by the graph projection manager 156 and stored in the graph projection database 162 can be a knowledge graph and is an integration point. For example, the graph projections can represent floor plans and systems associated with each floor. Furthermore, the graph projections can include events, e.g., telemetry data of the building subsystems 122. The graph projections can show application services as nodes and API calls between the services as edges in the graph. The graph projections can illustrate the capabilities of spaces, users, and/or devices. The graph projections can include indications of the building subsystems 122, e.g., thermostats, cameras, VAVs, etc. The graph projection database 162 can store graph projections that keep up a current state of a building.
The graph projections of the graph projection database 162 can be digital twins of a building. Digital twins can be digital replicas of physical entities that enable an in-depth analysis of data of the physical entities and provide the potential to monitor systems to mitigate risks, manage issues, and utilize simulations to test future solutions. Digital twins can play an important role in helping technicians find the root cause of issues and solve problems faster, in supporting safety and security protocols, and in supporting building managers in more efficient use of energy and other facilities resources. Digital twins can be used to enable and unify security systems, employee experience, facilities management, sustainability, etc.
In some embodiments the enrichment manager 138 can use a graph projection of the graph projection database 162 to enrich events. In some embodiments, the enrichment manager 138 can identify nodes and relationships that are associated with, and are pertinent to, the device that generated the event. For example, the enrichment manager 138 could identify a thermostat generating a temperature measurement event within the graph. The enrichment manager 138 can identify relationships between the thermostat and spaces, e.g., a zone that the thermostat is located in. The enrichment manager 138 can add an indication of the zone to the event.
Furthermore, the command processor 136 can be configured to utilize the graph projections to command the building subsystems 122. The command processor 136 can identify a policy for a commanding entity within the graph projection to determine whether the commanding entity has the ability to make the command. For example, the command processor 136, before allowing a user to make a command, determine, based on the graph projection database 162, to determine that the user has a policy to be able to make the command.
In some embodiments, the policies can be conditional based policies. For example, the building data platform 100 can apply one or more conditional rules to determine whether a particular system has the ability to perform an action. In some embodiments, the rules analyze a behavioral based biometric. For example, a behavioral based biometric can indicate normal behavior and/or normal behavior rules for a system. In some embodiments, when the building data platform 100 determines, based on the one or more conditional rules, that an action requested by a system does not match a normal behavior, the building data platform 100 can deny the system the ability to perform the action and/or request approval from a higher level system.
For example, a behavior rule could indicate that a user has access to log into a system with a particular IP address between 8 A.M. through 5 P.M. However, if the user logs in to the system at 7 P.M., the building data platform 100 may contact an administrator to determine whether to give the user permission to log in.
The change feed generator 152 can be configured to generate a feed of events that indicate changes to the digital twin, e.g., to the graph. The change feed generator 152 can track changes to the entities, relationships, and/or events of the graph. For example, the change feed generator 152 can detect an addition, deletion, and/or modification of a node or edge of the graph, e.g., changing the entities, relationships, and/or events within the database 160. In response to detecting a change to the graph, the change feed generator 152 can generate an event summarizing the change. The event can indicate what nodes and/or edges have changed and how the nodes and edges have changed. The events can be posted to a topic by the change feed generator 152.
The change feed generator 152 can implement a change feed of a knowledge graph. The building data platform 100 can implement a subscription to changes in the knowledge graph. When the change feed generator 152 posts events in the change feed, subscribing systems or applications can receive the change feed event. By generating a record of all changes that have happened, a system can stage data in different ways, and then replay the data back in whatever order the system wishes. This can include running the changes sequentially one by one and/or by jumping from one major change to the next. For example, to generate a graph at a particular time, all change feed events up to the particular time can be used to construct the graph.
The change feed can track the changes in each node in the graph and the relationships related to them, in some embodiments. If a user wants to subscribe to these changes and the user has proper access, the user can simply submit a web API call to have sequential notifications of each change that happens in the graph. A user and/or system can replay the changes one by one to reinstitute the graph at any given time slice. Even though the messages are “thin” and only include notification of change and the reference “id/seq id,” the change feed can keep a copy of every state of each node and/or relationship so that a user and/or system can retrieve those past states at any time for each node. Furthermore, a consumer of the change feed could also create dynamic “views” allowing different “snapshots” in time of what the graph looks like from a particular context. While the twin manager 108 may contain the history and the current state of the graph based upon schema evaluation, a consumer can retain a copy of that data, and thereby create dynamic views using the change feed.
The schema and ontology 154 can define the message schema and graph ontology of the twin manager 108. The message schema can define what format messages received by the messaging manager 140 should have, e.g., what parameters, what formats, etc. The ontology can define graph projections, e.g., the ontology that a user wishes to view. For example, various systems, applications, and/or users can be associated with a graph ontology. Accordingly, when the graph projection manager 156 generates an graph projection for a user, system, or subscription, the graph projection manager 156 can generate a graph projection according to the ontology specific to the user. For example, the ontology can define what types of entities are related in what order in a graph, for example, for the ontology for a subscription of “Customer A,” the graph projection manager 156 can create relationships for a graph projection based on the rule:
For the ontology of a subscription of “Customer B,” the graph projection manager 156 can create relationships based on the rule:
The policy manager 158 can be configured to respond to requests from other applications and/or systems for policies. The policy manager 158 can consult a graph projection to determine what permissions different applications, users, and/or devices have. The graph projection can indicate various permissions that different types of entities have and the policy manager 158 can search the graph projection to identify the permissions of a particular entity. The policy manager 158 can facilitate fine grain access control with user permissions. The policy manager 158 can apply permissions across a graph, e.g., if “user can view all data associated with floor 1” then they see all subsystem data for that floor, e.g., surveillance cameras, HVAC devices, fire detection and response devices, etc.
The twin manager 108 includes a query manager 165 and a twin function manager 167. The query manger 164 can be configured to handle queries received from a requesting system, e.g., the user device 176, the applications 110, and/or any other system. The query manager 165 can receive queries that include query parameters and context. The query manager 165 can query the graph projection database 162 with the query parameters to retrieve a result. The query manager 165 can then cause an event processor, e.g., a twin function, to operate based on the result and the context. In some embodiments, the query manager 165 can select the twin function based on the context and/or perform operates based on the context. In some embodiments, the query manager 165 is configured to perform the operations described with reference to
The twin function manager 167 can be configured to manage the execution of twin functions. The twin function manager 167 can receive an indication of a context query that identifies a particular data element and/or pattern in the graph projection database 162. Responsive to the particular data element and/or pattern occurring in the graph projection database 162 (e.g., based on a new data event added to the graph projection database 162 and/or change to nodes or edges of the graph projection database 162, the twin function manager 167 can cause a particular twin function to execute. The twin function can execute based on an event, context, and/or rules. The event can be data that the twin function executes against. The context can be information that provides a contextual description of the data, e.g., what device the event is associated with, what control point should be updated based on the event, etc. The twin function manager 167 can be configured to perform the operations of the
Referring now to
The graph projection 200 includes a device hub 202 which may represent a software service that facilitates the communication of data and commands between the cloud platform 106 and a device of the building subsystems 122, e.g., door actuator 214. The device hub 202 is related to a connector 204, an external system 206, and a digital asset “Door Actuator” 208 by edge 250, edge 252, and edge 254.
The cloud platform 106 can be configured to identify the device hub 202, the connector 204, the external system 206 related to the door actuator 214 by searching the graph projection 200 and identifying the edges 250-254 and edge 258. The graph projection 200 includes a digital representation of the “Door Actuator,” node 208. The digital asset “Door Actuator” 208 includes a “DeviceNameSpace” represented by node 207 and related to the digital asset “Door Actuator” 208 by the “Property of Object” edge 256.
The “Door Actuator” 214 has points and timeseries. The “Door Actuator” 214 is related to “Point A” 216 by a “has_a” edge 260. The “Door Actuator” 214 is related to “Point B” 218 by a “has_A” edge 258. Furthermore, timeseries associated with the points A and B are represented by nodes “TS” 220 and “TS” 222. The timeseries are related to the points A and B by “has_a” edge 264 and “has_a” edge 262. The timeseries “TS” 220 has particular samples, sample 210 and 212 each related to “TS” 220 with edges 268 and 266 respectively. Each sample includes a time and a value. Each sample may be an event received from the door actuator that the cloud platform 106 ingests into the entity, relationship, and event database 160, e.g., ingests into the graph projection 200.
The graph projection 200 includes a building 234 representing a physical building. The building includes a floor represented by floor 232 related to the building 234 by the “has_a” edge from the building 234 to the floor 232. The floor has a space indicated by the edge “has_a” 270 between the floor 232 and the space 230. The space has particular capabilities, e.g., is a room that can be booked for a meeting, conference, private study time, etc. Furthermore, the booking can be canceled. The capabilities for the floor 232 are represented by capabilities 228 related to space 230 by edge 280. The capabilities 228 are related to two different commands, command “book room” 224 and command “cancel booking” 226 related to capabilities 228 by edge 284 and edge 282 respectively.
If the cloud platform 106 receives a command to book the space represented by the node, space 230, the cloud platform 106 can search the graph projection 200 for the capabilities for the 228 related to the space 230 to determine whether the cloud platform 106 can book the room.
In some embodiments, the cloud platform 106 could receive a request to book a room in a particular building, e.g., the building 234. The cloud platform 106 could search the graph projection 200 to identify spaces that have the capabilities to be booked, e.g., identify the space 230 based on the capabilities 228 related to the space 230. The cloud platform 106 can reply to the request with an indication of the space and allow the requesting entity to book the space 230.
The graph projection 200 includes a policy 236 for the floor 232. The policy 236 is related set for the floor 232 based on a “To Floor” edge 274 between the policy 236 and the floor 232. The policy 236 is related to different roles for the floor 232, read events 238 via edge 276 and send command 240 via edge 278. The policy 236 is set for the entity 203 based on has edge 251 between the entity 203 and the policy 236.
The twin manager 108 can identify policies for particular entities, e.g., users, software applications, systems, devices, etc. based on the policy 236. For example, if the cloud platform 106 receives a command to book the space 230. The cloud platform 106 can communicate with the twin manager 108 to verify that the entity requesting to book the space 230 has a policy to book the space. The twin manager 108 can identify the entity requesting to book the space as the entity 203 by searching the graph projection 200. Furthermore, the twin manager 108 can further identify the edge has 251 between the entity 203 and the policy 236 and the edge 1178 between the policy 236 and the command 240.
Furthermore, the twin manager 108 can identify that the entity 203 has the ability to command the space 230 based on the edge 1174 between the policy 236 and the edge 270 between the floor 232 and the space 230. In response to identifying the entity 203 has the ability to book the space 230, the twin manager 108 can provide an indication to the cloud platform 106.
Furthermore, if the entity makes a request to read events for the space 230, e.g., the sample 210 and the sample 212, the twin manager 108 can identify the edge has 251 between the entity 203 and the policy 236, the edge 1178 between the policy 236 and the read events 238, the edge 1174 between the policy 236 and the floor 232, the “has_a” edge 270 between the floor 232 and the space 230, the edge 268 between the space 230 and the door actuator 214, the edge 260 between the door actuator 214 and the point A 216, the “has_a” edge 264 between the point A 216 and the TS 220, and the edges 268 and 266 between the TS 220 and the samples 210 and 212 respectively.
Referring now to
The connection broker 354 is related to an agent that optimizes a space 356 via edge 398b. The agent represented by the node 356 can book and cancel bookings for the space represented by the node 230 based on the edge 398b between the connection broker 354 and the node 356 and the edge 398a between the capabilities 228 and the connection broker 354.
The connection broker 354 is related to a cluster 308 by edge 398c. Cluster 308 is related to connector B 302 via edge 398e and connector A 306 via edge 398d. The connector A 306 is related to an external subscription service 304. A connection broker 310 is related to cluster 308 via an edge 311 representing a rest call that the connection broker represented by node 310 can make to the cluster represented by cluster 308.
The connection broker 310 is related to a virtual meeting platform 312 by an edge 354. The node 312 represents an external system that represents a virtual meeting platform. The connection broker represented by node 310 can represent a software component that facilitates a connection between the cloud platform 106 and the virtual meeting platform represented by node 312. When the cloud platform 106 needs to communicate with the virtual meeting platform represented by the node 312, the cloud platform 106 can identify the edge 354 between the connection broker 310 and the virtual meeting platform 312 and select the connection broker represented by the node 310 to facilitate communication with the virtual meeting platform represented by the node 312.
A capabilities node 318 can be connected to the connection broker 310 via edge 360. The capabilities 318 can be capabilities of the virtual meeting platform represented by the node 312 and can be related to the node 312 through the edge 360 to the connection broker 310 and the edge 354 between the connection broker 310 and the node 312. The capabilities 318 can define capabilities of the virtual meeting platform represented by the node 312. The node 320 is related to capabilities 318 via edge 362. The capabilities may be an invite bob command represented by node 316 and an email bob command represented by node 314. The capabilities 318 can be linked to a node 320 representing a user, Bob. The cloud platform 106 can facilitate email commands to send emails to the user Bob via the email service represented by the node 304. The node 304 is related to the connect a node 306 via edge 398f Furthermore, the cloud platform 106 can facilitate sending an invite for a virtual meeting via the virtual meeting platform represented by the node 312 linked to the node 318 via the edge 358.
The node 320 for the user Bob can be associated with the policy 236 via the “has” edge 364. Furthermore, the node 320 can have a “check policy” edge 366 with a portal node 324. The device API node 328 has a check policy edge 370 to the policy node 236. The portal node 324 has an edge 368 to the policy node 236. The portal node 324 has an edge 323 to a node 326 representing a user input manager (UIM). The portal node 324 is related to the UIM node 326 via an edge 323. The UIM node 326 has an edge 323 to a device API node 328. The UIM node 326 is related to the door actuator node 214 via edge 372. The door actuator node 214 has an edge 374 to the device API node 328. The door actuator 214 has an edge 335 to the connector virtual object 334. The device hub 332 is related to the connector virtual object via edge 380. The device API node 328 can be an API for the door actuator 214. The connector virtual object 334 is related to the device API node 328 via the edge 331.
The device API node 328 is related to a transport connection broker 330 via an edge 329. The transport connection broker 330 is related to a device hub 332 via an edge 378. The device hub represented by node 332 can be a software component that hands the communication of data and commands for the door actuator 214. The cloud platform 106 can identify where to store data within the graph projection 300 received from the door actuator by identifying the nodes and edges between the points 216 and 218 and the device hub node 332. Similarly, the cloud platform 308 can identify commands for the door actuator that can be facilitated by the device hub represented by the node 332, e.g., by identifying edges between the device hub node 332 and an open door node 352 and an lock door node 350. The door actuator 114 has an edge “has mapped an asset” 280 between the node 214 and a capabilities node 348. The capabilities node 348 and the nodes 352 and 350 are linked by edges 396 and 394.
The device hub 332 is linked to a cluster 336 via an edge 384. The cluster 336 is linked to connector A 340 and connector B 338 by edges 386 and the edge 389. The connector A 340 and the connector B 338 is linked to an external system 344 via edges 388 and 390. The external system 344 is linked to a door actuator 342 via an edge 392.
Referring now to
A building node 404 represents a particular building that includes two floors. A floor 1 node 402 is linked to the building node 404 via edge 460 while a floor 2 node 406 is linked to the building node 404 via edge 462. The floor 2 includes a particular room 2023 represented by edge 464 between floor 2 node 406 and room 2023 node 408. Various pieces of equipment are included within the room 2023. A light represented by light node 416, a bedside lamp node 414, a bedside lamp node 412, and a hallway light node 410 are related to room 2023 node 408 via edge 466, edge 472, edge 470, and edge 468.
The light represented by light node 416 is related to a light connector 426 via edge 484. The light connector 426 is related to multiple commands for the light represented by the light node 416 via edges 484, 486, and 488. The commands may be a brightness setpoint 424, an on command 425, and a hue setpoint 428. The cloud platform 106 can receive a request to identify commands for the light represented by the light 416 and can identify the nodes 424-428 and provide an indication of the commands represented by the node 424-428 to the requesting entity. The requesting entity can then send commands for the commands represented by the nodes 424-428.
The bedside lamp node 414 is linked to a bedside lamp connector 481 via an edge 413. The connector 481 is related to commands for the bedside lamp represented by the bedside lamp node 414 via edges 492, 496, and 494. The command nodes are a brightness setpoint node 432, an on command node 434, and a color command 436. The hallway light 410 is related to a hallway light connector 446 via an edge 498d. The hallway light connector 446 is linked to multiple commands for the hallway light node 410 via edges 498g, 498f, and 498e. The commands are represented by an on command node 452, a hue setpoint node 450, and a light bulb activity node 448.
The graph projection 400 includes a name space node 422 related to a server A node 418 and a server B node 420 via edges 474 and 476. The name space node 422 is related to the bedside lamp connector 481, the bedside lamp connector 444, and the hallway light connector 446 via edges 482, 480, and 478. The bedside lamp connector 444 is related to commands, e.g., the color command node 440, the hue setpoint command 438, a brightness setpoint command 456, and an on command 454 via edges 498c, 498b, 498a, and 498.
Referring now to
A digital twin (or a shadow) may be a computing entity that describes a physical thing (e.g., a building, spaces of a building, devices of a building, people of the building, equipment of the building, etc.) through modeling the physical thing through a set of attributes that define the physical thing. A digital twin can refer to a digital replica of physical assets (a physical device twin) and can be extended to store processes, people, places, systems that can be used for various purposes. The digital twin can include both the ingestion of information and actions learned and executed through artificial intelligence agents.
In
The twin manager 108 stores the graph 529 which may be a graph data structure including various nodes and edges interrelating the nodes. The graph 529 may be the same as, or similar to, the graph projections described herein with reference to
The floor node 522 is related to the zone node 518 by the “has” edge 540 indicating that the floor represented by the floor node 522 has another zone represented by the zone node 518. The floor node 522 is related to another zone node 524 via a “has” edge 542 representing that the floor represented by the floor node 522 has a third zone represented by the zone node 524.
The graph 529 includes an AHU node 514 representing an AHU of the building represented by the building node 526. The AHU node 514 is related by a “supplies” edge 530 to the VAV node 512 to represent that the AHU represented by the AHU node 514 supplies air to the VAV represented by the VAV node 512. The AHU node 514 is related by a “supplies” edge 536 to the VAV node 520 to represent that the AHU represented by the AHU node 514 supplies air to the VAV represented by the VAV node 520. The AHU node 514 is related by a “supplies” edge 532 to the VAV node 516 to represent that the AHU represented by the AHU node 514 supplies air to the VAV represented by the VAV node 516.
The VAV node 516 is related to the zone node 518 via the “serves” edge 534 to represent that the VAV represented by the VAV node 516 serves (e.g., heats or cools) the zone represented by the zone node 518. The VAV node 520 is related to the zone node 524 via the “serves” edge 538 to represent that the VAV represented by the VAV node 520 serves (e.g., heats or cools) the zone represented by the zone node 524. The VAV node 512 is related to the zone node 510 via the “serves” edge 528 to represent that the VAV represented by the VAV node 512 serves (e.g., heats or cools) the zone represented by the zone node 510.
Furthermore, the graph 529 includes an edge 533 related to a timeseries node 564. The timeseries node 564 can be information stored within the graph 529 and/or can be information stored outside the graph 529 in a different database (e.g., a timeseries database). In some embodiments, the timeseries node 564 stores timeseries data (or any other type of data) for a data point of the VAV represented by the VAV node 516. The data of the timeseries node 564 can be aggregated and/or collected telemetry data of the timeseries node 564.
Furthermore, the graph 529 includes an edge 537 related to a timeseries node 566. The timeseries node 566 can be information stored within the graph 529 and/or can be information stored outside the graph 529 in a different database (e.g., a timeseries database). In some embodiments, the timeseries node 566 stores timeseries data (or any other type of data) for a data point of the VAV represented by the VAV node 516. The data of the timeseries node 564 can be inferred information, e.g., data inferred by one of the artificial intelligence agents 570 and written into the timeseries node 564 by the artificial intelligence agent 570. In some embodiments, the timeseries 546 and/or 566 are stored in the graph 529 but are stored as references to timeseries data stored in a timeseries database.
The twin manager 108 includes various software components. For example, the twin manager 108 includes a device management component 548 for managing devices of a building. The twin manager 108 includes a tenant management component 550 for managing various tenant subscriptions. The twin manager 108 includes an event routing component 552 for routing various events. The twin manager 108 includes an authentication and access component 554 for performing user and/or system authentication and grating the user and/or system access to various spaces, pieces of software, devices, etc. The twin manager 108 includes a commanding component 556 allowing a software application and/or user to send commands to physical devices. The twin manager 108 includes an entitlement component 558 that analyzes the entitlements of a user and/or system and grants the user and/or system abilities based on the entitlements. The twin manager 108 includes a telemetry component 560 that can receive telemetry data from physical systems and/or devices and ingest the telemetry data into the graph 529. Furthermore, the twin manager 108 includes an integrations component 562 allowing the twin manager 108 to integrate with other applications.
The twin manager 108 includes a gateway 506 and a twin connector 508. The gateway 506 can be configured to integrate with other systems and the twin connector 508 can be configured to allow the gateway 506 to integrate with the twin manager 108. The gateway 506 and/or the twin connector 508 can receive an entitlement request 502 and/or an inference request 504. The entitlement request 502 can be a request received from a system and/or a user requesting that an AI agent action be taken by the AI agent 570. The entitlement request 502 can be checked against entitlements for the system and/or user to verify that the action requested by the system and/or user is allowed for the user and/or system. The inference request 504 can be a request that the AI agent 570 generates an inference, e.g., a projection of information, a prediction of a future data measurement, an extrapolated data value, etc.
The cloud platform 106 is shown to receive a manual entitlement request 586. The request 586 can be received from a system, application, and/or user device (e.g., from the applications 110, the building subsystems 122, and/or the user device 176). The manual entitlement request 586 may be a request for the AI agent 570 to perform an action, e.g., an action that the requesting system and/or user has an entitlement for. The cloud platform 106 can receive the manual entitlement request 586 and check the manual entitlement request 586 against an entitlement database 584 storing a set of entitlements to verify that the requesting system has access to the user and/or system. The cloud platform 106, responsive to the manual entitlement request 586 being approved, can create a job for the AI agent 570 to perform. The created job can be added to a job request topic 580 of a set of topics 578.
The job request topic 580 can be fed to AI agents 570. For example, the topics 580 can be fanned out to various AI agents 570 based on the AI agent that each of the topics 580 pertains to (e.g., based on an identifier that identifies an agent and is included in each job of the topic 580). The AI agents 570 include a service client 572, a connector 574, and a model 576. The model 576 can be loaded into the AI agent 570 from a set of AI models stored in the AI model storage 568. The AI model storage 568 can store models for making energy load predictions for a building, weather forecasting models for predicting a weather forecast, action/decision models to take certain actions responsive to certain conditions being met, an occupancy model for predicting occupancy of a space and/or a building, etc. The models of the AI model storage 568 can be neural networks (e.g., convolutional neural networks, recurrent neural networks, deep learning networks, etc.), decision trees, support vector machines, and/or any other type of artificial intelligence, machine learning, and/or deep learning category. In some embodiments, the models are rule based triggers and actions that include various parameters for setting a condition and defining an action.
The AI agent 570 can include triggers 595 and actions 597. The triggers 595 can be conditional rules that, when met, cause one or more of the actions 597. The triggers 595 can be executed based on information stored in the graph 529 and/or data received from the building subsystems 122. The actions 597 can be executed to determine commands, actions, and/or outputs. The output of the actions 597 can be stored in the graph 529 and/or communicated to the building subsystems 122.
The AI agent 570 can include a service client 572 that causes an instance of an AI agent to run. The instance can be hosted by the artificial intelligence service client 588. The client 588 can cause a client instance 592 to run and communicate with the AI agent 570 via a gateway 590. The client instance 592 can include a service application 594 that interfaces with a core algorithm 598 via a functional interface 596. The core algorithm 598 can run the model 576, e.g., train the model 576 and/or use the model 576 to make inferences and/or predictions.
In some embodiments, the core algorithm 598 can be configured to perform learning based on the graph 529. In some embodiments, the core algorithm 598 can read and/or analyze the nodes and relationships of the graph 529 to make decisions. In some embodiments, the core algorithm 598 can be configured to use telemetry data (e.g., the timeseries data 564) from the graph 529 to make inferences on and/or perform model learning. In some embodiments, the result of the inferences can be the timeseries 566. In some embodiments, the timeseries 564 is an input into the model 576 that predicts the timeseries 566.
In some embodiments, the core algorithm 598 can generate the timeseries 566 as an inference for a data point, e.g., a prediction of values for the data point at future times. The timeseries 564 may be actual data for the data point. In this regard, the core algorithm 598 can learn and train by comparing the inferred data values against the true data values. In this regard, the model 576 can be trained by the core algorithm 598 to improve the inferences made by the model 576.
Referring now to
In step 602, the twin manager 108 receives information from a physical device and stores the information, or a link to the information, in the graph 529. For example, the telemetry component 560 can receive telemetry data from physical devices, e.g., the building subsystems 122. The telemetry can be measured data values, a log of historical equipment commands, etc. The telemetry component 560 can store the received information in the graph 529 by relating a node storing the information to a node representing the physical device. For example, the telemetry component 560 can store timeseries data as the timeseries 566 along by identifying that the physical device is a VAV represented by the VAV node 516 and that an edge 537 relates the VAV node 516 to the timeseries node 566.
In step 604, the twin manager 108 and/or the cloud platform 106 receives an indication to execute an artificial intelligence agent of an entity represented in the graph 529, the AI agent being associated with a model. In some embodiments, the indication is created by a user and provided via the user device 176. In some embodiments, the indication is created by an application, e.g., one of the applications 110. In some embodiments, the indication is a triggering event that triggers the agent and is received from the building subsystems 122 and/or another agent (e.g., an output of one agent fed into another agent).
In some embodiments, the AI agent is an agent for a specific entity represented in the graph 529. For example, the agent could be a VAV maintenance agent configured to identify whether a VAV (e.g., a VAV represented by the nodes 512, 530, and/or 516) should have maintenance performed at a specific time. Another agent could be a floor occupant prediction agent that is configure to predict the occupancy of a particular floor of a building, e.g., the floor represented by the floor node 522.
Responsive to receiving the indication, in step 606, the AI agent 570 causes a client instance 592 to run the model 576 based on the information received in step 602. In some embodiments, the information received in step 602 is provided directly to the AI agent 570. In some embodiments, the information is read from the graph 529 by the AI agent 570.
In step 608, the AI agent 570 stores the inferred and/or predicted information in the graph 529 (or stores the inferred and/or predicted information in a separate data structure with a link to the graph 529). In some embodiments, the AI agent 570 identifies that the node that represents the physical entity that the AI agent 570 inferred and/or predicted information for, e.g., the VAV represented by the VAV 516. The AI agent 570 can identify that the timeseries node 566 stores the inferred and/or predicted information by identifying the edge 537 between the VAV node 516 and the timeseries node 566.
In step 610, the AI agent 570 can retrieve the inferred or predicted information from the graph 529 responsive to receiving an indication to execute the model of the AI agent 570 of the inferred or predicted information, e.g., similar to the step 604. In step 612, the AI agent 570 can execute one or more actions based on the inferred and/or predicted information of the step 610 based the inferred and/or predicted information retrieved from the graph 529. In some embodiments, the AI agent 570 executes the model 576 based on the inferred and/or predicted information.
In step 614, the AI agent 570 can train the model 576 based on the inferred or predicted information read from the graph 529 and received actual values for the inferred or predicted information. In some embodiments, the AI agent 570 can train and update parameters of the model 576. For example, the timeseries 564 may represent actual values for a data point of the VAV represented by the VAV node 516. The timeseries 566 can be the inferred and/or predicted information. The AI agent 570 can compare the timeseries 564 and the timeseries 566 to determine an error in the inferences and/or predictions of the model 576. The error can be used by the model 576 to update and train the model 576.
Referring now to
Referring now to
In some cases, the devices can also be actuated on (told to perform an action). For example, a thermostat has sensors to measure temperature and humidity. A thermostat can also be asked to perform an action of setting the setpoint for a HVAC system. In this regard, the digital twin 800 can be configured so that information that the digital twin 800 can be made aware of can be stored by the digital twin 800 and there are also actions that the digital twin 800 can take.
The digital twin 800 can include a connector 804 that ingests device telemetry into the graph 808 and/or update the digital twin attributes stored in the graph 808. In some embodiments, the connectors 804 can ingest external data received from external data sources into the graph 808. The external data could be weather data, calendar data, etc. In some embodiments, the connectors 804 can send commands back to the devices, e.g., the actions determined by the actions 806.
The digital twin 800 includes triggers 802 which can set conditional logic for triggering the actions 706. The digital twin 800 can apply the attributes stored in the graph 808 against a rule of the triggers 802. When a particular condition of the rule of the triggers 802 involving that attribute is met, the actions 706 can execute. One example of a trigger could be a conditional question, “when the temperature of the zone managed by the thermostat reaches x degrees Fahrenheit.” When the question is met by the attributes store din the graph 808, a rule of the actions 706 can execute.
The digital twin 800 can, when executing the actions 806, update an attribute of the graph 808, e.g., a setpoint, an operating setting, etc. These attributes can be translated into commands that the building data platform 100 can send to physical devices that operate based on the setpoint, the operating setting, etc. An example of an action rule for the actions 806 could be the statement, “update the setpoint of the HVAC system for a zone to x Degrees Fahrenheit.”
In some embodiments, the triggers 802 and/or the actions 806 are predefined and/or manually defined through user input of the user device 176. In some cases, it may be difficult for a user to determine what the parameter values of the trigger rule should be (e.g., what values maximize a particular reward or minimize a particular penalty). Similarly, it may be difficult for a user to determine what the parameter values of the action rule should be (e.g., what values maximize the particular reward or minimize the particular penalty). Furthermore, even if the user is able to identify the ideal parameter values for the triggers 802 and the actions 806, the ideal values for the parameters may not be constant and may instead change over time. Therefore, it would be desirable if the values of the attributes for the triggers 802 and the actions 806 are tuned optimally and automatically by the building data platform 100 by observing the responses from other related digital twins.
Causal patterns between one or more digital twins having their triggering conditions satisfied and one or more digital twins (including the triggering digital twin) actuating by sending specific commands to their physical counterparts could be learned and defined by the building data platform 100. Automated learning can be used by the building data platform 100 during real operations, by running simulations using digital twins, or using predicted inference within the digital twin. There may not even be the need for all standard operating procedures in building systems to be defined upfront by a user since patterns of interaction between digital twins can be learned by the building data platform 100 to define and recommend those to building and facility owners.
Referring now to
The system 900 further includes an HVAC digital twin 904 that includes triggers 914, connectors 916, and actions 918. The system further includes the facility manager 906 that includes triggers 920, connectors 922, and actions 924. In some embodiments, the graph 926, the graph 928, and the graph 930 are the same graph or different graphs. In some embodiments, the graphs 926-930 are the graph 529.
In the system 900, the actions 912 are connected to the triggers 914 and the triggers 920. In this regard, whatever action is taken by the people counter digital twin 902, the result of the action will be provided to the HVAC digital twin 904 and the facility manager digital twin 906. The people counter digital twin 902 can output a “low occupancy” attribute which can be stored in the graph 926 and/or provided to the HVAC digital twin 904 and/or the facility manager digital twin 906. In some embodiments, if all of the digital twins use and/or have access to the same graph, if the people counter digital twin 902 stores the low occupancy indicator in the graph, the HVAC digital twin 904 and the facility manager digital twin 906 can read the attribute from the graph.
In some embodiments, the trigger 908 is the logical condition, “when there are less than twenty people in a particular area.” Responsive to an occupancy count of the particular area is less than twenty, which the people counter digital twin 902 can determine from models and/or information of the graph 926, a low occupancy indication can be generated by the actions 912. The low occupancy indication can be provided to the HVAC digital twin 904.
In some embodiments, the trigger 914 of the HVAC digital twin 904 can be the logical condition, “if there is low occupancy.” Similarly, the trigger 920 of the facility manager digital twin 906 can be the logical condition, “if there is low occupancy.” Responsive to the trigger 914 being triggered, the actions 918 can execute to switch an HVAC mode to an economy mode. The economy mode status for an HVAC system can be stored in the graph 928 and/or communicated to an HVAC controller to execute on. Responsive to the trigger 920 being triggered, the actions 924 can execute to notify a facility manager of the low occupancy status, e.g., send a notification to a user device of the facility manager.
In some embodiments, the digital twins of the system 900 can be solution twins, e.g., the people counter twin 902, the HVAC digital twin 904, the facility manager twin 906, etc. The digital twin can be a solution twin because it represents a particular software solutions for the building. For example, in some embodiments, an occupancy sensor digital twin of a zone could be triggered with under-utilized criteria (e.g., the triggering of the people counter digital twin 902 shown in
Referring now to
In
The employee digital twin 1002 can generate an “occupant near office” indication via the actions 1012 responsive to the trigger 1010 triggering when a particular occupant is a particular instance (e.g., 250 meters) from their office. The digital twin 1002 can identify the occupant, the occupant's office, and the location of the office through analyzing the nodes and/or edge of the graph 1032. The calendar digital twin 1006 determines, based on calendar data (e.g., calendar data stored in the graph 1036), whether it is a work day via the trigger 1014 (e.g., is a day Monday through Friday). Responsive to determining that it is a work day, the calendar digital twin 1006 generates an indication that it is a work day via the actions 1018.
The meeting room digital twin 1004 can receive the work day indication from the calendar digital twin 1006 and can receive the occupant near office indication from the employee digital twin 1002. The meeting room digital twin 1004 can take actions to reserve a meeting room via the actions 1024 responsive to the trigger 1020 indicating that the occupant is near their office and it is a work day. The cafeteria digital twin 1008 can receive the “occupant near office” indication from the employee digital twin 1002 and can receive the “it is a work day” indication from the calendar digital twin 1006. The cafeteria digital twin 1008 can trigger the ordering of a coffee for the occupant via the trigger 1030 responsive to the trigger 1026 being triggered.
Referring now to
In step 1102, the building data platform can store an agent 570 in a data structure. The agent 570 can include a trigger rule indicating a condition for executing an action rule and an action rule indicating an action to be performed responsive to the condition being met. In some embodiments, the model 576 includes, or can be replaced with, the trigger rule and the action rule. The trigger rule and the action rule can be logical statements and/or conditions that include parameter values and/or create an output action. The parameter values can, in some embodiments, be identified through a learning process, e.g., as described through
In step 1104, the agent 570 can receive information from at least one of a physical device and/or from the graph 529. The information can be generated by a physical device, e.g., the building subsystems 122. The building data platform 100 can, in some embodiments, receive the information from the physical device, ingest the information into the graph 529, and the agent 570 can read the information from the graph 529. In some embodiments, the agent 570 can check the information of the graph 529 against a trigger rule at a set period.
In step 1106, the agent 570 determines whether the information received in the step 1104 causes the condition to be met. The agent 570 can apply the information to the trigger rule to determine whether the trigger rule is triggered, i.e., the condition of the trigger rule being met.
In step 1108, the agent 570 can perform the action responsive to the condition being met by the information determined in step 1106. The action may cause a physical device to be operated or information be sent to another agent including another trigger rule and another action rule. In some embodiments, the action can be performed by executing the action rule of the agent 570. The action rule can perform an action based on one or more parameter value of the action rule. In some embodiments, the action output of the action rule can be sent directly to the physical device, e.g., the building subsystems 122. In some embodiments, the action output can be stored into the graph 529. Another operating component of the building data platform 100, e.g., the command processor 136, can read the action from the graph 529 can communicate a corresponding command to the building subsystems 122.
Referring generally to
In some embodiments, the building data platform 100 can perform q-learning (Reinforcement Learning) to train and/or retrain the triggers and/or actions of the agents. In some embodiments, the data used to train and/or retrain the triggers and/or actions can be simulated data determined by another digital twin.
One digital twin may have trigger conditions such as, “when the outside temperature is x0,” “when the inside humidity is x %,” “when an AI-driven algorithm's threshold is reached,” and “when it is a certain day of the week.” In responsive to one or multiple triggers being met, the digital twin can perform actions (e.g., capabilities of a device either inherent and/or digital twin enhanced). The actions can include setting a setpoint to a value x0. The actions can be to run a fan for x minutes. The actions can be to start an AI-driven energy saving schedule. The actions can be to change a mode status to an away status. In some embodiments, the building data platform 100 can user other digital twins to simulate a reward for various values of the triggers and/or actions. The reward can be optimized to determine values for the parameters of the triggers and/or actions.
In some embodiments, allowing the digital twin to learn and adjust the parameters of the triggers and/or rules allows the digital twin to optimize responses to internal and/or external events in real-time. In some embodiments, the digital twin performs operations with the correlation of contextual relationships to provide spatial intelligence. In some embodiments, the digital twin allows for AI-based self-learning solutions to operate on top of the digital twin. The digital twin can capture comprehensive data that drives rich analytics for security, compliance, etc. In some embodiments, the digital twin can enable and perform simulations.
In some embodiments, the building data platform 100 can identify events and/or event patterns if the building data platform 100 identifies a pattern that suggests a trigger and/or action should be updated. For example, if the building data platform 100 identifies a pattern occurring in a building, the building data platform 100 can set triggers and/or actions in digital twins to allow the pattern to occur automatically. For example, if a user closes their blinds at 5:00 P.M. regularly on weekdays, this could indicate that the user desires the blinds to be closed at 5:00 P.M. each day. The building data platform 100 can set a blind control digital twin to trigger a blind closing action at 5:00 P.M. each day.
In some embodiments, an agent of a digital twin can predict an inference in the future indicating that some action should be performed in the future. The building data platform 100 can identify that the action should be performed in the future and can set up a flow so that a prediction of one digital twin can be fed into another digital twin that can perform the action.
Referring now to
The building data platform 100 can perturb parameters, ε1 and ε2 of the trigger rule 1202 of the thermostat digital twin. The trigger rule 1202 may be that if a number of occupants is greater than ε1 and a zone temperature is less than ε2° C. the rule is triggered and a corresponding action be performed. The corresponding action can be to increase a supply air temperature setpoint of an AHU to 22° C. The perturbation of the parameters can be increasing or decreasing the parameters in set amounts from existing values. The perturbation of the parameters can be selecting a space of values for the parameters and/or randomizing the parameters and/or parameter space.
With the perturbed values for ε1 and ε2, the AHU digital twin 1204 can simulate the state of the AHU via the AHU digital twin 1204 for various conditions of occupant number and zone temperature. The result of the various states of the AHU digital twin 1204. The simulation can be performed by the AI agent 570 via the model 576. The output of the model 576 can be the simulated states, e.g., timeseries 566.
The building data platform 100 can analyze the states produced by the AHU digital twin 1204 to determine energy and comfort results from the states of the AHU digital twin 1204. For example, an energy score can be generated for each state. For example, a power consumption level can be determined for each state. Similarly, a comfort violation score can be determined for each state. The comfort violation can indicate whether or not a temperature, humidity, or other condition of a physical space controlled by the AHU would be uncomfortable for a user (e.g., go below or above certain levels).
The building data platform 100 can generate accumulated training data. The accumulated training data can include the values of the parameters ε1 and ε2, the state of the AHU digital twin 1204 for each value of the parameters, and the energy score and comfort violation score for each state. In some embodiments, the triggers and/or actions that can be recommended for the thermostat digital twin can be determined by observing the responses of other digital twins on perturbed thresholds of existing triggers and/or actions.
The building data platform 100 can generate neural networks 1210 for predicting an energy score based on the parameters ε1 and ε2. Furthermore, the neural networks 1210 can indicate a comfort violation score for the parameters ε1 and ε2. The neural networks 1210 can be trained by the building data platform 100 based on the accumulated training data 1208.
Based on the trained neural network models 1210, the building data platform 100 can determine optimal values for the parameters ε1 and ε2. The building data platform 100 can search a space of potential values for ε1 and ε2 that consider predicted energy scores and/or comfort violation scores predicted by the trained neural network models 1210. The optimization can be the relation 1400 shown in
In some embodiments, the optimal values for the parameters found by the building data platform 100 can be presented to a user for review and/or approval via a user interface, e.g., via the user device 176. In some embodiments, the recommendations produced by the building data platform 100 through the components 1202-1212 can be restricted by only looking at state/value changes of digital twins that are nearest neighbors in the graph 529, e.g., two nodes are directed related by one edge, e.g., a thermostat node for the thermostat digital twin is directed to an AHU node for the AHU digital twin 1204. In some embodiments, the building data platform 100 can use spatial correlation to assume contextual relationship between assets that can affect each other's attribute states/values.
Referring now to
In step 1302, the building data platform 100 can perturb a thermostat digital twin (e.g., the thermostat digital twin rule 1202) with various value for thresholds and/or other parameters, E. The result of the perturbed parameters can result in various states, s. The states can be states predicted by the thermostat digital twin or another digital twin that operates based on the thresholds and/or parameters E, e.g., the AHU digital twin 1204. The perturbations and simulated states can result in pairs (S, ε). The pairs can be used to determine feedback for energy and/or comfort, e.g., (E, C).
In step 1304, the building data platform 100 can building neural network models, e.g., the neural networks 1210 based on the data determined in step 1302. The neural networks 1210 can predict energy rewards as a function of the state and the parameters, e.g., E=f (s, ε). Furthermore, the neural networks 1210 can predict comfort rewards as a function of the state and the parameters, e.g., C=f (s, ε).
In step 1306, the building data platform 100 can determine a value for the parameter, ε that minimizes a relation, (α1·E+α2·C). The minimization is shown in relation 1400 of FIG. 14. The values of α1 and α2 can weigh the various rewards in the relation that is minimized, e.g., the energy reward and/or the comfort reward. In step 1308, the building data platform 100 can periodically repeat the steps 1302-1306. For example. For example, the building data platform 100 can repeat the steps at a defined time period. In some embodiments, the building data platform 100 can compute rewards for the actions of the thermostat digital twin. If the rewards indicate that the thermostat digital twin need retraining, the building data platform 100 can repeat the steps 1302-1308.
Referring now to
The thermostat digital twin rule 1502 can be an action rule that if a trigger is met (e.g., the trigger 1402), the action rule 1502 executes to command the AHU digital twin 1204. The trigger rule may be to execute the action rule if an occupant count is greater than ten and a zone temperature is less than twenty degrees Celsius. The action rule 1502 may be to increase an AHU supply air temperature setpoint to a value, e.g., ε. The value can, in some embodiments, be 22 degrees Celsius.
The building data platform 100 can predict states resulting from perturbed values of E by executing the AHU digital twin 1204 to simulate the states. The building data platform 100 can collect rule feedback 1206 to construct accumulated training data 1208. Furthermore, the building data platform 100 can train neural network models 1210 based on the accumulated training data 1208 and find optimal values for the parameter E based on the trained neural network models 1210
Referring now to
The list 1602 includes states for an AHU. The states can be supply air temperature, supply air flow rate, return air temperature, return air flow rate, outdoor air flow rate, etc. These states can be predicted and/or determined based on a digital twin for an AHU (e.g., the AHU digital twin 1204) based on perturbed parameter values for a trigger rule, an action rule, etc. In this regard, the rule feedback 1206 in some embodiments, can be generated based on the digital twin for the AHU and used to tune the values of the parameters for the trigger rule 1402 and/or the action rule 1502.
Referring now to
The building data platform 100 can perturb the values for the parameters ε1 and ε2 of the reactor feed digital twin trigger rule 1702 (e.g., pseudo-randomly, increasing and/or decreasing in a particular number of predefined increments, etc.). A chemical reactor digital twin 1704 can simulate a state of the chemical reactor for the various perturbed parameters ε1 and ε2. The building data platform 100 can determine a rule feedback 1706 for the state simulate by the chemical reactor digital twin 1704. The rule feedback 1706 can identify scores for production throughput (P) and chemical property (C).
The building data platform 100 can accumulate training data 1708. The accumulated training data 1708 can include the feedback 1706, the state simulated by the chemical reactor digital twin 1704, and/or the parameter values for ε1 and ε2. The building data platform 100 can train neural network models 1710 to predict production throughput and/or chemical property for the various parameter and/or state pairs, e.g., the state resulting from the parameters of the trigger rule 1702. The building data platform 100 can use the trained neural network models 1710 to identify optimal values for ε1 and ε2. In element 1712, the building data platform 100 can identify values for ε1 and ε2 that minimize the relation 1900 shown in
Referring now to
In step 1802, the building data platform 100 can perturb a reactor digital twin 1704 with various values of a threshold E of a trigger rule 1702 with various values which cause the reactor digital twin to determine resulting states for the various values of the threshold, E. The states and the values for the threshold E can create state threshold pairs. The pairs can be used to determine feedback, e.g., production throughput and chemical property.
In step 1804, after some accumulation of feedback data, the building data platform 100 can build neural network models 1710 based on the pairs that predict production throughput and chemical property based on the values for the threshold E. In step 1806, the building data platform 100 can determine a value for the threshold E that maximizes a reward and/or minimizes a penalty. The building data platform 100 can minimize the relation 1900 of
Referring now to
Referring now to
The list 2102 includes states for a chemical reactor. The states can be product concentration, cooling coil temperature, product temperature, etc. These states can be predicted and/or determined based on a digital twin for a chemical reactor (e.g., the reactor digital twin 1704) based on perturbed parameter values for a trigger rule, an action rule, etc. In this regard, the rule feedback 1706 in some embodiments, can be generated based on the digital twin for the chemical reactor and used to tune the values of the parameters for the trigger rule 1702 and/or the action rule 2002.
Referring now to
In some embodiments, the building data platform 100 can search the graph 529 to identify information related to the space, e.g., related pieces of equipment, spaces, people, etc. For example, the building data platform 100 can identify which entities of the graph 529 are related and operate to affect each other. The building data platform 100 can identify which actions each entity can perform and/or what measurements each entity can make, e.g., by identifying related data nodes for each entity. The identified entities, measurements, and/or commands can be combined into the rule 2202 by the building data platform 100.
In some embodiments, the learning service 2206, which may be a component of the building data platform 100, can run a learning process with the rule 2202 and/or one or more reward functions (e.g., comfort reward function, carbon footprint reduction reward function, the energy reduction reward function 2204, etc.). The learning service 2206 can learn the rule 2208 from the rule 2202 and/or the energy reduction reward function 2204.
The learning service 2206 can run an optimization to determine combinations between measurements and actions triggered based on the measurements. The learning service 2206 can determine values for each measurement and/or action. Furthermore, the learning service 2206 can identify the relational operations for causing a trigger, e.g., equals to, greater than, less than, not equal to, etc. Furthermore, the learning service 2206 can identify action operations, e.g., increase by a particular amount, decrease by a particular amount, set an output equal to a value, run a particular algorithm, etc.
Referring now to
In step 2302, the building data platform 100 can determine actions that a particular entity can take and data that the entity can measure by analyzing a graph 529. The entity can be a thermostat, an air handler unit, a zone of a building, a person, a VAV unit, and/or any other entity. For example, if the entity is a thermostat the building data platform 100 could identify room temperature measurements for a thermostat and/or a cooling stage command, a heating stage command, a fan command, etc. that the thermostat can perform. Responsive to identifying data that the entity can measure, the building data platform 100 can generate a trigger condition based on the data type, e.g., when the temperature is equal to, less than, greater than, and/or not equal to some parameter value, trigger an action.
In step 2304, the building data platform 100 identifies, based on the graph 529, entities related to the entity and actions that the entities can take and data that the entities can measure. For example, if the entity is for a thermostat for a zone, the building data platform 100, could identify a shade control system for controlling a shade of the zone, an air handler unit that serves the zone, a VAV that serves the zone, etc. For example, the building data platform 100 can identify, based on the building graph 529, that a binds node is associated with a zone node that the thermostat node is related to. The building data platform 100 can identify a list of actions that the entities can perform, e.g., setting blind position from 0% (fully open) to 100% (fully closed).
In some 2306, the building data platform 100 can simulate various combinations of triggers that tare based on the data that the entity and/or entities can measure and actions that are based on the actions that the entity and/or entities can make. The building data platform 100 can simulate various combinations, trigger operations, action operations, and/or parameters.
In step 2308, the building data platform 100 can identify a combination of triggers and actions that maximizes a reward. The building data platform 100 can search the simulated combinations of triggers and/or actions to identify a trigger and/or action that maximizes a reward and/or minimizes a reward. In some embodiments, the building data platform 100 uses a policy gradient and value function instead of brute force to try out combinations of the triggers and/or actions in the steps 2306-2308.
In some embodiments, the building data platform 100 can identify the operations for the triggers and/or actions. For example, the operation could be comparing a measurement to a threshold, determining whether a measurement is less than a threshold, determining whether a measurement is greater than the threshold, determining whether the measurement is not equal to the threshold, etc.
In step 2310, the building data platform 100 can generate a digital twin for the entity. The entity can include (or reference) the graph 529 and include an agent that operates the triggers and/or actions. The triggers and/or actions can operate based on the graph 529 and/or based on data received building equipment, e.g., the building subsystems 122.
In step 2312, the building data platform 100 can run a building system of a building and monitor the behavior of the entity and entities of the building. In some embodiments, the building system can be the building subsystems 122. In step 2314, the building data platform 100 can identify relationships between the measurements and actions of the entity and/or the entities based on the monitored behavior. The building data platform 100 can discover existing relationships by identifying how the measurements are currently affecting actions based on the monitored behavior. In step 2316, the building data platform 100 can optimize the identified relationships between the measurements and the actions by maximizing a reward or minimizing a penalty.
Referring now to
The system 2400 includes a schema infusing tool 2404. The schema infusing tool can infuse a particular schema, the schema 2402, into various systems, services, and/or equipment in order to integrate the data of the various systems, services, and/or equipment into the building data platform 100. The schema 2402 may be the BRICK schema, in some embodiments. In some embodiment, the schema 2402 may be a schema that uses portions and/or all of the BRICK schema but also includes unique class, relationship types, and/or unique schema rules. The schema infusing tool 2404 can infuse the schema 2402 into systems such as systems that manage and/or produce building information model (BIM) data 2418, building automation system (BAS) systems that produce BAS data 2420, and/or access control and video surveillance (ACVS) systems that produce ACVS data 2422. In some embodiments, the BIM data 2418 can be generated by BIM automation utilities 2501.
The BIM data 2418 can include data such as Revit data 2424 (e.g., Navisworks data), industrial foundation class (IFC) data 2426, gbxml data 2428, and/or CoBie data 2430. The BAS data 2420 can include Modelica data 2432 (e.g., Control Description Language (CDL) data), Project Haystack data 2434, Bacnet data 2436, Metasys data 2438, and/or Easy IO data 2440. All of this data can utilizing the schema 2402 and/or be capable of being mapped into the schema 2402.
The BAS data 2420 and/or the ACVS data 2422 may include timeseries data 2408. The timeseries data 2408 can include trends of data points over time, e.g., a time correlated set of data values each corresponding to time stamps. The timeseries data can be a timeseries of data measurements, e.g., temperature measurements, pressure measurements, etc. Furthermore, the timeseries data can be a timeseries of inferred and/or predicted information, e.g., an inferred temperature value, an inferred energy load, a predicted weather forecast, identities of individuals granted access to a facility over time, etc. The timeseries data 2408 can further indicate command and/or control data, e.g., the damper position of a VAV over time, the setpoint of a thermostat over time, etc.
The system 2400 includes a schema mapping toolchain 2412. The schema mapping toolchain 2412 can map the data of the metadata sources 2406 into data of the schema 2402, e.g., the data in schema 2414. The data in schema 2414 may be in a schema that can be integrated by an integration toolchain 2416 with the building data platform 100 (e.g. ingested into the databases, graphs, and/or knowledge bases of the building data platform 100) and/or provided to the AI services and applications 2410 for execution).
The AI services and applications 2410 include building control 2442, analytics 2444, microgrid management 2446, and various other applications 2448. The building control 2442 can include various control applications that may utilize AI, ML, and/or any other software technique for managing control of a building. The building control 2442 can include auto sequence of operation, optimal supervisory controls, etc. The analytics 2444 include clean air optimization (CAO) applications, energy prediction model (EPM) applications, and/or any other type of analytics.
Referring now to
The system 2500 includes various tools for converting the metadata sources 2406 into the data in schema 2414. Various mapping tools 2502-2512 can map data from an existing schema into the schema 2402. For example, the mapping tools 2502-2512 can utilize a dictionary that provides mapping rules and syntax substitutions. In some embodiments, that data sources can have the schema 2402 activated, e.g., schema enable 2518-2522. If the schema 2402 is enabled for a Metasys data source, an easy IO data source, or an ACVS data sources, the output data by said systems can be in the schema 2402. Examples of schema mapping techniques can be found in U.S. patent application Ser. No. 16/663,623 filed Oct. 25, 2019, U.S. patent application Ser. No. 16/885,968 filed May 28, 2020, and U.S. patent application Ser. No. 16/885,959 filed May 28, 2020, the entireties of which are incorporated by reference herein.
For the EasyIO data 2440, the EasyIO controller objects could be tagged with classes of the schema 2402. For the Revit data 2424, the metadata of a REVIT model could be converted into the schema 2402, e.g., into a resource description format (RDF). For the Metasys data 2438, Metasys SCT data could be converted into RDF. An OpenRefine aided mapping tool 2514 and/or a natural language aided mapping tool 2516 could perform the schema translation for the Bacnet data 2436.
The schema data output by the tools 2502-2522 can be provided to a reconciliation tool 2530. The reconciliation tool 2530 can be configured to merge complementary or duplicate information and/or resolve any conflicts in the data received from the tools 2502-2522. The result of the reconciliation performed by the reconciliation tool 2530 can be the data in schema 2414 which can be ingested into the building data platform 100 by the ingestion tool 2532. The ingestion tool 2532 can generate and/or update one or more graphs managed and/or stored by the twin manager 108. The graph could be the graphs described with reference to
The system 2500 includes agents that perform operations on behalf of the AI services and applications 2410. For example, as shown in the system 2500, the analytics 2444 are related to various agents, a CAO AI agent 2524, an EPM AI agent 2526, and various other AI agents 2528. The agents 2524-2528 can receive data from the building data platform 100, e.g., the data that the ingestion tool 2532 ingests into the building data platform 100, and generate analytics data for the analytics 2444.
Referring now to
The nodes may represent various entities of a building and/or buildings. The entities may be a campus, a building, a floor, a space, a zone, a piece of equipment, a person, a control point, a data measurement point, a sensor, an actuator, telemetry data, a piece of timeseries data, etc. The edges 2644-2678 can interrelate the nodes 2608-2642 to represent the relationships between the various entities of the building. The edges 2644-2678 can be semantic language based edges 2644-2678. The edges can include words and/or phrases that represent the relationship. The words and/or phrases can include at least one predicate, in some cases.
The knowledge graph 2602 includes a building node 2636 representing a building. The building can include floors, represented by the building node 2636 being related to a floor 1 node 2640 via an “hasPart” edge 2670 and the building node 2636 being related to a floor 3 node 2638 via a “feeds” edge 2668. The building includes an energy prediction, e.g., a value or a timeseries of values indicating energy usage of the building. This can be represented by the building node 2636 being related to an energy prediction node 2630 via the edge 2672.
The floor 1 includes zones, indicated by the floor 1 node 2640 being related to a zone1-1 node 2608 via a “hasPart” edge 2644 and the floor 1 node 2640 being related to a zone1-2 node 2642 via a “hasPart” edge 2645. Furthermore, the floor 1 can be fed by a particular AHU, AHU1. This is indicated by an AHU1 node 2622 being related to the floor 1 node 2640 via the “feeds” edge 2660. The zone1-2 can include a VAV1-2 that feeds air to it. This can be indicated by the VAV1-2 node 2610 being related to the zone1-2 node 2642 by the “feeds” node 2662. The AHU1 can feed the VAV1-2, indicated by the AHU1 node 2622 being related to the VAV1-2 node 2610 by the “feeds” edge 2658.
An AHU 3 can feed air to the floor 3, indicated by the “AHU3” node 2634 being related to the “floor 3” node 2638 by the “feeds” edge 2666. The “AHU3” node 2634 is related to a CAO inference 2628 via an edge 2674 representing that the AHU has a clean air optimization inference that could be determined by the CAO AI service 2604. The knowledge graph 2602 includes an AHU2 node 2632. The AHU2 2632 is related to a floor 2 node 2641 via a “feeds” edge 2664 indicating that the AHU2 feeds air to the floor 2. The AHU2 node 2632 is related to a CAO inference 2626 via the edge 2676. The CAO inference 2626 can indicate an inference made by the CAO AI service 2604 for the AHU2. The AHU1 2622 is related to a CAO inference 2624 via edge 2678. The CAO inference 2624 can indicate a clean air optimization inference made by the CAO AI service 2604 for the AHU1.
The knowledge graph 2602 includes a VAV1-1 node 2612 indicating a VAV1-1. The VAV1-1 node 2612 is related to the zone1-1 node 2608 via a “feeds” edge 2646 indicating that the VAV1-1 feeds air to the zone1-1. The AHU1 can feed air to the VAV1-1 indicated by the AHU1 node 2622 being related to the VAV1-1 node 2612 via the “feeds” edge 2656. The VAV1-1 node 2612 includes various points, e.g., a zone temperature point (represented by the VAV1.ZN-T node 2614) and a zone relative humidity point (represented by the VAV1.ZN-RH node 2616). The VAV1-1 node 2612 is related to the VAV1.ZN-T node 2614 via the “hasPoint” edge 2648. The VAV1-1 node 2612 is related to the VAV1.ZN-RH node 2616 via the “hasPoint” edge 2654.
The VAV1.ZN-T point includes a timeseries node 2618 representing and/or storing a timeseries for the zone temperature, indicated by the timeseries node 2618 being related to the VAV1.ZN-T node 2614 via the “hasTimeseriesID” node 2650. The VAV1.ZN-RH point includes a timeseries node 2620 representing and/or storing a timeseries for the zone humidity, indicated by the timeseries node 2620 being related to the VAV1.ZN-RH node 2616 via the “hasTimeseriesID” node 2652. In some embodiments, the timeseries node 2618 and the timeseries node 2620 are identifiers of a particular timeseries stored in a separate timeseries database, the identifier uniquely identifying the location of the timeseries so a system can consult the knowledge graph 2602 and use the identifiers to retrieve the timeseries data from the separate timeseries database.
The system 2600 includes a CAO AI service 2604. The CAO AI service 2604 can be configured to identify timeseries that it needs to execute on. For example, if the CAO AI service 2604 is executing for the AHU1, the CAO AI service 2604 could identify timeseries data linked to the AHU1. The CAO AI service 2604 can generate CAO inferences, e.g., can infer ideal settings for clean air. The ideal settings could be an AHU supply air temperature setpoint, an AHU minimum ventilation rate, etc. The ideal settings can be ingested into the knowledge graph 2602, e.g., as the CAO inferences 2624-2630.
In some embodiments, the CAO AI service 2604 (or an agent for the CAO AI service 2604) operates on behalf of a particular entity, e.g., the AHU 1. The CAO AI service 2604 can generate inferences with data of the AHU 1, e.g., by identifying timeseries data of the AHU 1 by identifying timeseries nodes of the knowledge graph 2602 via an edge. The inferences can be ingested into the knowledge graph 2602 by generating a new node and/or adding an edge between the new node and the node of the AHU 1, AHU 1 node 2622. Similarly, the inferences can be added to the knowledge graph 2602 by updating an existing node related to the AHU 1 node 2622 via an existing edge. In some embodiments, the inferences of the CAO AI service 2604 can generate a recommendation, e.g., a control setting for improving or optimizing air quality, which can be reviewed and approved by a user via the user device 176.
The EPM AI service 2606 can generate energy predictions for various buildings, spaces, or devices of a building, e.g., entities of the knowledge graph 2602. For example, the EPM AI service 2606 could predict a future energy consumption level of the building 2636, e.g., a future energy demand. The energy prediction can be a node 2630 related to the building node 2636 via the edge 2672. In some embodiments, the EPM AI service 2606 can generate the energy prediction node 2630 responsive to generating the future building energy consumption and cause the node 2630 to include a value for the future building energy consumption. The node 2630 can be added to the graph 2602 and the edge 2672 can be generated by the EPM AI service 2606 and added between the building node 2636 and the energy prediction 2630.
In some embodiments, the energy prediction node 2630 already exists within the knowledge graph 2602. In this example, the EPM AI service 2606 can identify the energy prediction node 2630 by identifying an edge 2672 between the building 2636 and the energy prediction node 2630. The EPM AI service 2606 can then ingest the energy prediction into the node 2630.
Referring now to
The CAO AI agent 2524 can operate on behalf of the CAO AI service 2604. Similarly, the EPM AI agent 2526 can operate on behalf of the EPM AI service 2606. Furthermore a service bus 2714 can interface with the agent 2524 and/or the agent 2526. A user can interface with the agents 2524-2526 via the user device 176. The user can provide an entitlement request, e.g., a request that the user is entitled to make and can be verified by the AI agent manager 2716. The AI agent manager 2716 can send an AI job request based on a schedule to the service bus 2714 based on the entitlement request. The service bus 2714 can communicate the AI job request to the appropriate agent and/or communicate results for the AI job back to the user device 176.
In some embodiments, the CAO AI agent 2524 can provide a request for generating an inference to the CAO AI service 2604. The request can include data read from the knowledge graph 2602, in some embodiments.
The CAO AI agent 2524 includes a client 2702, a schema translator 2704, and a CAO client 2706. The client 2702 can be configured to interface with the knowledge graph 2602, e.g., read data out of the knowledge graph 2602. The client 2702 can further ingest inferences back into the knowledge graph 2602. For example, the client 2702 could identify timeseries nodes related to the AHU1 node 2622, e.g., timeseries nodes related to the AHU1 node 2622 via one or more edges. The client 2702 can then ingest the inference made by the CAO AI agent 2524 into the knowledge graph 2602, e.g., add the CAO inference 2624 or update the CAO inference 2624.
The client 2702 can provide data it reads from the knowledge graph 2602 to a schema translator 2704 that may translate the data into a specific format in a specific schema that is appropriate for consumption by the CAO client 2706 and/or the CAO AI service 2604. The CAO client 2706 can run one or more algorithms, software components, machine learning models, etc. to generate the inference and provide the inference to the client 2702. In some embodiments, the client 2706 can interface with the EPM AI service 2606 and provide the translated data to the EPM AI service 2606 for generating an inference. The inference can be returned by the EPM AI service 2606 to the CAO client 2706.
The EPM AI agent 2526 can operate in a similar manner to the CAO AI agent 2524, in some embodiments. The client 2708 can retrieve data from the knowledge graph 2602 and provide the data to the schema translator 2710. The schema translator 2710 can translate the data into a readable format by the CAO AI service 2604 and can provide the data to the EPM client 2712. The EPM client 2712 can provide the data along with a prediction request to the CAO AI service 2604. The CAO AI service 2604 can generate the prediction and provide the prediction to the EPM client 2712. The EPM client 2712 can provide the prediction to the client 2708 and the client 2708 can ingest the prediction into the knowledge graph 2602.
In some embodiments, the knowledge graph 2602 includes data necessary for the inferences and/or predictions that the agents 2524 and 2526 generate. For example, the knowledge graph 2602 can store information such as the size of a building, the number of floors of the building, the equipment of each floor of the building, the square footage of each floor, square footage of each zone, ceiling heights, etc. The data can be stored as nodes in the knowledge graph 2602 representing the physical characteristics of the building. In some embodiments, the EPM AI agent 2526 makes the predictions based on the characteristic data of the building and/or physical areas of the building.
Referring now to
The system 2800 includes a client 2802. The client 2802 can integrate with the knowledge graph 2602 and also with a graphical building model 2804 that can be rendered on a screen of the user device 176.
The client 2802 can retrieve information from the knowledge graph 2602, e.g., an inference generated by the CAO AI agent 2524, a prediction made by the EPM AI agent 2526, and/or operational data stored in the knowledge graph 2602. The client 2802 can ingest the values of the retrieved information into the graphical building model 2804 which can be displayed on the user device 176. In some embodiments, when a particular visual component is being displayed on the user device 176 for the virtual model 2804, e.g., a building, the corresponding information for the building can be displayed in the interface, e.g., inferences, predictions, and/or operational data.
For example, the client 2802 could identify a node of the building in the knowledge graph 2602, e.g., the building node 2636. The client 2802 could identify information linked to the building node 2636 via edges, e.g., the energy prediction 2630 related to the building node 2636 via edge 2672. The client 2802 can cause the energy prediction 2630 to be displayed in the graphical building model 2804.
In some embodiments, a user can provide input through the graphical building model 2804. The input may be a manual action that a user provides via the user device 176. The manual action can be ingested into the knowledge graph 2602 and stored as a node within the knowledge graph 2602. In some embodiments, the manual action can trigger one of the agents 2524-2526 causing the agent to generate an inference and/or prediction which is ingested into the knowledge graph 2602 and presented for user review in the model 2804.
In some embodiments, the agents 2524-2526 are the same as, or similar to, the agents described with reference to
In some embodiments, the animations of the model 2804 can be based on the inferences and/or predictions of the agents 2524-2526. In some embodiments, charts or graphs can be included within the model 2804, e.g., charting or graphing timeseries values of the inferences and/or predictions. For example, if an inference is an inference of a flow rate of a fluid (e.g., water, air, refrigerant, etc.) through a conduit, the speed at which arrows moving through the virtual conduit can be controlled based on the inferred flow rate inferred by an agent. Similarly, if the model 2804 provides a heat map indicating occupancy, e.g., red indicating high occupancy, blue indicating medium occupancy, and green indicating low occupancy, an agent could infer an occupancy level for each space of the building and the color coding for the heat map of the model 2804 could be based on the inference made by the agent.
In some embodiments, the graphical building model 2804 can be a three dimensional or two dimensional graphical building. The graphical building model 2804 can be a building information model (BIM), in some embodiments. The BIM can be generated and viewed based on the knowledge graph 2602. An example of rendering graph data and/or BIM data in a user interface is described in greater detail in U.S. patent application Ser. No. 17/136,752 filed Dec. 29, 2020, U.S. patent application Ser. No. 17/136,768 filed Dec. 29, 2020, and U.S. patent application Ser. No. 17/136,785 filed Dec. 29, 2020, the entirety of which is incorporated by reference herein.
In some embodiments, the graphical building model 2804 includes one or multiple three dimensional building elements 2806. The three dimensional building elements 2806 can form a building when combined, e.g., can form a building model of a building or a campus model of a campus. The building elements 2806 can include floors of a building, spaces of a building, equipment of a building, etc. Furthermore, each three dimensional building element 2806 can be linked to certain data inferences 2808, predictions 2810, and/or operational data 2812. The data 2808-2812 can be retrieved from the knowledge graph 2602 for display in an interface via the user device 176.
Referring now to
In
For example, the digital twin could be the knowledge graph 2602 combined with one or more agents, e.g., the CAO AI agent 2524 and/or the EPM AI agent 2526. The agents can generated inferred and/or predicted information for the AHU based at least in part on operational data stored in the knowledge graph 2602 (e.g., operational data of the AHU).
For example, the CAO AI agent 2524 could be configured to generate a minimum ventilation rate of the AHU, e.g., 29 CFM/occupant. The CAO AI agent 2524 can ingest the minimum ventilation rate into the knowledge graph 2602, e.g., into a node of the knowledge graph 2602 related by a relationship to a node for the AHU. The client 2802 can provide the minimum ventilation rate to the graphical building model 2804 which can be displayed in the element 2908 of the interface 2900. Similarly, EPM AI agent 2526 could generate the energy usage prediction for the AHU, e.g., 10 kWh. The energy prediction can be ingested into the knowledge graph 2602 and then provided to a user via the user interface 2900 in the element 2908.
In some embodiments, a user may provide an input to the interface 2900, e.g., a command, a manual action, a new data value, etc. The input can cause the digital twin to generate new inferences and/or predictions which can be ingested into the knowledge graph 2602 and then displayed in the interface 2900. In some embodiments, the input is a new occupancy value input via the element 2908. The new occupancy value can be ingested into the knowledge graph 2602 into a node related to a node for the AHU. The CAO AI agent 2524 can trigger (e.g., if it includes a trigger rule to execute based on a change in occupancy).
The CAO AI agent 2524 can re-compute the minimum ventilation rate based on the new occupancy value for clean air and can store the updated value in the knowledge graph 2602, thus updating the value displayed in the element 2908. Similarly, the energy usage displayed in the element 2908 could be reproduce by the EPM AI agent 2526.
Referring generally to the
In some embodiments, past information and/or future predictions can be ingested into the knowledge graph 2602 (e.g., into a digital twin or twins). In this regard, past, present, and/or future information can be displayed and/or visualized within a graphical building model. Furthermore, the client 2802 can, in some embodiments, provide a recorder and/or playback element (e.g., a widget) that allows a user to view past, present, and/or future states and values of the building within the graphical representation of the building.
In some embodiments, actionable insights can be included within the three dimensional view of the building. These actionable insights can aid a user in reviewing and understanding what issues may be affecting the building and what actions could be taken to resolve the issues. In some embodiments, the actionable insights can be acted on (e.g., approved and/or disapproved). The three dimensional view of the building can further display the status of the actionable insights and/or results of implementing the actionable insights, e.g., if the actions resolved in the issues.
In some embodiments, the three dimensional building can include indications of alerts. Furthermore, the three dimensional building can indicate virtual zones for occupancy and/or monitoring of a group of individuals for a particular metric. In some embodiments, an integration can exist (e.g., the client 2802) that can integrate any AI service output into the three dimensional building. The client 2802 may abstract the integration and store a template/configuration for integrating information of the knowledge graph 2602 into the three dimensional building.
In some embodiments, the user interface can display entity data for various spaces, assets, solutions, people, etc. The entity data can be data collected and ingested into a digital twin and/or inferred, derived, and/or predicted information generated by an agent running against the digital twin. The data could be live telemetry flowing from various entities and attributes of the entities stored in the digital twin. In some embodiments, the entity data can be a predicted telemetry stored as an event stream with future timestamps for a space, asset, person, etc. in the digital twin. In some embodiments, the entity data can be actionable insight and/or recommendations with future timestamps for spaces, assets, people, etc. In some embodiments, the entity data can include actionable insights for present and/or future times for spaces, assets, and/or people of a digital twin. The actionable insights can have present timestamps and/or future timestamps. The entity data can further indicate an actionable insight status view indicating status of a space, asset, and/or solution after a particular recommendation is accepted. The status could indicate submitted, confirmed, applied, etc.
In some embodiments, information for addressing an infectious disease (e.g., COVID-19, influenza, smallpox, etc.) can be displayed within the three dimensional building. The information could be performance information, actionable information, indoor air quality, reproductive number, energy consumption, etc. The information can further indicate diagnosis information and/or actionable insights derived by an AI service for resolving an issue (fault, poor performance, etc.) for a space, piece of equipment, or user.
The three dimensional graphical building 3000 can represent a building with settings, rooms, cabins rooms, AHUs, VAVs, ACs, etc. The three dimensional graphical building 3000 can include the floor 3100. The three dimensional graphical building 3000 and/or the floor 3100 can be updated with information based on time-series data (e.g., zone temperatures, supply fan outputs, maximum supply air flows, etc.) received from a building management system managing and/or operating the rooms, AHUs, VAVs, ACs, etc. The timeseries data can be ingested into a digital twin of the building. An agent can operate against the digital twin to derive predictions or inferences, and display information within the three dimensional graphical building 3000 based on the derived predictions or inferences.
Referring now to
Referring now to
Similarly, the reproductive number can be a value generated by an AI agent for various spaces of a building based on measurements made for the various spaces of the building and/or operating settings of equipment of the various areas of the building. The reproductive number may be the expected number of cases directly generated by one case in a population where all individuals are susceptible to infection. The reproductive number can be generated by the AI agent based on telemetry stored in the knowledge graph 2602 for the various spaces of the building. The clean air scores generated by the AI agent can be ingested by the AI agent into the knowledge graph 2602. The client 2802 and/or data formatter 3202 can read out the clean air score and/or reproductive number for the various spaces of the building and display indications of the clean air scores and/or reproductive numbers in the various spaces of the building within the floor 3300.
Referring now to
Referring now to
Referring now to
Referring now to
Referring now to
Referring now to
Responsive to interacting with the low clean air score, the client 2802 can provide diagnostic information at 3906 and/or provide action information at 3912. The diagnostic information and/or the action information can be information identified by an agent and ingested into the knowledge graph 2602. The client 2802 can read the diagnostic information and/or action information out of the knowledge graph 2602 and display the information in the floor 3904 or in a user interface element on the above, below, or on the side of the floor 3904 within a user interface. The diagnostic information can provide a reason for the clean air score, e.g., an indication 3908 that the ventilation rate is too low or an indication 3910 that a filter is not in use. The action information can include an indication 3914 that a supply air temperature setpoint point needs to be adjusted (e.g., to a particular value), an indication 3916 that a minimum ventilation rate needs to be adjusted (e.g., to a particular value), and/or an indication 3918 that a particular filter needs to be added to an AHU. A user can interact with the provided action information to cause the settings to automatically update and/or a work order to be generated to replace the filter.
Referring now to
Responsive to interacting with the high reproductive number, the client 2802 can provide diagnostic information at 4006 and/or provide action information at 4012. The diagnostic information and/or the action information can be information identified by an agent and ingested into the knowledge graph 2602. The client 2802 can read the diagnostic information and/or action information out of the knowledge graph 2602 and display the information in the floor 4004 or in a user interface element on the above, below, or on the side of the floor 4004 within a user interface. The diagnostic information can provide a reason or the high reproductive number, e.g., an indication 4008 that the ventilation rate is too low or an indication 4010 that an occupant density level is too high. The action information can include an indication 4014 to stay at home and work remotely, an indication 4016 to wear a mask, and/or an indication 4018 to maintain social distancing.
Referring now to
Referring now to
In some embodiments, the element 4202 allows a user to view “what if” scenarios by moving back and forth in time and testing various settings and configurations for equipment of the building and/or the building. In some embodiments, for various settings and/or configurations provided by a user, an agent can infer future states for various future times. The user can provide various settings and view the varying results for each set of settings or configuration.
In some embodiments, via the element 4200 various recommended action can be displayed to resolve a future predicted issue (e.g., fault, poor air quality, high reproduction rate, etc.). A user can set, via the element 4200, approval to automatically generate a ticket for maintenance, update operating settings of equipment, etc. The element 4200 can allow a user to approve a trigger to automatically perform action if a scenario simulated and displayed in the element 4200 does in fact occur.
Referring now to
The element 4300 can allow a user to view states of the building at a current time, one or more days into the future, and/or one or more days into the past. The element 4303 can be a slider allowing a user to flip through the states of the building at historical, current, or future times. The elements 4300 and 4302 can provide a “Building Player” allowing a user to “play” the building view the past statuses of the building (e.g., stored in a timeseries database and/or a digital twin), present statuses of the building, and/or future statuses of the building (e.g., simulated and/or inference/predictions generated by a digital twin). The predicted future statuses can be generated by an agent based on past and/or current data for the building, in some embodiments.
Referring now to
In some embodiments, the occupancy is predicted from schedules and/or calendars and/or predicted from historical occupancy trends. In some embodiments, the predicted occupancy could be fed into a clean air optimization model to predict the air quality. In some embodiments, the element 4400 can display actions for avoiding the future poor air quality scenario. The user can provide approval to perform the action and/or allow an agent to automatically operate to avoid the poor air scenario. In some cases, the poor air quality scenario may be unavoidable. In such a case, the element 4400 can include an indication that the scenario is unavoidable.
The construction and arrangement of the systems and methods as shown in the various exemplary embodiments are illustrative only. Although only a few embodiments have been described in detail in this disclosure, many modifications are possible (e.g., variations in sizes, dimensions, structures, shapes and proportions of the various elements, values of parameters, mounting arrangements, use of materials, colors, orientations, etc.). For example, the position of elements may be reversed or otherwise varied and the nature or number of discrete elements or positions may be altered or varied. Accordingly, all such modifications are intended to be included within the scope of the present disclosure. The order or sequence of any process or method steps may be varied or re-sequenced according to alternative embodiments. Other substitutions, modifications, changes, and omissions may be made in the design, operating conditions and arrangement of the exemplary embodiments without departing from the scope of the present disclosure.
The present disclosure contemplates methods, systems and program products on any machine-readable media for accomplishing various operations. The embodiments of the present disclosure may be implemented using existing computer processors, or by a special purpose computer processor for an appropriate system, incorporated for this or another purpose, or by a hardwired system. Embodiments within the scope of the present disclosure include program products comprising machine-readable media for carrying or having machine-executable instructions or data structures stored thereon. Such machine-readable media can be any available media that can be accessed by a general purpose or special purpose computer or other machine with a processor. By way of example, such machine-readable media can comprise RAM, ROM, EPROM, EEPROM, CD-ROM or other optical disk storage, magnetic disk storage or other magnetic storage devices, or any other medium which can be used to carry or store desired program code in the form of machine-executable instructions or data structures and which can be accessed by a general purpose or special purpose computer or other machine with a processor. When information is transferred or provided over a network or another communications connection (either hardwired, wireless, or a combination of hardwired or wireless) to a machine, the machine properly views the connection as a machine-readable medium. Thus, any such connection is properly termed a machine-readable medium. Combinations of the above are also included within the scope of machine-readable media. Machine-executable instructions include, for example, instructions and data which cause a general purpose computer, special purpose computer, or special purpose processing machines to perform a certain function or group of functions.
Although the figures show a specific order of method steps, the order of the steps may differ from what is depicted. Also two or more steps may be performed concurrently or with partial concurrence. Such variation will depend on the software and hardware systems chosen and on designer choice. All such variations are within the scope of the disclosure. Likewise, software implementations could be accomplished with standard programming techniques with rule based logic and other logic to accomplish the various connection steps, processing steps, comparison steps and decision steps.
In various implementations, the steps and operations described herein may be performed on one processor or in a combination of two or more processors. For example, in some implementations, the various operations could be performed in a central server or set of central servers configured to receive data from one or more devices (e.g., edge computing devices/controllers) and perform the operations. In some implementations, the operations may be performed by one or more local controllers or computing devices (e.g., edge devices), such as controllers dedicated to and/or located within a particular building or portion of a building. In some implementations, the operations may be performed by a combination of one or more central or offsite computing devices/servers and one or more local controllers/computing devices. All such implementations are contemplated within the scope of the present disclosure. Further, unless otherwise indicated, when the present disclosure refers to one or more computer-readable storage media and/or one or more controllers, such computer-readable storage media and/or one or more controllers may be implemented as one or more central servers, one or more local controllers or computing devices (e.g., edge devices), any combination thereof, or any other combination of storage media and/or controllers regardless of the location of such devices.
Number | Name | Date | Kind |
---|---|---|---|
5301109 | Landauer et al. | Apr 1994 | A |
5446677 | Jensen et al. | Aug 1995 | A |
5581478 | Cruse et al. | Dec 1996 | A |
5812962 | Kovac | Sep 1998 | A |
5960381 | Singers et al. | Sep 1999 | A |
5973662 | Singers et al. | Oct 1999 | A |
6014612 | Larson et al. | Jan 2000 | A |
6031547 | Kennedy | Feb 2000 | A |
6134511 | Subbarao | Oct 2000 | A |
6157943 | Meyer | Dec 2000 | A |
6285966 | Brown et al. | Sep 2001 | B1 |
6363422 | Hunter et al. | Mar 2002 | B1 |
6385510 | Hoog et al. | May 2002 | B1 |
6389331 | Jensen et al. | May 2002 | B1 |
6401027 | Xu et al. | Jun 2002 | B1 |
6437691 | Sandelman et al. | Aug 2002 | B1 |
6477518 | Li et al. | Nov 2002 | B1 |
6487457 | Hull et al. | Nov 2002 | B1 |
6493755 | Hansen et al. | Dec 2002 | B1 |
6577323 | Jamieson et al. | Jun 2003 | B1 |
6626366 | Kayahara et al. | Sep 2003 | B2 |
6646660 | Patty | Nov 2003 | B1 |
6704016 | Oliver et al. | Mar 2004 | B1 |
6732540 | Sugihara et al. | May 2004 | B2 |
6764019 | Kayahara et al. | Jul 2004 | B1 |
6782385 | Natsumeda et al. | Aug 2004 | B2 |
6813532 | Eryurek et al. | Nov 2004 | B2 |
6816811 | Seem | Nov 2004 | B2 |
6823680 | Jayanth | Nov 2004 | B2 |
6826454 | Sulfstede | Nov 2004 | B2 |
6865511 | Frerichs et al. | Mar 2005 | B2 |
6925338 | Eryurek et al. | Aug 2005 | B2 |
6986138 | Sakaguchi et al. | Jan 2006 | B1 |
7031880 | Seem et al. | Apr 2006 | B1 |
7401057 | Eder | Jul 2008 | B2 |
7552467 | Lindsay | Jun 2009 | B2 |
7627544 | Chkodrov et al. | Dec 2009 | B2 |
7818249 | Lovejoy et al. | Oct 2010 | B2 |
7889051 | Billig et al. | Feb 2011 | B1 |
7996488 | Casabella et al. | Aug 2011 | B1 |
8078330 | Brickfield et al. | Dec 2011 | B2 |
8104044 | Scofield et al. | Jan 2012 | B1 |
8229470 | Ranjan et al. | Jul 2012 | B1 |
8401991 | Wu et al. | Mar 2013 | B2 |
8495745 | Schrecker et al. | Jul 2013 | B1 |
8516016 | Park et al. | Aug 2013 | B2 |
8532808 | Drees et al. | Sep 2013 | B2 |
8532839 | Drees et al. | Sep 2013 | B2 |
8600556 | Nesler et al. | Dec 2013 | B2 |
8635182 | Mackay | Jan 2014 | B2 |
8682921 | Park et al. | Mar 2014 | B2 |
8731724 | Drees et al. | May 2014 | B2 |
8737334 | Ahn et al. | May 2014 | B2 |
8738334 | Jiang et al. | May 2014 | B2 |
8751487 | Byrne et al. | Jun 2014 | B2 |
8788097 | Drees et al. | Jul 2014 | B2 |
8805995 | Oliver | Aug 2014 | B1 |
8843238 | Wenzel et al. | Sep 2014 | B2 |
8874071 | Sherman et al. | Oct 2014 | B2 |
8941465 | Pineau et al. | Jan 2015 | B2 |
8990127 | Taylor | Mar 2015 | B2 |
9070113 | Shafiee et al. | Jun 2015 | B2 |
9116978 | Park et al. | Aug 2015 | B2 |
9185095 | Moritz et al. | Nov 2015 | B1 |
9189527 | Park et al. | Nov 2015 | B2 |
9196009 | Drees et al. | Nov 2015 | B2 |
9229966 | Aymeloglu et al. | Jan 2016 | B2 |
9286582 | Drees et al. | Mar 2016 | B2 |
9311807 | Schultz et al. | Apr 2016 | B2 |
9344751 | Ream et al. | May 2016 | B1 |
9354968 | Wenzel et al. | May 2016 | B2 |
9507686 | Horn et al. | Nov 2016 | B2 |
9524594 | Ouyang et al. | Dec 2016 | B2 |
9558196 | Johnston et al. | Jan 2017 | B2 |
9652813 | Gifford et al. | May 2017 | B2 |
9753455 | Drees | Sep 2017 | B2 |
9811249 | Chen et al. | Nov 2017 | B2 |
9838844 | Emeis et al. | Dec 2017 | B2 |
9886478 | Mukherjee | Feb 2018 | B2 |
9948359 | Horton | Apr 2018 | B2 |
10055114 | Shah et al. | Aug 2018 | B2 |
10055206 | Park et al. | Aug 2018 | B2 |
10116461 | Fairweather et al. | Oct 2018 | B2 |
10169454 | Ait-Mokhtar et al. | Jan 2019 | B2 |
10171297 | Stewart et al. | Jan 2019 | B2 |
10171586 | Shaashua et al. | Jan 2019 | B2 |
10187258 | Nagesh et al. | Jan 2019 | B2 |
10514963 | Shrivastava et al. | Dec 2019 | B2 |
10515098 | Park et al. | Dec 2019 | B2 |
10534326 | Sridharan et al. | Jan 2020 | B2 |
10536295 | Fairweather et al. | Jan 2020 | B2 |
10564993 | Deutsch et al. | Feb 2020 | B2 |
10705492 | Harvey | Jul 2020 | B2 |
10708078 | Harvey | Jul 2020 | B2 |
10760815 | Janakiraman et al. | Sep 2020 | B2 |
10762475 | Song et al. | Sep 2020 | B2 |
10824120 | Ahmed | Nov 2020 | B2 |
10845771 | Harvey | Nov 2020 | B2 |
10854194 | Park et al. | Dec 2020 | B2 |
10862928 | Badawy et al. | Dec 2020 | B1 |
10921760 | Harvey | Feb 2021 | B2 |
10921972 | Park et al. | Feb 2021 | B2 |
10969133 | Harvey | Apr 2021 | B2 |
10986121 | Stockdale et al. | Apr 2021 | B2 |
11016998 | Park et al. | May 2021 | B2 |
11024292 | Park et al. | Jun 2021 | B2 |
11038709 | Park et al. | Jun 2021 | B2 |
11041650 | Li et al. | Jun 2021 | B2 |
11054796 | Holaso | Jul 2021 | B2 |
11070390 | Park et al. | Jul 2021 | B2 |
11073976 | Park et al. | Jul 2021 | B2 |
11108587 | Park et al. | Aug 2021 | B2 |
11113295 | Park et al. | Sep 2021 | B2 |
11216663 | Ettinger | Jan 2022 | B1 |
11229138 | Harvey et al. | Jan 2022 | B1 |
11264140 | Tai | Mar 2022 | B1 |
11314726 | Park et al. | Apr 2022 | B2 |
11314788 | Park et al. | Apr 2022 | B2 |
20020010562 | Schleiss et al. | Jan 2002 | A1 |
20020016639 | Smith et al. | Feb 2002 | A1 |
20020059229 | Natsumeda et al. | May 2002 | A1 |
20020123864 | Eryurek et al. | Sep 2002 | A1 |
20020147506 | Eryurek et al. | Oct 2002 | A1 |
20020177909 | Fu et al. | Nov 2002 | A1 |
20030005486 | Ridolfo et al. | Jan 2003 | A1 |
20030014130 | Grumelart | Jan 2003 | A1 |
20030073432 | Meade, II | Apr 2003 | A1 |
20030158704 | Triginai et al. | Aug 2003 | A1 |
20030171851 | Brickfield et al. | Sep 2003 | A1 |
20030200059 | Ignatowski et al. | Oct 2003 | A1 |
20040068390 | Saunders | Apr 2004 | A1 |
20040128314 | Katibah et al. | Jul 2004 | A1 |
20040133314 | Ehlers et al. | Jul 2004 | A1 |
20040199360 | Friman et al. | Oct 2004 | A1 |
20050055308 | Meyer et al. | Mar 2005 | A1 |
20050108262 | Fawcett et al. | May 2005 | A1 |
20050154494 | Ahmed | Jul 2005 | A1 |
20050278703 | Lo et al. | Dec 2005 | A1 |
20050283337 | Sayal | Dec 2005 | A1 |
20060095521 | Patinkin | May 2006 | A1 |
20060140207 | Eschbach et al. | Jun 2006 | A1 |
20060184479 | Levine | Aug 2006 | A1 |
20060200476 | Gottumukkala et al. | Sep 2006 | A1 |
20060265751 | Cosquer et al. | Nov 2006 | A1 |
20060271589 | Horowitz et al. | Nov 2006 | A1 |
20070028179 | Levin et al. | Feb 2007 | A1 |
20070203693 | Estes | Aug 2007 | A1 |
20070261062 | Bansal et al. | Nov 2007 | A1 |
20070273497 | Kuroda et al. | Nov 2007 | A1 |
20070273610 | Baillot | Nov 2007 | A1 |
20080034425 | Overcash et al. | Feb 2008 | A1 |
20080094230 | Mock et al. | Apr 2008 | A1 |
20080097816 | Freire et al. | Apr 2008 | A1 |
20080186160 | Kim et al. | Aug 2008 | A1 |
20080249756 | Chaisuparasmikul | Oct 2008 | A1 |
20080252723 | Park | Oct 2008 | A1 |
20080281472 | Podgorny et al. | Nov 2008 | A1 |
20090195349 | Frader-Thompson et al. | Aug 2009 | A1 |
20100045439 | Tak et al. | Feb 2010 | A1 |
20100058248 | Park | Mar 2010 | A1 |
20100131533 | Ortiz | May 2010 | A1 |
20100274366 | Fata et al. | Oct 2010 | A1 |
20100281387 | Holland et al. | Nov 2010 | A1 |
20100286937 | Hedley et al. | Nov 2010 | A1 |
20100324962 | Nesler et al. | Dec 2010 | A1 |
20110015802 | Imes | Jan 2011 | A1 |
20110047418 | Drees et al. | Feb 2011 | A1 |
20110061015 | Drees et al. | Mar 2011 | A1 |
20110071685 | Huneycutt et al. | Mar 2011 | A1 |
20110077950 | Hughston | Mar 2011 | A1 |
20110087650 | Mackay et al. | Apr 2011 | A1 |
20110087988 | Ray et al. | Apr 2011 | A1 |
20110088000 | Mackay | Apr 2011 | A1 |
20110125737 | Pothering et al. | May 2011 | A1 |
20110137853 | Mackay | Jun 2011 | A1 |
20110153603 | Adiba et al. | Jun 2011 | A1 |
20110154363 | Karmarkar | Jun 2011 | A1 |
20110157357 | Weisensale et al. | Jun 2011 | A1 |
20110178977 | Drees | Jul 2011 | A1 |
20110191343 | Heaton et al. | Aug 2011 | A1 |
20110205022 | Cavallaro et al. | Aug 2011 | A1 |
20110218777 | Chen et al. | Sep 2011 | A1 |
20120011126 | Park et al. | Jan 2012 | A1 |
20120011141 | Park et al. | Jan 2012 | A1 |
20120022698 | Mackay | Jan 2012 | A1 |
20120062577 | Nixon | Mar 2012 | A1 |
20120064923 | Imes et al. | Mar 2012 | A1 |
20120083930 | Ilic et al. | Apr 2012 | A1 |
20120100825 | Sherman et al. | Apr 2012 | A1 |
20120101637 | Imes et al. | Apr 2012 | A1 |
20120135759 | Imes et al. | May 2012 | A1 |
20120136485 | Weber et al. | May 2012 | A1 |
20120158633 | Eder | Jun 2012 | A1 |
20120259583 | Noboa et al. | Oct 2012 | A1 |
20120272228 | Marndi et al. | Oct 2012 | A1 |
20120278051 | Jiang et al. | Nov 2012 | A1 |
20130007063 | Kalra et al. | Jan 2013 | A1 |
20130038430 | Blower et al. | Feb 2013 | A1 |
20130038707 | Cunningham et al. | Feb 2013 | A1 |
20130060820 | Bulusu et al. | Mar 2013 | A1 |
20130086497 | Ambuhl et al. | Apr 2013 | A1 |
20130097706 | Titonis et al. | Apr 2013 | A1 |
20130103221 | Raman et al. | Apr 2013 | A1 |
20130167035 | Imes et al. | Jun 2013 | A1 |
20130170710 | Kuoch et al. | Jul 2013 | A1 |
20130204836 | Choi et al. | Aug 2013 | A1 |
20130246916 | Reimann et al. | Sep 2013 | A1 |
20130247205 | Schrecker et al. | Sep 2013 | A1 |
20130262035 | Mills | Oct 2013 | A1 |
20130275174 | Bennett et al. | Oct 2013 | A1 |
20130275908 | Reichard | Oct 2013 | A1 |
20130297050 | Reichard et al. | Nov 2013 | A1 |
20130298244 | Kumar et al. | Nov 2013 | A1 |
20130331995 | Rosen | Dec 2013 | A1 |
20130338970 | Reghetti | Dec 2013 | A1 |
20140032506 | Hoey et al. | Jan 2014 | A1 |
20140059483 | Mairs et al. | Feb 2014 | A1 |
20140081652 | Klindworth | Mar 2014 | A1 |
20140135952 | Maehara | May 2014 | A1 |
20140152651 | Chen et al. | Jun 2014 | A1 |
20140172184 | Schmidt et al. | Jun 2014 | A1 |
20140189861 | Gupta et al. | Jul 2014 | A1 |
20140207282 | Angle et al. | Jul 2014 | A1 |
20140258052 | Khuti et al. | Sep 2014 | A1 |
20140269614 | Maguire et al. | Sep 2014 | A1 |
20140277765 | Karimi et al. | Sep 2014 | A1 |
20140278461 | Artz | Sep 2014 | A1 |
20140327555 | Sager et al. | Nov 2014 | A1 |
20150019174 | Kiff et al. | Jan 2015 | A1 |
20150042240 | Aggarwal et al. | Feb 2015 | A1 |
20150105917 | Sasaki et al. | Apr 2015 | A1 |
20150145468 | Ma et al. | May 2015 | A1 |
20150156031 | Fadell et al. | Jun 2015 | A1 |
20150168931 | Jin | Jun 2015 | A1 |
20150172300 | Cochenour | Jun 2015 | A1 |
20150178421 | Borrelli et al. | Jun 2015 | A1 |
20150185261 | Frader-Thompson et al. | Jul 2015 | A1 |
20150186777 | Lecue et al. | Jul 2015 | A1 |
20150202962 | Habashima et al. | Jul 2015 | A1 |
20150204563 | Imes et al. | Jul 2015 | A1 |
20150235267 | Steube et al. | Aug 2015 | A1 |
20150241895 | Lu et al. | Aug 2015 | A1 |
20150244730 | Vu et al. | Aug 2015 | A1 |
20150244732 | Golshan et al. | Aug 2015 | A1 |
20150261863 | Dey et al. | Sep 2015 | A1 |
20150263900 | Polyakov et al. | Sep 2015 | A1 |
20150286969 | Warner et al. | Oct 2015 | A1 |
20150295796 | Hsiao et al. | Oct 2015 | A1 |
20150304193 | Ishii et al. | Oct 2015 | A1 |
20150316918 | Schleiss et al. | Nov 2015 | A1 |
20150324422 | Elder | Nov 2015 | A1 |
20150341212 | Hsiao et al. | Nov 2015 | A1 |
20150348417 | Ignaczak et al. | Dec 2015 | A1 |
20150379080 | Jochimski | Dec 2015 | A1 |
20160011753 | McFarland et al. | Jan 2016 | A1 |
20160033946 | Zhu et al. | Feb 2016 | A1 |
20160035246 | Curtis | Feb 2016 | A1 |
20160065601 | Gong et al. | Mar 2016 | A1 |
20160070736 | Swan et al. | Mar 2016 | A1 |
20160078229 | Gong et al. | Mar 2016 | A1 |
20160090839 | Stolarczyk | Mar 2016 | A1 |
20160119434 | Dong et al. | Apr 2016 | A1 |
20160127712 | Ag | May 2016 | A1 |
20160139752 | Shim et al. | May 2016 | A1 |
20160163186 | Davidson et al. | Jun 2016 | A1 |
20160170390 | Xie et al. | Jun 2016 | A1 |
20160171862 | Das et al. | Jun 2016 | A1 |
20160173816 | Huenerfauth et al. | Jun 2016 | A1 |
20160179315 | Sarao et al. | Jun 2016 | A1 |
20160179342 | Sarao et al. | Jun 2016 | A1 |
20160179990 | Sarao et al. | Jun 2016 | A1 |
20160195856 | Spero | Jul 2016 | A1 |
20160212165 | Singla et al. | Jul 2016 | A1 |
20160239660 | Azvine et al. | Aug 2016 | A1 |
20160239756 | Aggour et al. | Aug 2016 | A1 |
20160247129 | Song et al. | Aug 2016 | A1 |
20160253679 | Venkatraman | Sep 2016 | A1 |
20160260063 | Harris et al. | Sep 2016 | A1 |
20160313751 | Risbeck et al. | Oct 2016 | A1 |
20160313752 | Przybylski | Oct 2016 | A1 |
20160313902 | Hill et al. | Oct 2016 | A1 |
20160350364 | Anicic et al. | Dec 2016 | A1 |
20160357521 | Zhang et al. | Dec 2016 | A1 |
20160357828 | Tobin et al. | Dec 2016 | A1 |
20160358432 | Branscomb et al. | Dec 2016 | A1 |
20160363336 | Roth et al. | Dec 2016 | A1 |
20160370258 | Perez | Dec 2016 | A1 |
20160378306 | Kresl et al. | Dec 2016 | A1 |
20160379326 | Chan-Gove et al. | Dec 2016 | A1 |
20170006135 | Siebel | Jan 2017 | A1 |
20170011318 | Vigano et al. | Jan 2017 | A1 |
20170017221 | Lamparter et al. | Jan 2017 | A1 |
20170039255 | Raj et al. | Feb 2017 | A1 |
20170052536 | Warner et al. | Feb 2017 | A1 |
20170053441 | Nadumane et al. | Feb 2017 | A1 |
20170063894 | Muddu et al. | Mar 2017 | A1 |
20170068409 | Nair | Mar 2017 | A1 |
20170070775 | Taxier et al. | Mar 2017 | A1 |
20170075984 | Deshpande et al. | Mar 2017 | A1 |
20170084168 | Janchookiat | Mar 2017 | A1 |
20170090437 | Veeramani et al. | Mar 2017 | A1 |
20170093700 | Gilley et al. | Mar 2017 | A1 |
20170098086 | Hoernecke et al. | Apr 2017 | A1 |
20170103327 | Penilla et al. | Apr 2017 | A1 |
20170103403 | Chu et al. | Apr 2017 | A1 |
20170123389 | Baez et al. | May 2017 | A1 |
20170134415 | Muddu et al. | May 2017 | A1 |
20170177715 | Chang et al. | Jun 2017 | A1 |
20170180147 | Brandman et al. | Jun 2017 | A1 |
20170188216 | Koskas et al. | Jun 2017 | A1 |
20170212482 | Boettcher et al. | Jul 2017 | A1 |
20170212668 | Shah et al. | Jul 2017 | A1 |
20170220641 | Chi et al. | Aug 2017 | A1 |
20170230930 | Frey | Aug 2017 | A1 |
20170235817 | Deodhar et al. | Aug 2017 | A1 |
20170251182 | Siminoff et al. | Aug 2017 | A1 |
20170270124 | Nagano et al. | Sep 2017 | A1 |
20170277769 | Pasupathy et al. | Sep 2017 | A1 |
20170278003 | Liu | Sep 2017 | A1 |
20170294132 | Colmenares | Oct 2017 | A1 |
20170315522 | Kwon et al. | Nov 2017 | A1 |
20170315697 | Jacobson et al. | Nov 2017 | A1 |
20170322534 | Sinha et al. | Nov 2017 | A1 |
20170323389 | Vavrasek | Nov 2017 | A1 |
20170329289 | Kohn et al. | Nov 2017 | A1 |
20170336770 | Macmillan | Nov 2017 | A1 |
20170345287 | Fuller et al. | Nov 2017 | A1 |
20170351957 | Lecue et al. | Dec 2017 | A1 |
20170357225 | Asp et al. | Dec 2017 | A1 |
20170357490 | Park et al. | Dec 2017 | A1 |
20170357908 | Cabadi et al. | Dec 2017 | A1 |
20180012159 | Kozloski et al. | Jan 2018 | A1 |
20180013579 | Fairweather et al. | Jan 2018 | A1 |
20180024520 | Sinha et al. | Jan 2018 | A1 |
20180039238 | Ag | Feb 2018 | A1 |
20180048485 | Pelton et al. | Feb 2018 | A1 |
20180069932 | Tiwari et al. | Mar 2018 | A1 |
20180114140 | Chen et al. | Apr 2018 | A1 |
20180137288 | Polyakov | May 2018 | A1 |
20180157930 | Rutschman et al. | Jun 2018 | A1 |
20180162400 | Abdar | Jun 2018 | A1 |
20180176241 | Manadhata et al. | Jun 2018 | A1 |
20180198627 | Mullins | Jul 2018 | A1 |
20180203961 | Aisu et al. | Jul 2018 | A1 |
20180239982 | Rutschman et al. | Aug 2018 | A1 |
20180275625 | Park et al. | Sep 2018 | A1 |
20180276962 | Butler et al. | Sep 2018 | A1 |
20180277246 | Zhong | Sep 2018 | A1 |
20180292797 | Lamparter et al. | Oct 2018 | A1 |
20180336785 | Ghannam et al. | Nov 2018 | A1 |
20180356775 | Harvey | Dec 2018 | A1 |
20180359111 | Harvey | Dec 2018 | A1 |
20180364654 | Locke et al. | Dec 2018 | A1 |
20190005025 | Malabarba | Jan 2019 | A1 |
20190013023 | Pourmohammad et al. | Jan 2019 | A1 |
20190025771 | Park et al. | Jan 2019 | A1 |
20190037135 | Hedge | Jan 2019 | A1 |
20190042988 | Brown et al. | Feb 2019 | A1 |
20190088106 | Grundstrom | Mar 2019 | A1 |
20190094824 | Xie et al. | Mar 2019 | A1 |
20190096217 | Pourmohammad et al. | Mar 2019 | A1 |
20190102840 | Perl et al. | Apr 2019 | A1 |
20190121801 | Jethwa et al. | Apr 2019 | A1 |
20190138512 | Pourmohammad et al. | May 2019 | A1 |
20190147883 | Mellenthin et al. | May 2019 | A1 |
20190158309 | Park et al. | May 2019 | A1 |
20190163152 | Worrall et al. | May 2019 | A1 |
20190268178 | Fairweather et al. | Aug 2019 | A1 |
20190278777 | Malik | Sep 2019 | A1 |
20190310979 | Masuzaki et al. | Oct 2019 | A1 |
20190377306 | Harvey | Dec 2019 | A1 |
20200133978 | Ramamurti et al. | Apr 2020 | A1 |
20200226156 | Borra et al. | Jul 2020 | A1 |
20200285203 | Thakur et al. | Sep 2020 | A1 |
20200336328 | Harvey | Oct 2020 | A1 |
20200348632 | Harvey | Nov 2020 | A1 |
20200387576 | Brett et al. | Dec 2020 | A1 |
20200396208 | Brett et al. | Dec 2020 | A1 |
20210042299 | Migliori | Feb 2021 | A1 |
20210043221 | Yelchuru et al. | Feb 2021 | A1 |
20210133670 | Cella | May 2021 | A1 |
20210200171 | Sridharan et al. | Jul 2021 | A1 |
20210200174 | Sridharan et al. | Jul 2021 | A1 |
20210200713 | Sridharan et al. | Jul 2021 | A1 |
20210279957 | Eder | Sep 2021 | A1 |
20210325070 | Endel et al. | Oct 2021 | A1 |
20210342961 | Winter et al. | Nov 2021 | A1 |
20210356153 | Nesler et al. | Nov 2021 | A1 |
20210373509 | Borah et al. | Dec 2021 | A1 |
20210373510 | Borah et al. | Dec 2021 | A1 |
20210381711 | Harvey et al. | Dec 2021 | A1 |
20210381712 | Harvey et al. | Dec 2021 | A1 |
20210382445 | Harvey et al. | Dec 2021 | A1 |
20210383041 | Harvey et al. | Dec 2021 | A1 |
20210383042 | Harvey et al. | Dec 2021 | A1 |
20210383200 | Harvey et al. | Dec 2021 | A1 |
20210383219 | Harvey et al. | Dec 2021 | A1 |
20210383235 | Harvey et al. | Dec 2021 | A1 |
20210383236 | Harvey et al. | Dec 2021 | A1 |
20220004672 | Santarone | Jan 2022 | A1 |
20220028159 | Vincent | Jan 2022 | A1 |
20220066402 | Harvey et al. | Mar 2022 | A1 |
20220066405 | Harvey | Mar 2022 | A1 |
20220066415 | Miller | Mar 2022 | A1 |
20220066432 | Harvey et al. | Mar 2022 | A1 |
20220066434 | Harvey et al. | Mar 2022 | A1 |
20220066528 | Harvey et al. | Mar 2022 | A1 |
20220066722 | Harvey et al. | Mar 2022 | A1 |
20220066754 | Harvey et al. | Mar 2022 | A1 |
20220066761 | Harvey et al. | Mar 2022 | A1 |
20220067226 | Harvey et al. | Mar 2022 | A1 |
20220067227 | Harvey et al. | Mar 2022 | A1 |
20220067230 | Harvey et al. | Mar 2022 | A1 |
20220069863 | Harvey et al. | Mar 2022 | A1 |
20220070293 | Harvey et al. | Mar 2022 | A1 |
20220082280 | Douglas et al. | Mar 2022 | A1 |
20220121965 | Chatterji et al. | Apr 2022 | A1 |
20220138684 | Harvey | May 2022 | A1 |
20220187818 | Kondejkar | Jun 2022 | A1 |
20220215264 | Harvey et al. | Jul 2022 | A1 |
20230010757 | Preciado | Jan 2023 | A1 |
20230071312 | Preciado et al. | Mar 2023 | A1 |
20230076011 | Preciado et al. | Mar 2023 | A1 |
20230083703 | Meiners | Mar 2023 | A1 |
Number | Date | Country |
---|---|---|
2019226217 | Nov 2020 | AU |
2019226264 | Nov 2020 | AU |
2019351573 | May 2021 | AU |
101415011 | Apr 2009 | CN |
102136099 | Jul 2011 | CN |
102136100 | Jul 2011 | CN |
102650876 | Aug 2012 | CN |
104040583 | Sep 2014 | CN |
104603832 | May 2015 | CN |
104919484 | Sep 2015 | CN |
106204392 | Dec 2016 | CN |
106406806 | Feb 2017 | CN |
106960269 | Jul 2017 | CN |
107147639 | Sep 2017 | CN |
107598928 | Jan 2018 | CN |
111860900 | Oct 2020 | CN |
110046316 | Dec 2020 | CN |
2 528 033 | Nov 2012 | EP |
3 268 821 | Jan 2018 | EP |
3 324 306 | May 2018 | EP |
H10-049552 | Feb 1998 | JP |
2003-162573 | Jun 2003 | JP |
2007-018322 | Jan 2007 | JP |
4073946 | Apr 2008 | JP |
2008-107930 | May 2008 | JP |
2013-152618 | Aug 2013 | JP |
2014-044457 | Mar 2014 | JP |
20160102923 | Aug 2016 | KR |
WO-2009020158 | Feb 2009 | WO |
WO-201 1100255 | Aug 2011 | WO |
WO-201 3050333 | Apr 2013 | WO |
WO-2015106702 | Jul 2015 | WO |
WO-201 5145648 | Oct 2015 | WO |
WO-201 7035536 | Mar 2017 | WO |
WO-2017192422 | Nov 2017 | WO |
WO-2017194244 | Nov 2017 | WO |
WO-2017205330 | Nov 2017 | WO |
WO-2017213918 | Dec 2017 | WO |
WO-2018132112 | Jul 2018 | WO |
WO-2020061621 | Apr 2020 | WO |
WO-2022042925 | Mar 2022 | WO |
Entry |
---|
Ettinger et al., U.S. Pat. No. 11,216,663B1, Jan. 2022. See the short version. |
Vincent et al., U.S. Patent Application Publication 2022/0028159, Jan. 2022, see the short version. |
Santarone et al., U.S. Patent Application Publication 2022/0004672, Jan. 2022, see the short version. |
Balaji et al., “Brick: Metadata schema for portable smart building applications,” Applied Energy, 2018 (20 pages). |
Balaji et al., “Brick: Metadata schema for portable smart building applications,” Applied Energy, Sep. 15, 2018, 3 pages, (Abstract). |
Balaji et al., “Demo Abstract: Portable Queries Using the Brick Schema for Building Applications,” BuildSys '16, Palo Alto, Ca, USA, Nov. 16-17, 2016 (2 pages). |
Balaji, B. et al., “Brick: Towards a Unified Metadata Schema For Buildings.” BuildSys '16, Palo Alto, Ca, USA, Nov. 16-17, 2016 (10 pages). |
Bhattacharya et al., “Short Paper: Analyzing Metadata Schemas for Buildings - The Good, The Bad and The Ugly,” BuildSys '15, Seoul, South Korea, Nov. 4-5, 2015 (4 pages). |
Bhattacharya, A., “Enabling Scalable Smart-Building Analytics,” Electrical Engineering and Computer Sciences, University of California at Berkeley, Technical Report No. UCB/EECS-2016-201, Dec. 15, 2016 (121 pages). |
Brick, “Brick Schema: Building Blocks for Smart Buildings,” URL: chrome-extension://efaidnbmnnnibpcajpcglclefindmkaj/https://www. memoori. com/wp-content/uploads/2016/06/Brick_Schema_Whitepaper.pdf, Mar. 2019 (17 pages). |
Brick, “Brick: Towards a Unified Metadata Schema For Buildings,” URL: chrome-extension://efaidnbmnnnibpcajpcglclefindmkaj/https://brickschema.org/papers/Brick_BuildSys_Present ation.pdf, Presented at BuildSys '16, Nov. 2016 (46 pages). |
Brick, “Metadata Schema for Buildings,” URL: https://brickschema.org/docs/Brick-Leaflet.pdf, retrieved from internet Dec. 24, 2019 (3 pages). |
Chinese Office Action on CN Appl. No. 201780003995.9 dated Apr. 8, 2021 (21 pages with English language translation). |
Chinese Office action on CN Appl. No. 201780043400.2 dated Apr. 25, 2021 (15 pages with English language translation). |
Curry, E. et al., “Linking building data in the cloud: Integrating cross-domain building data using linked data.” Advanced Engineering Informatics, 2013, 27 (pp. 206-219). |
Digital Platform Litigation Documents Part 1, includes cover letter, dismissal of case DDE-1 -21 -cv-01796, IPR2023-00022 (documents filed 1/26/23-Oct. 7, 22), and IPR2023-00085 (documents filed 1/26/Oct. 23, 20/22) (748 pages total). |
Digital Platform Litigation Documents Part 10, includes DDE-1-21-cv-01796 (documents filed 11/Jan. 22, 12/22/21 (1795 pages total). |
Digital Platform Litigation Documents Part 2, includes IPR2023-00085 (documents filed Oct. 20, 22) (172 pages total). |
Digital Platform Litigation Documents Part 3, includes IPR2023-00085 (documents filed Oct. 20, 22) and IPR2023-00170 (documents filed 11/28/22-Nov. 7, 22) (397 pages total). |
Digital Platform Litigation Documents Part 4, includes IPR2023-00170 (documents filed Nov. 7, 22) and IPR2023-00217 (documents filed 1/18/2023-Nov. 15, 22) (434 pages total). |
Digital Platform Litigation Documents Part 5, includes IPR2023-00217 (documents filed Nov. 15, 22) and IPR2023-00257 (documents filed 1/25/Nov. 23, 23/22) (316 pages total). |
Digital Platform Litigation Documents Part 6, includes IPR2023-00257 (documents filed Nov. 23, 22) and IPR 2023-00346 (documents filed 1/Mar. 23, 12/13/22) (295 pages total). |
Digital Platform Litigation Documents Part 7, includes IPR 2023-00346 (documents filed Dec. 13, 22) and IPR2023-00347 (documents filed 1/Mar. 23, 12/13/22) (217 pages total). |
Digital Platform Litigation Documents Part 8, includes IPR2023-00347 (documents filed Dec. 13, 22), EDTX-2-22-cv-00243 (documents filed 9/20/Jun. 22, 29/22), and DDE-1-21-cv-01796 (documents filed 2/3/Jan. 23, 10/23 (480 pages total). |
Digital Platform Litigation Documents Part 9, includes DDE-1-21-cv-01796 (documents filed 1/Oct. 23, 11/1/22 (203 pages total). |
El Kaed, C. et al., “Building management insights driven by a multi-system semantic representation approach,” 2016 IEEE 3rd World Forum on Internet of Things (WF-loT), Dec. 12-14, 2016, (pp. 520-525). |
Ellis, C. et al., “Creating a room connectivity graph of a building from per-room sensor units.” BuildSys '12, Toronto, ON, Canada, Novembers, 2012 (7 pages). |
Extended European Search Report on EP Application No. 18196948.6 dated Apr. 10, 2019 (9 pages). |
Fierro et al., “Beyond a House of Sticks: Formalizing Metadata Tags with Brick,” BuildSys '19, New York, Ny, USA, Nov. 13-14, 2019 (10 pages). |
Fierro et al., “Dataset: An Open Dataset and Collection Tool for BMS Point Labels,” DATA'19, New York, Ny, USA, Nov. 10, 2019 (3 pages). |
Fierro et al., “Design and Analysis of a Query Processor for Brick,” ACM Transactions on Sensor Networks, Jan. 2018, vol. 1, No. 1, art. 1 (25 pages). |
Fierro et al., “Design and Analysis of a Query Processor for Brick,” BuildSys '17, Delft, Netherlands, Nov. 8-9, 2017 (10 pages). |
Fierro et al., “Mortar: An Open Testbed for Portable Building Analytics,” BuildSys '18, Shenzhen, China, Nov. 7-8, 2018 (10 pages). |
Fierro et al., “Why Brick is a Game Changer for Smart Buildings,” URL: https://brickschema.org/papers/Brick_Memoori_Webinar_Presentation.pdf, Memoori Webinar, 2019 (67 pages). |
Fierro, “Writing Portable Building Analytics with the Brick Metadata Schema,” UC Berkeley, ACM E-Energy, 2019 (39 pages). |
Fierro, G., “Design of an Effective Ontology and Query Processor Enabling Portable Building Applications,” Electrical Engineering and Computer Sciences, University of California at Berkeley, Technical Report No. UCB/EECS-2019-106, Jue 27, 2019 (118 pages). |
File History for U.S. Appl. No. 12/776,159, filed May 7, 2010 (722 pages). |
Final Conference Program, ACM BuildSys 2016, Stanford, Ca, USA, Nov. 15-17, 2016 (7 pages). |
Gao et al., “A large-scale evaluation of automated metadata inference approaches on sensors from air handling units,” Advanced Engineering Informatics, 2018, 37 (pp. 14-30). |
Harvey, T., “Quantum Part 3: The Tools of Autonomy, How PassiveLogic's Quantum Creator and Autonomy Studio software works,” URL: https://www.automatedbuildings.com/news/Jan. 22/articles/passive/211224010000passive.html, Jan. 2022 (7 pages). |
Harvey, T., “Quantum: The Digital Twin Standard for Buildings,” URL: https://www.automatedbuildings.com/news/Feb. 21/articles/passivelogic/210127124501passivelogic.html, Feb. 2021 (6 pages). |
Hu, S et al., “Building performance optimisation: A hybrid architecture for the integration of contextual information and time-series data,” Automation in Construction, 2016, 70 (pp. 51-61). |
International Search Report and Written Opinion for PCT Appl. Ser. No. PCT/US2017/013831 dated Mar. 31, 2017 (14 pages). |
International Search Report and Written Opinion for PCT Appl. Ser. No. PCT/US2017/035524 dated Jul. 24, 2017 (14 pages). |
International Search Report and Written Opinion on PCT/US2017/052060, dated Oct. 5, 2017, 11 pages. |
International Search Report and Written Opinion on PCT/US2017/052633, dated Oct. 23, 2017, 9 pages. |
International Search Report and Written Opinion on PCT/US2017/052829, dated Nov. 27, 2017, 24 pages. |
International Search Report and Written Opinion on PCT/US2018/024068, dated Jun. 15, 2018, 22 pages. |
International Search Report and Written Opinion on PCT/US2018/052971, dated Mar. 1, 2019, 19 pages. |
International Search Report and Written Opinion on PCT/US2018/052974, dated Dec. 19, 2018, 13 pages. |
International Search Report and Written Opinion on PCT/US2018/052975, dated Jan. 2, 2019, 13 pages. |
International Search Report and Written Opinion on PCT/US2018/052994, dated Jan. 7, 2019, 15 pages. |
International Search Report and Written Opinion on PCT/US2019/015481, dated May 17, 2019, 78 pages. |
International Search Report and Written Opinion on PCT/US2020/058381, dated Jan. 27, 2021,30 pages. |
Japanese Office Action on JP Appl. No. 2018-534963 dated May 11, 2021 (16 pages with English language translation). |
Koh et al., “Plaster: An Integration, Benchmark, and Development Framework for Metadata Normalization Methods,” BuildSys '18, Shenzhen, China, Nov. 7-8, 2018 (10 pages). |
Koh et al., “Scrabble: Transferrable Semi-Automated Semantic Metadata Normalization using Intermediate Representation,” BuildSys '18, Shenzhen, China, Nov. 7-8, 2018 (10 pages). |
Koh et al., “Who can Access What, and When?” BuildSys '19, New York, Ny, USA, Nov. 13-14, 2019 (4 pages). |
Li et al., “Event Stream Processing with Out-of-Order Data Arrival,” International Conferences on Distributed Computing Systems, 2007, (8 pages). |
Nissin Electric Co., Ltd., “Smart power supply system (SPSS),” Outline of the scale verification plan, Nissin Electric Technical Report, Japan, Apr. 23, 2014, vol. 59, No. 1 (23 pages). |
PassiveLogic, “Explorer: Digital Twin Standard for Autonomous Systems. Made interactive.” URL: https://passivelogic.com/software/quantum-explorer/, retrieved from internet Jan. 4, 2023 (13 pages). |
PassiveLogic, “Quantum: The Digital Twin Standard for Autonomous Systems, A physics-based ontology for next-generation control and AI.” URL: https://passivelogic.com/software/quantum-standard/, retrieved from internet Jan. 4, 2023 (20 pages). |
Quantum Alliance, “Quantum Explorer Walkthrough,” 2022, (7 pages) (screenshots from video). |
Results of the Partial International Search for PCT/US2018/052971, dated Jan. 3, 2019, 3 pages. |
Sinha, Sudhi and Al Huraimel, Khaled, “Reimagining Businesses with Al” John Wiley & Sons, Inc., Hoboken, Nj, USA, 2021 (156 pages). |
Sinha, Sudhi R. and PARK, Youngchoon, “Building an Effective loT Ecosystem for Your Business,” Johnson Controls International, Springer International Publishing, 2017 (286 pages). |
Sinha, Sudhi, “Making Big Data Work For Your Business: A guide to effective Big Data analytics,” Impackt Publishing LTD., Birmingham, UK, Oct. 2014 (170 pages). |
The Virtual Nuclear Tourist, “Calvert Cliffs Nuclear Power Plant,” URL: http://www.nucleartourist.com/US/calvert.htm, Jan. 11, 2006 (2 pages). |
University of California At Berkeley, Eecs Department, “Enabling Scalable Smart-Building Analytics,” URL: https://www2.eecs.berkeley.edu/Pubs/TechRpts/2016/EECS-2016-201.html, retrieved from internet Feb. 15, 2022 (7 pages). |
Van Hoof, Bert, “Announcing Azure Digital Twins: Create digital replicas of spaces and infrastructure using cloud, Al and loT,” URL: https://azure.microsoft.com/en-US/blog/announcing-azure-digital-twins-create-digital-replicas-of-spaces-and-infrastructure-using-cloud-ai-and-iot/, Sep. 24, 2018 (11 pages). |
W3C, “Sparql: Query Language for RDF,” located on The Wayback Machine, URL: https://web.archive.Org/web/20l61230061728/http://www.w3.org/TR/rdf-sparql-query/), retrieved from internet Nov. 15, 2022 (89 pages). |
Wei et al., “Development and Implementation of Software Gateways of Fire Fighting Subsystem Running on EBI,” Control, Automation and Systems Engineering, IITA International Conference on, IEEE, Jul. 2009 (pp. 9-12). |
White et al., “Reduce building maintenance costs with AWS loT TwinMaker Knowledge Graph,” The Internet of Things on AWS—Official Blog, URL: https://aws.amazon.com/blogs/iot/reduce-building-maintenance-costs-with-aws-iot-twinmaker-knowledge-graph/, Nov. 18, 2022 (10 pages). |
Zhou, Q. et al., “Knowledge-infused and Consistent Complex Event Processing over Real-time and Persistent Streams,” Further Generation Computer Systems, 2017, 76 (pp. 391-406). |
U.S. Appl. No. 17,566,029, PassiveLogic, Inc. |
U.S. Appl. No. 17,567,275, PassiveLogic, Inc. |
U.S. Appl. No. 17,722,115, PassiveLogic, Inc. |
Number | Date | Country | |
---|---|---|---|
20230169220 A1 | Jun 2023 | US |