This application relates to the field of building systems and, more particularly, to systems and methods for fault detection and diagnostics of building automation systems.
Building automation systems encompass a wide variety of systems that aid in the monitoring and control of building operations. In particular, building automation systems detect faults and other error or abnormal conditions in a building automation system. To use data points (e.g., room temperature sensors and pressure sensors) for fault detection, it is important to understand the actual task of a data point within a building, such as one data point for measuring outside air temperature and another data point for measuring supply air temperature.
These tasks of data points are represented by data point labels. For fault detection and diagnostics (FDD) of a building automation system, the system must know the labels of the data points within a building in order to understand its task. This task of labeling is done manually and can be difficult for installation and maintenance personnel. The task is also error prone, especially if the data point names in the building are not standardized. Accordingly, conventional systems fail to determine whether assigned labels adequately characterize data points, resulting in problems with analysis and diagnostics task.
In accordance with one embodiment of the disclosure, there is provided a label plausibility approach for building automation systems. The approach is independent of naming conventions to check whether data points are understood correctly by the system and labelled accurately. The checking procedure of the label plausibility approach determines whether the designated label of a data point corresponds to its actual task by checking whether the corresponding timeseries data is plausible for that label.
One aspect is a method for fault detection and diagnostics of a building automation system. Timeseries data are received from the building automation system. A label plausibility is determined for each set of timeseries data and the corresponding label associated with the set of timeseries data based on a tree-based classifier and an image transformation classifier. The tree-based classifier and the image transformation classifier receive the same data input and operating distinctly from each other.
Another aspect is a system for fault detection and diagnostics of a building automation system comprising an input component and a processor. The input component receives timeseries data from the building automation system. The processor determines a label plausibility for each set of timeseries data and the corresponding label associated with the set of timeseries data based on a tree-based classifier and an image transformation classifier. The tree-based classifier and the image transformation classifier receive the same data input and operating distinctly from each other.
The above-described features and advantages, as well as others, will become more readily apparent to those of ordinary skill in the art by reference to the following detailed description and accompanying drawings. While it would be desirable to provide one or more of these or other advantageous features, the teachings disclosed herein extend to those embodiments which fall within the scope of the appended claims, regardless of whether they accomplish one or more of the above-mentioned advantages.
For a more complete understanding of the present disclosure, and the advantages thereof, reference is now made to the following descriptions taken in conjunction with the accompanying drawings, wherein like numbers designate like objects.
Various technologies that pertain to systems and methods that facilitate label plausibility checking will now be described with reference to the drawings, where like reference numerals represent like elements throughout. The drawings discussed below, and the various embodiments used to describe the principles of the present disclosure in this patent document are by way of illustration only and should not be construed in any way to limit the scope of the disclosure. Those skilled in the art will understand that the principles of the present disclosure may be implemented in any suitably arranged apparatus. It is to be understood that functionality that is described as being carried out by certain system elements may be performed by multiple elements. Similarly, for instance, an element may be configured to perform functionality that is described as being carried out by multiple elements. The numerous innovative teachings of the present application will be described with reference to exemplary non-limiting embodiments.
Referring to
For the illustrated embodiment of
Referring to
The operation 200 represents a method for fault detection and diagnostics of the building automation system 100. A processor, which may include cooperative processors, of the management station 104, 106, 108 receive (202) timeseries data from one or more field devices 124, 126 of the building automation system 100. For some embodiments, the management station 104, 106, 108 may parse (204) the timeseries data into predetermined time periods, such as time periods associated with a calendar period type (days, weeks, months, years, etc.). For some embodiments, the operation 200 may extract (206) statistical features for the timeseries data. For example, hundreds of statistical features 208, 210 for each individual time period of the predetermined time periods may be extracted. For some embodiments, the processor of the management station 104, 106, 108 may identify (212) a data type associated with one or more field devices of the building automation system. For example, the data type may be determined based on the type of field device providing the timeseries data. Often, the data types may be determined by a manual process that sometimes results in wrong labels that the operation would like to detect.
In response to receiving (202) the timeseries data, the processor of management station 104, 106, 108 determines (214, 216, 218, 220, 222) a label plausibility for each set of timeseries data and the corresponding label associated with the set of timeseries data. The operation 200 determine the label plausibility based on multiple classifiers including a tree-based classifier 224 and an image transformation classifier 226. The tree-based classifier 224 and the image transformation classifier 226 receive the same data input and operating distinctly from each other. In this manner, the management station 104, 106, 108 operates two separate subprocesses 224, 226 that are orchestrated, and thus combined, to determine the label plausibility. The subprocesses 224, 226 are orchestrated to one ensemble algorithm.
As stated above, the subprocesses include the tree-based classifier 224 and the image transformation classifier 226. Based on the tree-based classifier 224, the processor of the management station 104, 106, 108 determines the label plausibility by, in part, generating (214) a first probability corresponding to an association between each set of timeseries data and the corresponding label. For example, the first probability may indicate a probability of how well the label fits the corresponding timeseries data. For some embodiments, the tree-based classifier may be a random forest classifier in which the statistical features 208, 210 may serve as inputs to the random forest classifier. The result of the tree-based classifier is a first probability associated with how well the timeseries data corresponds to the currently assigned label. For some embodiments, if the first probability is low, then the label of the datapoint may be wrong or some abnormal behavior may be exhibited. The first probability may be determined to be low if the probability fails to exceed a predetermined first probability threshold.
Based on the image transformation classifier 226, the processor of the management station 104, 106, 108 determines the label plausibility by, in part, generating (216-222) a second probability corresponding to an association between each set of timeseries data and the corresponding label. For example, the second probability of how well the label fits the corresponding timeseries data. The management station 104, 106, 108 may initiate generation (216-222) of the second probability by transforming (216) each set of timeseries data to an RGB image 228, 230. In response to transforming (216) to timeseries data, the management station 104, 106, 108 performs imaging processing 232 to process the transformed images at various stages (218-222) of generating (216-222) the second probability. In particular, when generating (216-222) the second probability, the management station 104, 106, 108 applies (218, 222) one or more neural networks to the transformed RGB images. For some embodiments, the neural networks include a convolutional neural network (CNN) and, thereafter, a multi-layer perceptron. For example, the CNN may be built on an Alexnet architecture and extract the most relevant information of the images (218). The result of this process is some kind of image, that we are combining with the most important statistical features (220). This image may be made to one flat vector and used in a multi-layer-perceptron. For some embodiments, the management station 104, 106, 108 may combine (220) one or more statistical features with each transformed RGB image. For some embodiments, the image transformation classifier includes data preprocessing of a Markov-Transition-Field transformation, a Gramian-Angular-Field transformation, or both.
The processor of the management station 104, 106, 108 determines the label plausibility by determining (234) the label plausibility based on a first probability of the tree-based classifier and a second probability of the image transformation classifier. The label plausibility may be represented in various ways indicating how well the current function label corresponds to the actual timeseries data. For example, the label plausibility may include a percentage or ratio representing how well each label associates with the timeseries data corresponding to the label. In response to determining (214, 216, 218, 220, 222) the label plausibility, the processor of the management station 104, 106, 108 may initiate a process to modify one or more labels based on the label plausibility, output the label plausibility at an output device, send the label plausibility information to a remote device, and the like.
The operation 200 for the label plausibility checker automatically analyzes the incoming timeseries data and checks the corresponding label using a statistical model or a machine learning system based on artificial intelligence. For a machine learning system, a classifier may be trained to learn how a timeseries of a certain data type may appear. Subsequent to training, the system may check timeseries data during operation to determine whether the data fits the learned pattern. Accordingly, each of the tree-based classifier and the image transformation classifier may be trained to identify the pattern associated each set of timeseries data and the corresponding label, for a particular data type. For some embodiments, the classifiers may be trained by inputting positive examples and negative examples to the classifier. For example, the positive examples may include data points of the same label and the negative examples may include data points of different labels.
Both classifiers or models may be pretrained on a subset of all available datapoints of the building automation system, which may be associated with one, multiple, or all buildings. The classifiers learn the appropriate positive and negative examples of timeseries data for each designated label. Each classifier learns to differentiate the current label of a set of data points from all other labels and learned how a typical timeseries of the label should correspond. Since we got 193 different types of sensors, we also got 193 labels. A separate classifier has been trained for each type of data point so different classifiers are used for the tree-based classifier 224 and the image transformation classifier 226.
Referring to
In response to transforming (216) to timeseries data, the management station 104, 106, 108 performs image processing 232 to process the transformed images at various stages (218-222) of generating (216-222) the second probability. In particular, when generating (216-222) the second probability, the operation applies (218, 222) one or more neural networks to the transformed RGB images. For some embodiments, the neural networks include a Convolutional Neural Network (CNN) and, thereafter, a multi-layer perceptron. For example, one neural network may be a CNN with Alexnet-like architecture. For some embodiments, the CNN 302, based on a pretrained Alexnet architecture, may be combined with a Multi-Layer Perceptron 304 to get a percentage how plausible the label is for this datapoint. For some embodiments, the management station 104, 106, 108 may combine (220) one or more statistical features 306 with each transformed RGB image. For such embodiments, the images 228, 230 of each set of timeseries data may correspond to neural network output 308 and a fully connected layer 310. Then, the most important statistical features 306 may be combined with the image and multiple fully connected layers 304 may be formed. The fully connected layers 304 operate on a flattened input where each input is connected to all neurons, which may be used to optimize objectives 312.
Referring to
Referring to
The processor or processors 606 may send data to, and process commands received from, other components of the device components 600, such as information of the communication component 604 or the memory component 608. Each application includes executable code to provide specific functionality for the processor 606 and/or remaining components of the management device 104, 106, 108. Examples of applications executable by the processor 606 include, but are not limited to, a tree-based module 610 and a BERND module 612.
Data stored at the memory component 608 is information that may be referenced and/or manipulated by a module of the processor 606 for performing functions of the management stations 104, 106, 108. Examples of data associated with the management station 104, 106, 108 and stored by the memory component 608 may include, but are not limited to, statistical features and images 614 and plausibility data 616.
The device components 600 may include an input component 618 that manages one or more input components and/or an output component 620 that manages one or more output components. The input components 618 and output components 620 of the device components 600 may include one or more visual, audio, mechanical, and/or other components. For some embodiments, the input and output components 618, 620 may include a user interface 622 for interaction with a user of the device. The user interface 622 may include a combination of hardware and software to provide a user with a desired user experience.
It is to be understood that
Those skilled in the art will recognize that, for simplicity and clarity, the full structure and operation of all data processing systems suitable for use with the present disclosure are not being depicted or described herein. Also, none of the various features or processes described herein should be considered essential to any or all embodiments, except as described herein. Various features may be omitted or duplicated in various embodiments. Various processes described may be omitted, repeated, performed sequentially, concurrently, or in a different order. Various features and processes described herein can be combined in still other embodiments as may be described in the claims.
It is important to note that while the disclosure includes a description in the context of a fully functional system, those skilled in the art will appreciate that at least portions of the mechanism of the present disclosure are capable of being distributed in the form of instructions contained within a machine-usable, computer-usable, or computer-readable medium in any of a variety of forms, and that the present disclosure applies equally regardless of the particular type of instruction or signal bearing medium or storage medium utilized to actually carry out the distribution. Examples of machine usable/readable or computer usable/readable mediums include nonvolatile, hard-coded type mediums such as read only memories (ROMs) or erasable, electrically programmable read only memories (EEPROMs), and user-recordable type mediums such as floppy disks, hard disk drives and compact disk read only memories (CD-ROMs) or digital versatile disks (DVDs).
Although an example embodiment of the present disclosure has been described in detail, those skilled in the art will understand that various changes, substitutions, variations, and improvements disclosed herein may be made without departing from the spirit and scope of the disclosure in its broadest form.