The present disclosure relates to anomaly detection using artificial intelligence, and in particular to detecting physical anomalies in image data.
Physical assets, such as buildings, equipment, etc., require monitoring for security, safety, operational efficiencies, etc. Typically, such physical assets are monitored by sensors and/or by visual inspection, which have several limitations and drawbacks.
For example, businesses often have several real estate assets that contain electrical/mechanical equipment, such as generator rooms, battery rooms, HVAC rooms, machine rooms, etc. One example of a major concern relating to these rooms is the possibility of a leak occurring, leading to contamination. To monitor for leaks, many rooms are equipped with physical sensors that are configured to send alerts to an operator. However, the sensors rely upon a lot of manual configuration, are limited in terms of tracked surface area, and may not always trigger. The use of cameras has been considered to detect leaks, but these likewise require manual configuration of alerts. Accordingly, businesses will send technicians periodically to these rooms to look for leaks and other issues, which is both time consuming and inefficient, particularly since a leak may have occurred well before the technician checks on the room and can result in significant damage.
Current techniques for monitoring physical assets therefore rely on manual and time-consuming efforts. Accordingly, systems and methods that enable anomaly detection of physical assets remain highly desirable.
In accordance with one aspect of the present disclosure, an anomaly detection method is disclosed, comprising: receiving image data from one or more cameras configured to capture an image of a physical asset; determining a probability of a physical anomaly being present in the image data using an artificial intelligence model that is trained to detect anomalous image data; and outputting an alert when the probability of the physical anomaly being present in the image data exceeds a threshold value.
In some aspects, the method further comprises: analyzing the image data to determine that the physical anomaly is a specific type of anomaly; and outputting the alert including information on the specific type of anomaly present in the image data.
In some aspects, analyzing the image data to determine that the physical anomaly is the specific type of anomaly comprises determining a probability that the physical anomaly is the specific type of anomaly using one or more secondary artificial intelligence models that are trained to predict the physical anomaly as being one or more types of anomalies.
In some aspects, the one or more secondary artificial intelligence models comprise a model that is trained to perform image segmentation on the image data to classify the anomaly as being the specific type of anomaly.
In some aspects, the one or more secondary artificial intelligence models comprise a multi-modal generative AI model.
In some aspects, the method further comprises receiving a context of the image data, wherein the multi-modal generative AI model uses the context of the image data to determine the probability that the physical anomaly is the specific type of anomaly.
In some aspects, the method further comprises receiving audio data and/or vibration data associated with the physical asset, wherein the multi-modal generative AI model uses the received audio and/or vibration data to determine the probability that the physical anomaly is the specific type of anomaly.
In some aspects, the multi-modal generative AI model is configured to generate an output comprising one or both of: a description of the specific type of anomaly, and a suggested action for responding to the specific type of anomaly.
In some aspects, determining that the physical anomaly is the specific type of anomaly comprises applying one or more rules to the image data.
In some aspects, the method further comprises receiving auxiliary data associated with the physical asset from one or more sensors, and wherein determining that the physical anomaly is the specific type of anomaly is further based on the auxiliary data.
In some aspects, analyzing the image data to determine that the physical anomaly is the specific type of anomaly is performed automatically when the probability that the physical anomaly is present in the image data exceeds the threshold value.
In some aspects, analyzing the image data to determine that the physical anomaly is a specific type of anomaly is performed in response to a prompt to identify the specific type of anomaly.
In some aspects, the specific type of anomaly is one of: equipment overheat, presence of a human, presence of an animal, misplaced tools or equipment, fire, flood, falling material, and leaks.
In some aspects, the method further comprises receiving user feedback on the specific type of anomaly, and updating the one or more secondary artificial intelligence models based on the user feedback.
In some aspects, the method further comprises, in response to outputting the alert, receiving user feedback that the image data is normal, and updating the artificial intelligence model based on the user feedback.
In some aspects, the one or more cameras are configured to capture an image of an area of a building, an area of a site, or a piece of equipment.
In some aspects, the building comprises one of: a warehouse, a hospital, a mechanical room, a server room, and an electrical room.
In some aspects, the site comprises one of: an underground cable tunnel, a manufacturing site, an industrial site, and a mining site.
In some aspects, the piece of equipment is one of: a generator, a battery, a machine, a utility cabinet, and a heating, ventilation, and air conditioning (HVAC) unit.
In some aspects, the image data is received as a single image or a stream of images.
In some aspects, the image data is received in association with metadata comprising one or more of a camera identifier and a camera location.
In some aspects, the one or more cameras comprise one or more of: thermal cameras, near-infrared cameras, and RGB cameras.
In some aspects, the physical anomaly is detected as a thermal anomaly in the image data or a visual anomaly in the image data.
In accordance with another aspect of the present disclosure, an anomaly detection system is disclosed, comprising: a processor; and a non-transitory computer-readable memory having stored thereon computer-executable instructions which, when executed by the processor, configure the anomaly detection system to perform the anomaly detection method of any one of the above aspects.
In accordance with another aspect of the present disclosure, a method of training an anomaly detection model is disclosed, comprising: obtaining training images comprising normal image data; and training an artificial intelligence model to determine a probability of a physical anomaly being present in the image data.
In some aspects, the method further comprises: obtaining training images comprising anomalous image data that have a known anomaly; and training one or more secondary artificial intelligence models to determine the known anomaly present in the anomalous image data.
In some aspects, the one or more second artificial intelligence models comprise a multi-modal generative AI model, and the method further comprises: obtaining additional training data comprising an additional input type; and training the multi-modal generative AI model to determine the known anomaly present in the anomalous image data based on the training images comprising anomalous image data and the additional training data.
In some aspects, the method further comprises training the multi-modal generative AI model to generate outputs associated with the known anomaly in response to different input prompts.
In some aspects, the training images are obtained from one or more cameras.
In some aspects, the one or more cameras comprise one or more of: thermal cameras, near-infrared cameras, and RGB cameras.
In some aspects, the training images are obtained from a plurality of cameras, and wherein the artificial intelligence model is trained on the training images from each of the plurality of cameras.
In some aspects, the method further comprises generating the training images by recording images for a pre-set period of time under normal operating conditions.
In accordance with another aspect of the present disclosure, an anomaly detection model is disclosed that is trained in accordance with the method of training an anomaly detection model of any one of the above aspects.
Further features and advantages of the present disclosure will become apparent from the following detailed description, taken in combination with the appended drawings, in which:
It will be noted that throughout the appended drawings, like features are identified by like reference numerals.
In accordance with the present disclosure, anomaly detection systems and methods are disclosed for automatically detecting physical anomalies using image data. The anomaly detection systems and methods disclosed herein can be used to detect anomalies in physical assets, such as anomalies present within a building, at a site (e.g. a manufacturing/industrial/mining site), and/or anomalies associated with a piece of equipment. Image data is received from one or more cameras that are configured to capture image data of a physical asset. The cameras may be thermal cameras, near-infrared cameras, or RGB cameras. The image data is analyzed to determine a probability of an anomaly being present in the image using an artificial intelligence model that is trained to detect anomalous image data. The image data may be further analyzed to determine that the physical anomaly is a specific type of anomaly, which may be performed using one or more secondary artificial intelligence models to analyze the image data and/or a deterministic algorithm that applies one or more rules to the image data. An alert is output when the probability of the physical anomaly being present in the image exceeds a threshold. A method of training an anomaly detection model is also disclosed.
Advantageously, systems and methods in accordance with the present disclosure allow for near-real time detection of anomalies, can identify small anomalies within a wide imaging area, can detect various types of anomalies using the same system, and requires no physical installation apart from one or more cameras. In some aspects, a multi-modal generative AI model may be used to make predictions on specific types of anomalies based on multiple types of inputs, and can generate various outputs such as a description of the anomaly, a suggested action for responding to the anomaly, etc., in addition to outputting an alert when the anomalous image data is detected.
Embodiments are described below, by way of example only, with reference to
In accordance with the present disclosure, the server(s) 102 are configured to receive the image data from the cameras 132a, 132b, and perform an anomaly detection method to determine if an anomaly is present in the image data (e.g. classify the image data as anomalous or nominal) using an artificial intelligence model, thus determining the presence of an anomaly within the rooms 130a, 130b or the equipment therein. The rooms 130a, 130b may for example be electrical/mechanical rooms, but it will be appreciated that cameras can be installed in various physical locations to identify anomalies associated with physical assets, including but not limited to buildings or rooms of buildings such as a warehouse, a hospital, a mechanical room, a server room, an electrical room, etc., and/or sites such as an underground cable tunnel, a manufacturing site, an industrial site, a mining site, etc. The cameras 132a, 132b in this example representation are configured to capture image data associated with equipment in the rooms 130a, 130b. The rooms 130a, 130b may contain various equipment such as one or more generators, batteries, machines, utility cabinets, HVAC units, etc. It will be appreciated that equipment being monitored can vary according to the particular application/implementation. For example, a mining site may comprise equipment such as crushers, conveyors, etc., as well as other physical assets such as high-pressure shipping containers with oxygen, etc., that may be monitored. It will be appreciated that there may be multiple cameras configured to capture image data of the same physical asset and/or equipment, and that the cameras 132a, 132b can be configured to capture image data of a room with no equipment in it, and/or to capture image data of equipment outside of a physical room.
In accordance with the present disclosure, the server(s) 102 are configured to implement an anomaly detection method that comprises determining a probability of a physical anomaly being present in the image data using an artificial intelligence model that is trained to detect anomalous image data. By determining if an anomaly is present in the image data, a near real-time warning of the anomaly can be output without requiring manual inspection of the physical asset to first detect the anomaly. An anomaly could be anything present in the building, at the site, or associated with equipment that should not be there in a normal situation. For generator rooms, for example, an anomaly could be caused by a water leak, cooling unit leak, oil leak, injector leak, diesel probe leak, floor leak systems, etc. As described in more detail below, the artificial intelligence model that is trained to detect anomalous image data may be trained using image data representing normal operating conditions, and thus the artificial intelligence model can detect when anything outside of normal is present in the image data.
The systems and methods disclosed herein provided by the server(s) 102 can be further configured to determine/classify the physical anomaly as a specific type of anomaly, for example by using one or more secondary artificial intelligence models trained to predict one or more types of anomalies, and/or by using a deterministic algorithm that applies one or more rules to the image data. As an example, the one or more secondary artificial intelligence models may comprise a classifier such as a YOLO neural network classifier to compare the object in a detected anomaly zone with known objects and indicate if it's a human, an animal or other common objects. The one or more secondary artificial intelligence models may additionally or alternatively comprise a multi-modal generative AI model that can receive multiple types of inputs, and can also generate various outputs related to the anomaly prediction. In some aspects, audio data may also be received from an audio capture device such as a microphone (not shown) associated with the physical asset and used to predict the specific type of anomaly. In some aspects, vibration data may also be received from a vibration sensor (not shown) associated with the physical asset and used to predict the specific type of anomaly. Various prompts may be provided to the multi-modal generative AI model and used for prediction and/or specifying desired output data. The prompts may be stored at the server(s) 102 and automatically called or triggered depending on the anomaly detection. Additionally or alternatively, the server(s) 102 may also provide a user interface that a user can interact with to investigate alerts and provide prompts. The prompts may for example include a prompt comprising textual data describing the context of the image data, and/or prompts specifying desired output data. One or more of the secondary artificial intelligence models may be hosted on a separate server (e.g. server 104) and the server(s) 102 may communicate with the server 104 and run said model(s).
Additionally, auxiliary data may be received from one or more sensors associated with the physical asset, which may be used for predicting the specific type of anomaly. For example, the system may receive light switch data indicative of whether a light in the room 130a or 130b is turned on. The system may determine, using said secondary artificial intelligence model(s) and/or a deterministic algorithm, that based on a shape of the physical anomaly detected, in conjunction with the information that the light is on when it is normally off, the anomaly is a human present in the room. For example, a classifier may learn or be trained to identify that if a light is on, a human is present in the room and the anomaly detection system may filter out events while the light is on and thus only monitor anomalous activities while the light is off (i.e. under normal operations). Another example of using auxiliary data may comprise receiving data from an loT connectivity event hub. The system could detect that a user has scanned their card to access the room and therefore pause anomaly detection until the person leaves the room based on an event of door open and close. Another example may be that when scheduled maintenance is being performed on equipment, the anomaly detection system could consult a rules engine and establish how critical alerts are supposed to be handled during the course of maintenance.
It is worth noting that the present disclosure is concerned with detecting physical anomalies associated with physical assets (e.g. associated with a physical location such as a building, room, or site, and/or associated with equipment) by analysis of image data acquired by the cameras. That is to say, the present disclosure is not concerned with identifying anomalies introduced by the camera, i.e. during image acquisition or processing. Accordingly, the present disclosure is not concerned with detecting anomalies to determine if a camera is functioning properly, but rather to detect a physical anomaly associated with a physical asset captured in image data.
As described above, the server(s) 102 are configured to receive the image data from the cameras 132a, 132b, and to determine if an anomaly is present in the rooms 130a, 130b and/or associated with equipment in the rooms by analyzing the image data using an artificial intelligence model. The servers 102 each comprise a CPU 110, a non-transitory computer-readable memory 112, a non-volatile storage 114, an input/output interface 116, and graphical processing units (“GPU”) 118. The non-transitory computer-readable memory 112 comprises computer-executable instructions stored thereon at runtime which, when executed by the CPU 110, configure the server to perform an anomaly detection method on image data as described in more detail herein. The non-volatile storage 114 has stored on it computer-executable instructions that are loaded into the non-transitory computer-readable memory 112 at runtime. The input/output interface 116 allows the server to communicate with one or more external devices (e.g. via network 120), including cameras 132a, 132b, external server(s) 104, as well as an operator device 140. The non-transitory computer-readable memory 112 also comprises an artificial intelligence model that is trained to classify the image data as being normal or containing an anomaly. The non-transitory computer-readable memory 112 may also store one or more secondary artificial intelligence models that analyzes the image data to determine a specific type of anomaly present in the image data. The GPU 118 may be used to control a display and may be used to run the artificial intelligence model(s) to analyze the image data and determine a probability of an anomaly being present in the image. When the probability of the anomaly being present in the image exceeds a threshold value, the servers 102 may output an alert to operator device 140, such as by sending an email, text, call, other message, etc. Further, the server(s) 102 may run one or more secondary artificial intelligence model(s) to determine a specific type of anomaly present in the image data, which can also be output to the operator device 140. It will be appreciated that there may be multiple servers 102 implemented to perform the anomaly detection methods. Multiple servers 102 may be networked together and collectively perform the anomaly detection method using distributed computing. In some aspects, as noted above, one or more artificial intelligence models may be stored at a separate server 104. The computing components of the server 104 are similar to those shown for server(s) 102.
The cameras 132a, 132b also have hardware components for capturing the image data, processing the image data, and sending the image data to the servers 102. As shown for camera 132a, the hardware components include CPU 133, non-transitory computer-readable memory 134, non-volatile storage 135, and input/output interface 136. It is also possible that in certain implementations, the cameras 132a, 132b may be configured to analyze image data themselves before sending to the servers 102, and have appropriate processing capabilities thereon.
As shown in
Multiple techniques may be used to increase accuracy of the model. As one example, small variations of raw training data may be injected in order to induce noise and movement of the images ingested by the system. This makes the system more reliable on vibration that could occurs when a generator or other equipment is running. Various transformations to the image data may also be applied, such as brightness adjustment of a few levels up and down, constraints, etc., in order to cover a wider range of potential variation in the images captured by the cameras. Another approach to help filter out anomalies is to apply auto segmentation of the image to extract masks. This allows the system to increase the score of anomalies for events detected on the floor compared to other locations less probable for the type of events intended to be captured.
In accordance with the present disclosure, the first artificial intelligence model trained to detect anomalous image data is not necessarily trained to classify specific types of anomalies, but rather to learn what is considered normal image data, and to determine when there is anything anomalous in the image data that deviates from the normal image data. Of course, a stand-alone artificial intelligence model could be further trained to identify specific types of anomalies in addition to determining when image data is anomalous, and the training data captured at 202 and stored at 204 may comprise image data of examples of anomalies to help improve further anomaly detection, e.g. water dripping on the floor. However, training a model to determine specific types of anomalies generally rely upon a large amount of manually labelled data, and such image data may be difficult to obtain (for example, there may only be so many images of oil leaks of for generators within a specific operating environment). Training the first artificial intelligence model using normal image data such that the artificial intelligence model is trained to classify anything anomalous in the image data was found to improve the accuracy of detecting any anomaly while requiring less training data. In an example implementation, the artificial intelligence model may be an algorithm supported by Anomalib. Numerous algorithms available within the Anomalib package were evaluated, and CFA was found to be suitable in light of its memory consumption and computer usage that allowed for scalability. PADIM was also initially considered to be a suitable algorithm, but the memory requirements increased exponentially as the number of supported points of view were trained.
One or more secondary artificial intelligence models may be trained to predict/determine that an anomaly is a specific type of anomaly. The specific types of anomalies can be determined through training of the one or more secondary artificial intelligence models over time, based on manual labelling of image data that has been identified to contain a physical anomaly. The training data captured at 202 and stored at 204 may thus comprise image data representative of various anomalies and be used to train the one or more secondary artificial intelligence models to classify specific types of anomalies. As an example, a YOLO classifier may be trained to classify and prioritize anomalies or to filter out common expected abnormalities (e.g. humans, tools, etc.). The model may be trained to perform image segmentation to classify specific types of anomalies. As an example, a model could be trained to establish confidence that the location and shape of the anomalies are a leak. For example, the system could extract a coordinate of the leak and establish that it is indeed on the floor. Then the model (or a further model) could appreciate the typical pattern of leaks based on trained models. The typical pattern of leaks may be determined using randomly generated physical possible leaks on a digital twin replica or a bank of images of typical leaks. The same method could also be applied for other type of anomalies, for example fire, animals, trip hazards, etc. Additionally or alternatively, the second artificial intelligence model may be a multi-modal generative AI model, that can receive multiple types of inputs (images, video, audio, text), predict the type of anomaly present, and generate various outputs. The training data captured at 202 and stored at 204 would comprise training data for these different types of inputs collected over time. As one example, the training data for additional types of inputs may comprise audio data of different alarms in a building. The multi-modal generate AI model may be an existing model that is trained specifically using training data associated with physical assets as disclosed herein, or a custom-built and trained model.
Advantageously, with the implementation of an anomaly detection model comprising a chain of artificial intelligence models as described above, the models can help one another learn. By chaining multiple AI models a more accurate final prediction can be created compared to taking only one single approach. A wide array of probabilities from each model can be combined and patterns can be derived that would better assess the situation and provide more accurate prediction for the operators. The training process can also be automated to avoid labelling or identifying by hand the context for each point of view saving numerous man hours of manually executing these tasks. The first artificial intelligence model can produce labelled data that the secondary artificial intelligence model(s) can consume to learn what anomalies look like, and the secondary artificial intelligence model(s) can provide new normal frames that would otherwise look like a false-positive (for example, a recognizable tool left on site), which can be added to the training images and used to update the first artificial intelligence model. Model chaining reduces the probability of triggering false positive by increasing the number of inputs used to calculate the final score of the prediction. This chain of decision can be customized for wider scenarios and specific special context for rooms that would by atypical. This approach drives more value for the operators as they receive fewer false alarms in the end.
When the probability of the physical anomaly being present in the image exceeds a threshold value, an alert is generated to indicate an anomaly event. The event may be analyzed (e.g. by a human operator) (306), who would review the image and confirm the presence of the physical anomaly and optionally the specific type of physical anomaly present in the image. A determination is made as to whether the anomaly detection model classification yielded a false positive (308). If there is no false positive (NO at 308), that is, there is indeed a physical anomaly present in the image, the operator can send instructions/communications for personnel to intervene to address the anomaly associated with the physical asset (310).
If however the operator determines that the classification is a false positive (YES at 308), for example because the anomaly detection model classification was incorrect, or because the operator wishes for this anomaly to be within the realm of normal operating conditions, the image(s) can be added back to the training data set (312) and used to retrain the anomaly detection model as described above with reference to
Image feed(s) (320), comprising one or more images captured by one or more cameras, are provided to a first artificial intelligence model (332) of an anomaly detection solution (330). The first artificial intelligence model (332) is trained to detect anomalous image data and determine a probability of a physical anomaly being present in the image data.
When an anomaly is detected (i.e. the probability of a physical anomaly being present in the image data exceeds a threshold value), one or more secondary artificial intelligence models (334) may be used to determine that the physical anomaly is a specific type of anomaly. In the example representation shown in
For example, a prompt providing image context (324) to the multimodal generative AI (336) may be as follows: You are a system in charge of assessing how critical alerts from a generator room monitored by CCTV. You will receive highlighted area of interest, you need to analyze the highlighted zone and indicate one of the following: (1) You are unable to identify the root cause in the received image, a human operative will need to be analyzing the situation; (2) You are able to identify the root cause of the anomaly and it's not critical. For example a ladder has been left in the middle of the room. There is no immediate danger but the ladder shouldn't be there none the less. So emit a warning; (3) You are able to identify the root cause of the anomaly and it's a problem that should be addressed by the technical team asap. Describe what you perceive and emit a alert.
Further, special context can be included in the prompt providing image context (324), for example if a room has a special situation. An example could be a window that shows an external dynamic scene. Typically, this area would be digitally masked, but the generative AI (336) could still receive the prompt instruction that a zone has been masked and it's normal. Another example of special context could be a specific equipment that may have numerous tools or objects around the area and often trigger anomalies. (The tools or other objects around the area could instead be labelled normal and used to retrain the models, as described above, and thus avoid this prompt).
The prompts with specific questions or tasks (326) may be automatically triggered. The prompts with specific questions or tasks (326) may in particular be prompts to generate an output comprising a description of the specific type of anomaly, and/or a suggested action for responding to the specific type of anomaly. For example, prompts with specific questions or tasks may be:
Is the anomaly in the zone of interest a leak or fire or person?
Classify the analysis of images/videos/sounds as results to prioritize the anomalies, e.g. normal in a generator room (tools, person standing, ladder, etc.), abnormal in a generator room (person lying on the floor, weapon, animal, etc.), typical anomalies linked to leaks (fire, smoke, electrical sparks, severe physical hazards, etc.).
Describe how big the leak is and/or a severity of the leak.
Link the anomaly to a maintenance/repair schedule.
Describe the anomaly in a summary fashion, such as an alert ticket to a support desk/office (with attached content) to help support staff confirm/take actions.
Once the anomaly is resolved, document/close the ticket, and add learnings to a database for knowledge transfer/sharing.
Once the anomaly is resolved, create a feedback loop with false positives, confirmed anomalies, etc.
The anomaly detection solution (330), based on the determinations by the first artificial intelligence model (332) and the one or more secondary artificial intelligence models (334), generates an output (340). The output (340) may for example be an alert comprising an image with the anomaly highlighted, and an indication of the specific type of anomaly, a percentage confidence, etc. The output (340) may also comprise any other description generated by the generative AI model (336), for example depending on the prompts with specific questions/tasks (326).
The prompts may be generated in advance by a development team based on the tasks needed to evaluate the image data, and may be automatically called based on pre-configured prompt. A collection of prompts can be built over time and depending on the physical assets being monitored and specific prompts could be reused.
An anomaly detection system interface may also allow end-users to ask specific questions. For example, a user may be able to ask specific questions such as: What did this room look like before the alerts? Was a door open prior to the alerts within the last hour of the alert? Etc.
An anomaly is detected (350) using a first artificial intelligence model as described above. The anomaly is analyzed (352) using one or more secondary artificial intelligence models. The one or more secondary artificial intelligence models can be used to determine the specific type of anomaly present as the detected anomaly, as well as provide additional information such as how critical the anomaly is and what actions should be taken based on the determinations.
Analyzing the anomaly using the secondary artificial intelligence models can be performed by chaining a number of specialized AI models. As represented in
The decision module (366) receives the output from the various AI models and decides what action should be taken with respect to the anomaly. The decision module (366) may be a typical business rule engine or another generative AI that have some rules applied to help handle the solution. The AI models emit a statistical analysis about how confident they are about a prediction with respect to the detected anomaly. In the case of conflicting information weight and accuracy will be considered by the decision module rule engine. For example, if images show leaks but the audio sounds normal, the image analysis would have precedence over audio analysis. On the other hand, if an abnormal mechanical noise was detected compared to regular operation, then even if the image analysis doesn't show an abnormal feature the decision module (366) may send an audio snippet to the operator for them to assess abnormality of the audio. The output and appropriate action generated by the decision module is dispatched (368).
The anomaly detection model may provide an output comprising the image in which the anomaly is detected and highlighting the anomaly detected in the image, and optionally a confidence of the presence of the anomaly and/or type of anomaly. For example, in images 420 and 422, an output image may place a square surrounding the anomaly/object with an automatic label that may say: Garbage Can 97.43% and Cone 99.98%, respectively.
One or more second artificial intelligence model(s) may also be trained to classify a type of anomaly present in an image (606). To train this second artificial intelligence model, training images are obtained at 602 comprising anomalous image data that have a known anomaly. In some aspects, as described above, the one or more second artificial intelligence models may comprise a multi-modal generative AI model. Training the multi-modal generative AI model may comprise obtaining additional training data at 602 comprising an additional input type; and training the multi-modal generative AI model at 606 to determine the known anomaly present in the anomalous image data based on the training images comprising anomalous image data and the additional training data. Training the multi-modal generative AI model at 606 may further comprise training the multi-modal generative AI model to generate outputs associated with the known anomaly in response to different input prompts.
Image data is received from one or more cameras configured to capture an image of a physical asset (702). The one or more cameras may be configured to capture an image of an area of an area of a building, an area of a site, or a piece of equipment. For example, the building may comprise one of: a hospital, a warehouse, a mechanical room, a server room, and an electrical room. The site may comprise one of: an underground cable tunnel, a manufacturing site, an industrial site, and a mining site. The piece of equipment may be one of: a generator, a battery, a machine, a utility cabinet, and a heating, ventilation, and air conditioning (HVAC) unit. The image data may be received as a single image or a stream of images. The image may be received in association with metadata comprising one or more of a camera identifier and a camera location, so that the anomaly detection model can evaluate anomalies with respect to the camera identifier/camera location. The one or more cameras may comprise one or more of: thermal cameras, near-infrared cameras, and RGB cameras
The image data is analyzed to determine a probability of a physical anomaly being present in the image using an artificial intelligence model that is trained to detect anomalous image data (704). Depending on the type of image data received, the anomaly may be detected as a thermal anomaly in the image data or a visual anomaly in the image data. A determination is made as to whether an anomaly is present in the image (706). The determination is made based on whether the probability of the physical anomaly being present in the image exceeds a threshold value. If there is no anomaly present in the image (NO at 706), i.e. the probability of the physical anomaly being present in the image does not exceed the threshold value, the method returns to 702 and continues to receive further image data. If there is an anomaly present in the image (YES at 706), i.e. the probability of the physical anomaly being present in the image exceeds the threshold value, an alert is output (710). In some implementations, the alert may include the image and a timestamp. The image may include an indicator to highlight the anomaly.
In some embodiments, when it is determined that a physical anomaly is present in the image (YES at 706), the method may comprises analyzing the image data to determine that the physical anomaly is a specific type of anomaly (708). For example, specific types of anomalies may include one or more of: equipment overheat, presence of a human, presence of an animal, misplaced tools or equipment, fire, flood, falling material, and leaks. Determining that the physical anomaly is the specific type of anomaly may comprise determining a probability that the physical anomaly is the specific type of anomaly using one or more secondary artificial intelligence models that are trained to predict the physical anomaly as being one or more types of anomalies. For example, the one or more secondary artificial intelligence models may comprise a model that is trained to perform image segmentation on the image data to classify the anomaly as being the specific type of anomaly. Additionally or alternatively, the one or more secondary artificial intelligence models may comprise a multi-modal generative AI model. In some aspects, the method may further comprise receiving a context of the image data, and the multi-modal generative AI model uses the context of the image data to determine the probability that the physical anomaly is the specific type of anomaly. In some aspects, the method may further comprise receiving audio data and/or vibration data associated with the physical asset, and the multi-modal generative AI model uses the received audio and/or vibration data to determine the probability that the physical anomaly is the specific type of anomaly. The multi-modal generative AI model may be configured to generate an output comprising one or both of: a description of the specific type of anomaly, and a suggested action for responding to the specific type of anomaly. Additionally or alternatively, determining that the physical anomaly is the specific type of anomaly may comprise applying one or more rules to the image data. Further, auxiliary data associated with the physical asset may be received from one or more sensors, and determining that the physical anomaly is the specific type of anomaly may be further based on the auxiliary data. The specific type of anomaly, along with other generated description o the anomaly, may also be included in the alert output at 710.
The method may further comprise updating/retraining the first and/or secondary artificial intelligence models as part of a feedback loop. For example, the method may further comprise receiving user feedback on the specific type of anomaly, and updating the one or more secondary artificial intelligence models based on the user feedback. Additionally or alternatively, the method may further comprise in response to outputting the alert, receiving user feedback that the image data is normal, and updating the artificial intelligence model based on the user feedback.
The server 810 may comprise a messaging software or queue management system 812, such as RabbitMQ™, that is used to communicate with the frame grabber 804 and acquire images corresponding to the camera frames. The images are provided to anomaly detection functionality 814, which runs anomaly detection model 816 to classify the image data and identify anomalies in the images. When the probability of the physical anomaly being present in the image exceeds a threshold value, an alert is output to one or more operator devices and/or to a portal 820. Event data, such as the anomalous image, a predicted type of anomaly, etc., may also be stored in a file system and packaged in a messaging program such as Slack™ or to a Ul connector 818 and output to the operator devices and/or portal 820. A snooze functionality may be provided in the anomaly detection functionality 814, which recognizes that once an alert is sent, it may be unnecessary to continue sending alerts for the same anomaly. The snooze functionality may snooze alerts for X minutes before alerting to an anomaly at the same asset.
The operator devices 820 can also send commands to the messaging program 818 to obtain images, and the messaging program can send commands to the messaging software 812 to request images from frame grabber 804.
It would be appreciated by one of ordinary skill in the art that the system and components shown in the figures may include components not shown in the drawings. For simplicity and clarity of the illustration, elements in the figures are not necessarily to scale, are only schematic and are non-limiting of the elements structures. It will be apparent to persons skilled in the art that a number of variations and modifications can be made without departing from the scope of the invention as described herein.
It is contemplated that any part of any aspect or embodiment discussed in this specification can be implemented or combined with any part of any other aspect or embodiment discussed in this specification.
It should be recognized that features and aspects of the various examples provided above can be combined into further examples that also fall within the scope of the present disclosure.
When used in this specification and claims, the terms “comprises” and “comprising” and variations thereof mean that the specified features, steps or integers are included. The terms are not to be interpreted to exclude the presence of other features, steps or components.
The invention may also broadly consist in the parts, elements, steps, examples and/or features referred to or indicated in the specification individually or collectively in any and all combinations of two or more said parts, elements, steps, examples and/or features. In particular, one or more features in any of the embodiments described herein may be combined with one or more features from any other embodiment(s) described herein.
This application claims priority to U.S. Provisional Patent Application 63/472,067, filed on Jun. 9, 2023, the entire contents of which is incorporated herein by reference for all purposes.
Number | Date | Country | |
---|---|---|---|
63472067 | Jun 2023 | US |