This patent application claims priority to Indian Patent Application No. 202241039424, filed on Jul. 8, 2022, which is incorporated herein in its entirety by reference.
The present disclosure generally relates to control systems in industries. More particularly, the present disclosure relates to a method and a control system for dynamic provisioning of visual contents to a user using machine learning.
Control systems are used in industrial systems to monitor processes for any abnormality in the processes. The control systems gather data from sensors located in the industrial systems and processes the data to provide visual content (e.g., a graphic of an industrial process) to users. The visual content provides visualization of the processes in real-time. The visualization aids in monitoring the processes in the industrial systems. Such control systems for example, include, Supervisory Control and Data Acquisition (SCADA), Distributed Control System (DCS), and the like. In a large industrial automation system, the control system continuously monitors a large set of complex processes for identifying any abnormality in the processes. Due to heavy load in the processing, interface of the control system, for example, Human Machine Interface (HMI), takes significant amount of time to provide the visual content. The amount of time taken by the interface is termed as call up time in the present description. It is crucial that the visual contents requested by the user are provided to the user without any delay. Also, when there are multiple visual contents, the user has to select and analyze various visual contents for monitoring the processes. The visual contents in which a certain user is interested in or is the most critical one needs to be provided within a short call up time. Hence, the call up time needs to be decreased for the control system to provide enhanced user experience.
The information disclosed in this background of the disclosure section is only for enhancement of understanding of the general background of the invention and should not be taken as an acknowledgement or any form of suggestion that this information forms the prior art already known to a person skilled in the art.
In an embodiment, the present disclosure discloses a method of providing visual contents to a user for monitoring processes in an industrial system. The method comprises receiving a plurality of variables associated with a plurality of processes in an industrial system, from one or more sources. Further, the method comprises determining a presence of one or more critical variables from the plurality of variables based on one or more parameters, using a machine learning model. Furthermore, the method comprises identifying one or more first visual contents from a plurality of visual contents associated with the plurality of variables by associating the one or more critical variables with the plurality of visual contents, upon determination. Each of the plurality of visual contents represents one or more processes from the plurality of processes and corresponding variables. Moreover, the method comprises identifying one or more second visual contents based on availability of behavior data of a user, using the machine learning model. Thereafter, the method comprises providing the one or more first visual contents and the one or more second visual contents to the user, for monitoring at least one process of the plurality of processes in the industrial system.
In an embodiment, the present disclosure discloses a control system for providing visual contents to a user for monitoring processes in an industrial system. The control system comprises one or more processors and a memory. The one or more processors are configured to receive a plurality of variables associated with a plurality of processes in an industrial system, from one or more sources. Further, the one or more processors are configured to determine a presence of one or more critical variables from the plurality of variables based on one or more parameters, using a machine learning model. Furthermore, the one or more processors are configured to identify one or more first visual contents from a plurality of visual contents associated with the plurality of variables by associating the one or more critical variables with the plurality of visual contents, upon determination. Each of the plurality of visual contents represents one or more processes from the plurality of processes and corresponding variables. Moreover, the one or more processors are configured to identify one or more second visual contents based on availability of behavior data of a user, using the machine learning model. Thereafter, the one or more processors are configured to provide the one or more first visual contents and the one or more second visual contents to the user, for monitoring at least one process of the plurality of processes in the industrial system.
The foregoing summary is illustrative only and is not intended to be in any way limiting. In addition to the illustrative aspects, embodiments, and features described above, further aspects, embodiments, and features will become apparent by reference to the drawings and the following detailed description.
It should be appreciated by those skilled in the art that any block diagram herein represents conceptual views of illustrative systems embodying the principles of the present subject matter. Similarly, it will be appreciated that any flow charts, flow diagrams, state transition diagrams, pseudo code, and the like represent various processes which may be substantially represented in computer readable medium and executed by a computer or processor, whether or not such computer or processor is explicitly shown.
In the present document, the word “exemplary” is used herein to mean “serving as an example, instance, or illustration.” Any embodiment or implementation of the present subject matter described herein as “exemplary” is not necessarily to be construed as preferred or advantageous over other embodiments.
While the disclosure is susceptible to various modifications and alternative forms, specific embodiment thereof has been shown by way of example in the drawings and will be described in detail below. It should be understood, however that it is not intended to limit the disclosure to the particular forms disclosed, but on the contrary, the disclosure is to cover all modifications, equivalents, and alternatives falling within the scope of the disclosure.
The terms “comprises,” “comprising,” or any other variations thereof, are intended to cover a non-exclusive inclusion, such that a setup, device or method that comprises a list of components or steps does not include only those components or steps but may include other components or steps not expressly listed or inherent to such setup or device or method. In other words, one or more elements in a system or apparatus proceeded by “comprises . . . a” does not, without more constraints, preclude the existence of other elements or additional elements in the system or apparatus.
Control systems are used in industrial systems to monitor processes for any abnormality in the processes and provide visual content to users. The users use the visual content for monitoring the processes in the industrial systems. Such control systems for example, include, Supervisory Control and Data Acquisition (SCADA), Distributed Control System (DCS), and the like. In a large industrial automation system, due to heavy load in the processing, interface of the control system takes significant amount of time to provide the visual content. The amount of time taken by the interface is referred as call up time. The call up time is required to be decreased for the control system to provide enhanced user experience.
The present disclosure provides a method and a control system for providing the visual contents to the user for monitoring the processes in the industrial system. The present disclosure determines critical variables from multiple variables of the processes in the industrial system based on certain parameters of the variables, using a machine learning model. The critical variables that need immediate attention of the users are determined. These critical variables are associated with visual contents in the control system to identify first visual contents from the visual contents. Hence, the first visual contents are identified based on the critical variables of the processes in the industrial system. Further, second visual contents is identified based on behavior data of a user, using the machine learning model. The first visual contents and the second visual contents are provided to the user, for monitoring the processes in the industrial system. The present disclosure identifies a set of visual contents (the first visual contents and the second visual contents) from the multiple contents in the control system based on the critical variables and the behavior data of the user. Hence, the visual contents are dynamically provisioned to the user based on the critical variables and the behavior data of the user. The number of visual contents provided to the user is a reduced set of visual contents, which reduces the call up time of the control system. This enhances the user experience. Further, the reduced set of visual contents (i.e., the first visual contents and the second visual contents are stored in a cache of the control system to enable quick and easy access of the visual contents, based on a request from the user. This further reduces the call up time of the control system. Also, since the reduced set of visual contents is stored in the cache, storage in the cache is optimized. Also, since the first visual contents are identified based on the critical variables that need immediate attention of the users, the monitoring of the processes is enhanced. Also, since the second visual contents are identified based on the behavior data of the user, visual contents relevant to the user can be identified.
In one implementation, the modules 205 may include, for example, an input module 212, a determination module 213, a first identification module 214, a second identification module 215, an output module 216, and auxiliary modules 217. It will be appreciated that such aforementioned modules 205 may be represented as a single module or a combination of different modules. In one implementation, the computation data 204 may include, for example, input data 206, determination data 207, first identification data 208, second identification data, output data 210, and auxiliary data 211.
In an embodiment, the input module 212 may be configured to receive the plurality of variables associated with the plurality of processes 101 in the industrial system. The plurality of variables may comprise at least one of, electrical variables, mechanical variables, and thermal variables. A person skilled in the art will appreciate that the plurality of variables may include any variables other than the above-mentioned variables.
Referring back to
Referring again to the example shown in
Referring back to
Referring back to
Referring back to
The auxiliary data 211 may store data, including temporary data and temporary files, generated by the one or more modules 205 for performing the various functions of the control system 102. The one or more modules 205 may also include the auxiliary modules 217 to perform various miscellaneous functionalities of the control system 102. The auxiliary data 211 may be stored in the memory 202. It will be appreciated that the one or more modules 205 may be represented as a single module or a combination of different modules.
The order in which the method 400 is described is not intended to be construed as a limitation, and any number of the described method blocks can be combined in any order to implement the method. Additionally, individual blocks may be deleted from the methods without departing from the scope of the subject matter described herein. Furthermore, the method can be implemented in any suitable hardware, software, firmware, or combination thereof.
At step 401, the control system 102 receives the plurality of variables associated with the plurality of processes 101 in the industrial system. The plurality of variables may comprise at least one of, electrical variables, mechanical variables, and thermal variables. The control system 102 may receive the plurality of variables from the one or more sources comprising one or more sensors in the industrial system.
At step 402, the control system 102 determines a presence of the one or more critical variables from the plurality of variables based on one or more parameters, using the machine learning model. The one or more parameters may comprise at least one of, alarm severity level associated with the plurality of variables and threshold values associated with a rate of change of the plurality of variables. The alarm severity level associated with the plurality of variables may indicate a level of severity of an alarm generated for the plurality of variables. The threshold values associated with the rate of change of the plurality of variables may indicate higher rate of change or lower rate of change. The threshold values may be pre-determined or learnt by the machine learning model. The machine learning model may determine the one or more critical variables based on a pre-defined alarm severity level or threshold level.
At step 403, the control system 102 identifies the one or more first visual contents from the plurality of visual contents associated with the plurality of variables. Each of the plurality of visual contents represents the one or more processes from the plurality of processes 101 and corresponding variables. The plurality of visual contents may be generated by the control system 102. The control system 102 may identify the one or more first visual contents by associating the one or more critical variables with the plurality of visual contents.
At step 404, the control system 102 identifies the one or more second visual contents based on availability of the behavior data of the user 104, using the machine learning model. The behavior data may comprise at least one of, historic visual content selected by the user 104, a selection pattern associated with the historic visual content, identification details of the user 104 and a role of the user 104 in the industrial system. The behavior data may be updated based on a learning by the machine learning model when the user 104 selects one or more visual contents other than the one or more first visual contents and the one or more second visual contents.
At step 405, the control system 102 provides the one or more first visual contents and the second visual contents from the cache upon receiving a request from the user 104. The number of the one or more first visual contents and the one or more second visual contents provided to the user 104 is less than a pre-defined threshold value based on a size limit of the interface. The control system 102 may provide the one or more second visual contents to the user 104 based on the behavior data, upon determining an absence of the one or more critical variables. The control system 102 may provide the one or more first visual contents to the user 104 based on the association of the one or more critical variables with the plurality of visual contents, upon determining unavailability of the behavior data.
COMPUTER SYSTEM:
The processor 502 may be disposed in communication with one or more input/output (I/O) devices (not shown) via I/O interface 501. The I/O interface 501 may employ communication protocols/methods such as, without limitation, audio, analog, digital, monoaural, RCA, stereo, IEEE (Institute of Electrical and Electronics Engineers)-1394, serial bus, universal serial bus (USB), infrared, PS/2, BNC, coaxial, component, composite, digital visual interface (DVI), high-definition multimedia interface (HDMI), Radio Frequency (RF) antennas, S-Video, VGA, IEEE 802.n/b/g/n/x, Bluetooth, cellular (e.g., code-division multiple access (CDMA), high-speed packet access (HSPA+), global system for mobile communications (GSM), long-term evolution (LTE), WiMax, or the like), etc.
Using the I/O interface 501, the computer system 500 may communicate with one or more I/O devices. For example, the input device 510 may be an antenna, keyboard, mouse, joystick, (infrared) remote control, camera, card reader, fax machine, dongle, biometric reader, microphone, touch screen, touchpad, trackball, stylus, scanner, storage device, transceiver, video device/source, etc. The output device 511 may be a printer, fax machine, video display (e.g., cathode ray tube (CRT), liquid crystal display (LCD), light-emitting diode (LED), plasma, Plasma display panel (PDP), Organic light-emitting diode display (OLED) or the like), audio speaker, etc.
The processor 502 may be disposed in communication with the communication network 509 via a network interface 503. The network interface 503 may communicate with the communication network 509. The network interface 503 may employ connection protocols including, without limitation, direct connect, Ethernet (e.g., twisted pair 10/100/1000 Base T), transmission control protocol/internet protocol (TCP/IP), token ring, IEEE 802.11a/b/g/n/x, etc. The communication network 509 may include, without limitation, a direct interconnection, local area network (LAN), wide area network (WAN), wireless network (e.g., using Wireless Application Protocol), the Internet, etc. The network interface 503 may employ connection protocols include, but not limited to, direct connect, Ethernet (e.g., twisted pair 10/100/1000 Base T), transmission control protocol/internet protocol (TCP/IP), token ring, IEEE 802.11a/b/g/n/x, etc.
The communication network 509 includes, but is not limited to, a direct interconnection, an e-commerce network, a peer to peer (P2P) network, local area network (LAN), wide area network (WAN), wireless network (e.g., using Wireless Application Protocol), the Internet, Wi-Fi, and such. The first network and the second network may either be a dedicated network or a shared network, which represents an association of the different types of networks that use a variety of protocols, for example, Hypertext Transfer Protocol (HTTP), Transmission Control Protocol/Internet Protocol (TCP/IP), Wireless Application Protocol (WAP), etc., to communicate with each other. Further, the first network and the second network may include a variety of network devices, including routers, bridges, servers, computing devices, storage devices, etc.
In some embodiments, the processor 502 may be disposed in communication with a memory 505 (e.g., RAM, ROM, etc. not shown in
The memory 505 may store a collection of program or database components, including, without limitation, user interface 506, an operating system 507, web browser 508 etc. In some embodiments, computer system 500 may store user/application data, such as, the data, variables, records, etc., as described in this disclosure. Such databases may be implemented as fault-tolerant, relational, scalable, secure databases such as Oracle® or Sybase®.
The operating system 507 may facilitate resource management and operation of the computer system 500. Examples of operating systems include, without limitation, APPLE MACINTOSH® OS X, UNIX®, UNIX-like system distributions (E.G., BERKELEY SOFTWARE DISTRIBUTION™ (BSD), FREEBSD™, NETBSD™, OPENBSD™, etc.), LINUX DISTRIBUTIONS™(E.G., RED HAT™, UBUNTU™, KUBUNTU™, etc.), IBM™ OS/2, MICROSOFT™ WINDOWS™ (XP™, VISTA™/7/8, 10 etc.), APPLE® IOS™, GOOGLE® ANDROID™, BLACKBERRY® OS, or the like.
In some embodiments, the computer system 500 may implement the web browser 508 stored program component. The web browser 508 may be a hypertext viewing application, for example MICROSOFT® INTERNET EXPLORER™, MICROSOFT® EDGE®, GOOGLE® CHROME™0, MOZILLA® FIREFOX™, APPLE® SAFARI™, etc. Secure web browsing may be provided using Secure Hypertext Transport Protocol (HTTPS), Secure Sockets Layer (SSL), Transport Layer Security (TLS), etc. Web browsers 508 may utilize facilities such as AJAX™, DHTML™, ADOBE® FLASH™, JAVASCRIPT™, TYPESCRIPT™ JAVA™, Application Programming Interfaces (APIs), etc. In some embodiments, the computer system 500 may implement a mail server (not shown in Figure) stored program component. The mail server may be an Internet mail server such as Microsoft Exchange, or the like. The mail server may utilize facilities such as ASP™, ACTIVEX™, ANSI™ C++/C#, MICROSOFT®,.NET™, CGI SCRIPTS™, JAVA™, JAVASCRIPT™, PERL™, PHP™, PYTHON™, WEBOBJECTS™, web assemblies, etc. The mail server may utilize communication protocols such as Internet Message Access Protocol (IMAP), Messaging Application Programming Interface (MAPI), MICROSOFT® exchange, Post Office Protocol (POP), Simple Mail Transfer Protocol (SMTP), web socket, or the like. The mail server may utilize communication technology such as, but not limited to, push technology. In some embodiments, the computer system 500 may implement a mail client stored program component. The mail client (not shown in Figure) may be a mail viewing application, such as APPLE® MAIL™, MICROSOFT® ENTOURAGE™, MICROSOFT® OUTLOOK™, MOZILLA® THUNDERBIRD™, etc.
Furthermore, one or more computer-readable storage media may be utilized in implementing embodiments consistent with the present disclosure. A computer-readable storage medium refers to any type of physical memory on which information or data readable by a processor may be stored. Thus, a computer-readable storage medium may store instructions for execution by one or more processors, including instructions for causing the processor(s) to perform steps or stages consistent with the embodiments described herein. The term “computer-readable medium” should be understood to include tangible items and exclude carrier waves and transient signals, i.e., be non-transitory. Examples include Random Access Memory (RAM), Read-Only Memory (ROM), volatile memory, non-volatile memory, hard drives, Compact Disc Read-Only Memory (CD ROMs), Digital Video Disc (DVDs), flash drives, disks, and any other known physical storage media.
The present disclosure provides a method and a control system for providing the visual contents to the user for monitoring the processes in the industrial system. The present disclosure determines critical variables from multiple variables of the processes in the industrial system based on certain parameters of the variables, using a machine learning model. The critical variables that need immediate attention of the users are determined. These critical variables are associated with visual contents in the control system to identify first visual contents from the visual contents. Hence, the first visual contents are identified based on the critical variables of the processes in the industrial system. Further, second visual contents are identified based on behavior data of a user, using the machine learning model. The first visual contents and the second visual contents are provided to the user, for monitoring the processes in the industrial system. The present disclosure identifies a set of visual contents (the first visual contents and the second visual contents) from the multiple contents in the control system based on the critical variables and the behavior data of the user. Hence, the visual contents are dynamically provisioned to the user based on the critical variables and the behavior data of the user. The number of visual contents provided to the user is a reduced set of visual contents, which reduces the call up time of the control system. This enhances the user experience. Further, the first visual contents and the second visual contents are stored in a cache of the control system and retrieved upon receiving a request for the visual contents from the user, thus providing quick and easy access to the visual contents. This further reduces the call up time of the control system. Also, since the first visual contents are identified based on the critical variables that need immediate attention of the users, the monitoring of the processes is enhanced. Also, since the second visual contents are identified based on the behavior data of the user, the user experience if further enhanced. Also, since the second visual contents are identified based on the behavior data of the user, visual contents relevant to the user can be identified.
The terms “an embodiment”, “embodiment”, “embodiments”, “the embodiment”, “the embodiments”, “one or more embodiments”, “some embodiments”, and “one embodiment” mean “one or more (but not all) embodiments of the invention(s)” unless expressly specified otherwise.
The terms “including”, “comprising”, “having” and variations thereof mean “including but not limited to”, unless expressly specified otherwise.
The enumerated listing of items does not imply that any or all of the items are mutually exclusive, unless expressly specified otherwise. The terms “a”, “an” and “the” mean “one or more”, unless expressly specified otherwise.
A description of an embodiment with several components in communication with each other does not imply that all such components are required. On the contrary a variety of optional components are described to illustrate the wide variety of possible embodiments of the invention.
When a single device or article is described herein, it will be readily apparent that more than one device/article (whether or not they cooperate) may be used in place of a single device/article. Similarly, where more than one device or article is described herein (whether or not they cooperate), it will be readily apparent that a single device/article may be used in place of the more than one device or article, or a different number of devices/articles may be used instead of the shown number of devices or programs. The functionality and/or the features of a device may be alternatively embodied by one or more other devices which are not explicitly described as having such functionality/features. Thus, other embodiments of the invention need not include the device itself.
The illustrated operations of
Finally, the language used in the specification has been principally selected for readability and instructional purposes, and it may not have been selected to delineate or circumscribe the inventive subject matter. It is therefore intended that the scope of the invention be limited not by this detailed description, but rather by any claims that issue on an application based here on. Accordingly, the disclosure of the embodiments of the invention is intended to be illustrative, but not limiting, of the scope of the invention, which is set forth in the following claims.
All references, including publications, patent applications, and patents, cited herein are hereby incorporated by reference to the same extent as if each reference were individually and specifically indicated to be incorporated by reference and were set forth in its entirety herein.
The use of the terms “a” and “an” and “the” and “at least one” and similar referents in the context of describing the invention (especially in the context of the following claims) are to be construed to cover both the singular and the plural, unless otherwise indicated herein or clearly contradicted by context. The use of the term “at least one” followed by a list of one or more items (for example, “at least one of A and B”) is to be construed to mean one item selected from the listed items (A or B) or any combination of two or more of the listed items (A and B), unless otherwise indicated herein or clearly contradicted by context. The terms “comprising,” “having,” “including,” and “containing” are to be construed as open-ended terms (i.e., meaning “including, but not limited to,”) unless otherwise noted. Recitation of ranges of values herein are merely intended to serve as a shorthand method of referring individually to each separate value falling within the range, unless otherwise indicated herein, and each separate value is incorporated into the specification as if it were individually recited herein. All methods described herein can be performed in any suitable order unless otherwise indicated herein or otherwise clearly contradicted by context. The use of any and all examples, or exemplary language (e.g., “such as”) provided herein, is intended merely to better illuminate the invention and does not pose a limitation on the scope of the invention unless otherwise claimed. No language in the specification should be construed as indicating any non-claimed element as essential to the practice of the invention.
Preferred embodiments of this invention are described herein, including the best mode known to the inventors for carrying out the invention. Variations of those preferred embodiments may become apparent to those of ordinary skill in the art upon reading the foregoing description. The inventors expect skilled artisans to employ such variations as appropriate, and the inventors intend for the invention to be practiced otherwise than as specifically described herein. Accordingly, this invention includes all modifications and equivalents of the subject matter recited in the claims appended hereto as permitted by applicable law. Moreover, any combination of the above-described elements in all possible variations thereof is encompassed by the invention unless otherwise indicated herein or otherwise clearly contradicted by context.
Number | Date | Country | Kind |
---|---|---|---|
202241039424 | Jul 2022 | IN | national |