The present disclosure relates generally to process modeling, optimization, and control systems, and more particularly to a method and system for performing multi-objective predictive modeling, monitoring, and update for an asset.
Predictive models are commonly used in a variety of business, industrial, and scientific applications. These models could be based on data-driven construction techniques, based on physics-based construction techniques, or based on a combination of these techniques.
Neural Network modeling, is a well-known instance of data-driven predictive modeling. Such data-driven models are trainable using mathematically well-defined algorithms (e.g., learning algorithms). That is, such models may be developed by training them to accurately map process inputs onto process outputs based upon measured or existing process data. This training requires the presentation of a diverse set of several input-output data vector tuples, to the training algorithm. The trained models may then accurately represent the input-output behavior of the underlying processes.
Predictive models may be interfaced with an optimizer once it is determined that they are capable of faithfully predicting various process outputs, given a set of inputs. This determination may be accomplished by comparing predicted versus actual values during a validation process performed on the models. Various methods of optimization may be interfaced, e.g., evolution algorithms (EAs), which are optimization techniques that simulate natural evolutionary processes, or gradient-descent optimization techniques. The predictive models coupled with an optimizer may be used for realizing a process controller (e.g., by applying the optimizer to manipulate process inputs in a manner that is known to result in desired model and process outputs).
Existing solutions utilize neural networks for nonlinear asset modeling and single-objective optimization techniques that probe these models in order to identify an optimal input-output vector for the process. These optimization techniques use a single-objective gradient-based, or evolutionary optimizer, which optimize a compound function (i.e., by means of an ad hoc linear or nonlinear combination) of objectives.
What is needed is a framework that provides modeling and optimization in a multi-objective space, where there is more than one objective of interest, the objectives may be mutually conflicting, and cannot be combined to compound functions. Such a framework would be able to achieve optimal trade-off solutions in this space of multiple, often conflicting, objectives. The optimal set of trade-off solutions in a space of conflicting objectives is commonly referred to as the Pareto Frontier.
In accordance with exemplary embodiments, a method and system for performing multi-objective predictive modeling, monitoring, and update for an asset is provided.
A method for performing multi-objective predictive modeling, monitoring, and update for an asset, includes determining a status of each of at least two predictive models for an asset as a result of monitoring predicted performance values. The status of each predictive model includes at least one of: acceptable performance values; validating model; and unacceptable performance values. Based upon the status of each predictive model, the method includes performing at least one of: terminating use of the predictive model for the asset; generating an alert for the asset of the status of the predictive model; and updating the predictive model based upon the status of the predictive model.
A system for performing multi-objective predictive modeling, monitoring, and update for an asset, including at least two predictive models relating to an asset, and a monitoring module in communication with the at least two predictive models. The monitoring module monitors predictive performance values for each predictive model and determines a status of each predictive model as a result of the monitoring. The status includes at least one of: acceptable performance values; validating model; and unacceptable performance values. Based upon the status of each predictive model, the system includes performing at least one of: terminating use of the predictive model for the asset; generating an alert for the asset of the status of the predictive model; and updating the predictive model based upon the status of the predictive model.
Other systems, methods, and/or computer program products according to exemplary embodiments will be or become apparent to one with skill in the art upon review of the following drawings and detailed description. It is intended that all such additional systems, methods, and/or computer program products be included within this description, be within the scope of the present invention, and be protected by the accompanying claims.
Referring to the exemplary drawings wherein like elements are numbered alike in the accompanying FIGURES:
In accordance with exemplary embodiments, a process management system is provided. The process management system performs closed-loop, model-based asset optimization and decision-making using a combination of data-driven and first-principles-based nonlinear models, and Pareto Frontier multi-objective optimization techniques based upon evolutionary algorithms and gradient descent. The process management system also performs on-line monitoring and adaptation of the nonlinear asset models. Predictive models refer to generalized models that are tuned to the specific equipment being measured and typically use sampled data in performing model generation and/or calibration. Pareto Frontier optimization techniques provide a framework for tradeoff analysis between, or among, desirable element attributes (e.g., where two opposing attributes for analysis may include turn rate versus range capabilities associated with an aircraft design, and the trade-off for an optimal turn rate (e.g., agility) may be the realization of diminished range capabilities).
A Pareto Frontier may provide a graphical depiction of all the possible optimal outcomes or solutions. Evolutionary algorithms (EAs) may be employed for use in implementing optimization functions. EAs are based on a paradigm of simulated natural evolution and use “genetic” operators that model simplified rules of biological evolution, which are then applied to create a new and desirably more superior population of solutions. Multi-objective EAs involve searches for, and maintenance of, multiple Pareto-optimal solutions during a given search which, in turn, allow the provision of an entire set of Pareto-optimal (Pareto Frontier) solutions via a single execution of the EA algorithm.
Optimization methods typically require starting points from which search is initiated. Unlike an EA that employs an initial population as a starting point, a gradient-based search algorithm employs an initial solution as a starting point (which may be randomly generated from the given search space).
In exemplary embodiments, nonlinear predictive, data-driven models trained and validated on an asset's historical data are constructed to represent the asset's input-output behavior. The asset's historical data refers to measurable input-output elements resulting from operation of the asset. For example, if the asset is a coal-fired boiler, the measurable elements may include emission levels of, e.g., nitrous oxides, carbon monoxide, and sulfur oxides. Historical data may also include operating conditions of the asset, such as fuel consumption and efficiency. Ambient conditions, such as air temperature and fuel quality may be also be measured and included with the historical data.
First-principles-based methods may be used in conjunction with the data-driven models for constructing predictive models representing the asset's input-output relationships. First-principles predictive models are based on a mathematical representation of the underlying natural physical principles governing the asset's input-output relationships. However, it may be necessary to first tune first-principles models based on the asset's historical data, before they are suitable for use. Given a set of ambient conditions for the asset of interest, a multi-objective optimizer probes the nonlinear predictive models of the asset to identify the Pareto-optimal set of input-output vector tuples that satisfy the asset's operational constraints. The multi-objective optimizer may utilize a set of historically similar operating points as seed points to initiate a flexible restricted search of the given search space around these points. A domain-based decision function is superimposed on the Pareto-optimal set of input-output vector tuples to filter and identify an optimal input-output vector tuple for the set of ambient conditions. The asset may be commanded to achieve this optimal state. This optimization process may be repeated as a function of time or as a function of changing operating and ambient conditions in the asset's state.
An online monitoring module (e.g., network-based processor) observes the prediction performance of the nonlinear models as a function of time, and initiates dynamic tuning and update of the various nonlinear predictive models to achieve high fidelity in modeling and closed-loop optimal operational decision-making.
While the invention is described with respect to assets found in a coal-fired plant, it will be understood that the process management system is equally adaptable for use in a variety of other industries and for a wide variety of assets (e.g., gas turbines, oil-fired boilers, refinery boilers, aircraft engines, marine engines, gasoline engines, diesel engines, hybrid engines, etc.). The invention is also adaptable for use in the optimal management of fleets of such assets. The coal-fired boiler embodiment described herein is provided for illustration and is not to be construed as limiting in scope.
Turning now to
The process manager 120 includes a user interface and monitor 107, predictive models 104, a multi-objective optimizer and decision-maker 106, and objective/fitness functions 108. The process manager 120 may be implemented via computer instructions (e.g., one or more software applications) executing on a server, or alternatively, on a computer device, such as user system 101. If executing on a server, the user system 101 may access the features of the process manager 120 over network 105. The user system 101 may be implemented using a general-purpose computer executing one or more computer programs for carrying out the processes described herein. The user system 101 may be a personal computer (e.g., a laptop, a personal digital assistant) or a host attached terminal. If the user system 101 is a personal computer, the processing described herein may be shared by the user system 101 and the host system server (e.g., by providing an applet to the user system 101). User system 101 may be operated by project team members or managers of the provider entity. Various methods of implementing the prediction and optimization functions may be employed as described further herein.
The network 105 may be any type of known network including, but not limited to, a wide area network (WAN), a local area network (LAN), a global network (e.g. Internet), a virtual private network (VPN), and an intranet. The network 105 may be implemented using a wireless network or any kind of physical network implementation known in the art.
The storage device 102 may be implemented using memory contained in the user system 101 or host system or it may be a separate physical device. The storage device 102 is logically addressable as a consolidated data source across a distributed environment that includes a network 105. Information stored in the storage device 102 may be retrieved and manipulated via the host system and may be viewed via the user system 101.
Turning now to
Turning now to
Data relating to outputs, or objectives (also referred to as ‘Y’) represent a third classification. ‘Y’ objectives refer to the target goals of a process such as heat rate, nitrous oxide emissions, etc. ‘Y’ constraints refer to a required constraint on the output, and may be a constraint such as required power output. This classified data is stored in memory (e.g., storage device 102) and is maintained for current and future use. The process management system enables filtering of data, an example of which is depicted in
Steps 302 and 304 may be implemented initially upon set up of the process management system and then updated periodically as needed. A predictive model may now be created using this information as described below.
At step 306, objectives and constraints of interest for the asset are identified. Multiple, sometimes conflicting objectives and constraints may be determined as desired. At step 308, controllable and uncontrollable variables (X, Z) are selected based upon their suitability for achieving a desired objective or required objective (Y). Analyzing the correlations between the (X, Z) variables and the Y objectives or constraints is an important step in determining the suitability of an (X, Z) variable in achieving a Y objective or constraint. An example of this correlation analysis is depicted in
The predictive model may be trained and validated for accuracy at step 312. The predictive model training and validation may proceed by inspection of an actual versus predicted graph 714 of
If the predictive model is valid, meaning that the predicted values coincide, or are in agreement, with the actual values, at step 314, then live data streams may be applied to the predictive models at step 316. If the predictive model is not valid at step 314, then the process returns to step 308 whereby alternative inputs (X, Z) are selected. These predictive models may then be used for optimization via the process management system.
Turning now to
Once these elements have been configured by the user, the process manager 120 identifies a corresponding Pareto Frontier at step 804 by applying a multi-objective optimization algorithm 106 to the predictive models 104. The objective/fitness functions 108 provide feedback to the multi-objective optimizer 106 in the identification of the Pareto Frontier. The Pareto Frontier provides optimal sets of input-output vector tuples that satisfy the operational constraints.
Optionally, a decision function may be applied to the Pareto Frontier at step 806. The decision function may be applied to the optimal sets of input-output vector tuples to reduce the number of input-output vector tuples in what may be referred to as a sub-frontier at step 808. One such decision function may be based on the application of costs or weights to objectives, whereby a subset of Pareto optimal solutions closest to an objectives weighting may be identified. Additional decision functions such as one that is capable of selecting one of the optimal input-output tuples that minimally perturbs the asset from its current state, may be applied. During this process, the process manager 120 provides a feature that enables a user to generate Pareto Frontier graphs that plot these data values. A sample Pareto Frontier graph is shown in
A user at step 101 or process manager in accordance with the user at step 120 may select a deployable input-output vector using the results of the decision functions at step 810. The selected deployable optimal input-output vector is then transmitted to the control system 103 or an operator of the asset at step 812.
Over time, the predictive models are monitored to ensure that they are accurate. In many asset modeling and optimization applications, it is necessary to tune/update the predictive models in order to effectively accommodate moderate changes (e.g., as a function of time) in asset behavior while minimizing the time required for training the predictive models. The process management system enables on-line tuning for predictive models as described in
Turning now to
Upon updating each current model, or alternatively, if the error ratio, ‘E’, does not exceed the pre-determined threshold, ‘Et’, the new data points are added to temporary storage at step 912. The temporary storage, or buffer, has a fixed size, ‘D’.
If adding the new data points to the temporary storage overflows the buffer (Db>D) at step 914, then a new training set is created at step 916. Otherwise, the process returns to step 902. At step 918, the current model is updated via a batch training technique and the temporary storage is emptied at step 920. The batch training trains the predictive model using the data set formed in step 916. Unlike incremental learning, the batch training is more thorough, and may include training, cross-validation, and model configuration optimization. Batch training may be performed at a fixed time interval or when the maximum data size of the buffer that stores the new data is reached. While incremental training of a predictive model allows it to adapt continually to changing asset conditions, batch training of a predictive model helps to periodically recalibrate the models using a more rigorous approach.
As described above, the embodiments of the invention may be embodied in the form of computer implemented processes and apparatuses for practicing those processes. Embodiments of the invention may also be embodied in the form of computer program code containing instructions embodied in tangible media, such as floppy diskettes, CD-ROMs, hard drives, or any other computer readable storage medium, wherein, when the computer program code is loaded into and executed by a computer, the computer becomes an apparatus for practicing the invention.
An embodiment of the present invention can also be embodied in the form of computer program code, for example, whether stored in a storage medium, loaded into and/or executed by a computer, or transmitted over some transmission medium, such as over electrical wiring or cabling, through fiber optics, or via electromagnetic radiation, wherein, when the computer program code is loaded into and executed by a computer, the computer becomes an apparatus for practicing the invention. When implemented on a general-purpose microprocessor, the computer program code segments configure the microprocessor to create specific logic circuits. The technical effect of the executable code is to facilitate prediction and optimization of model-based assets.
While the invention has been described with reference to exemplary embodiments, it will be understood by those skilled in the art that various changes may be made and equivalents may be substituted for elements thereof without departing from the scope of the invention. In addition, many modifications may be made to adapt a particular situation or material to the teachings of the invention without departing from the essential scope thereof. Therefore, it is intended that the invention not be limited to the particular embodiment disclosed as the best or only mode contemplated for carrying out this invention, but that the invention will include all embodiments falling within the scope of the appended claims. Moreover, the use of the terms first, second, etc. do not denote any order or importance, but rather the terms first, second, etc. are used to distinguish one element from another. Furthermore, the use of the terms a, an, etc. do not denote a limitation of quantity, but rather denote the presence of at least one of the referenced item.