This invention relates generally to predictive modeling and more particularly, to predictive modeling utilizing a committee of models and fusion with locally weighted learning.
Many different approaches have been utilized to optimize asset utilization. For example, asset optimization can be performed in connection with operation of a turbine or boiler for generating electricity supplied to a power grid. It is useful to predict and optimize for parameters such as Heat rate, NOx emissions, and plant load under various operating conditions, in order to identify a most optimum utilization of the turbine or the boiler.
Predictive modeling of an asset to be optimized is one known technique utilized in connection with decision-making for asset optimization. With a typical predictive model, however, local performance can vary over the prediction space. For example, a particular predictive model may provide very accurate results under one set of operating conditions, but may provide less accurate results under another set of operating condition.
Such prediction uncertainty can be caused by a wide variety of factors. For example, data provided to the model under certain conditions may contain noise, which leads to inaccuracy. Further, model parameter misspecification can result due to data-density variations in operating mode representation in the training set data, variations resulting from randomly sampling the training set data, non-deterministic training results, and different initial conditions. Also, model structure misspecification can occur if, for example, there are insufficient neurons in a neural network predictive model or if regression models are not specified with sufficient accuracy.
In one aspect, a method for controlling a process using a committee of predictive models is provided. The process has a plurality of control settings and at least one probe for generating data representative of state of the process. The method includes the steps of providing probe data to each model in the model committee so that each model generates a respective output, aggregating the model outputs, and generating a predictive output based on the aggregating.
In another aspect, a system for generating a predictive output related to a process is provided. The process has a plurality of control settings and at least one probe for generating data representative of state of the process. The system includes a committee of models comprising a plurality of predictive models. Each model is configured to generate a respective output based on data from the probe. The system includes a computer programmed to fuse the outputs from the models to generate at least one predictive output based on the model outputs.
In yet other aspect, a computer implemented method for generating a predictive output related to a process is provided. The process has a plurality of control settings and at least one probe for generating data representative of state of the process. The method includes supplying inputs to a committee of models comprising a plurality of predictive models, executing each model to generate a respective output based on data from the probe, and fusing the outputs from the models to generate at least one predictive output based on the model outputs.
As shown in
In one embodiment, each model 18 is a neural network based data-driven model trained and validated using historical data 20 and constructed to represent input-output relationships, as is well known in the art. For example, for a coal-fired boiler, there may be multiple model committees including multiple models in order to generate outputs representative of the various characteristics of the boiler. Example inputs include the various controllable and observable variables, and the outputs may include emissions characteristics such as NOx and CO, fuel usage characteristics such as heat rate, and operational characteristics such as bearable load.
With respect to
As explained above, the local performance of each model 18 of committee 12 may vary and may not be uniformly consistent over the entire prediction space. For example, in one particular set of operational conditions, one model 18 may have superior performance relative to the other models 18. In another set of operational conditions, however, a different model 18 may have superior performance and the performance of the one model 18 may be inferior. The outputs from models 18 of committee 12 therefore are, in one embodiment, locally weighted using the process described below in order to leverage the localized information so that models 18 are complementary to each other.
With respect to training multiple models, and referring to
For example, historical data 20 typically represents known variable inputs and known outputs. During training, the known output is compared with the model-generated output, and if there is a difference between the model generated output and the known output, the model is then adjusted (e.g., by altering the node weighting and/or connectivity for a neural network model) so that the model generates the known output.
Again, and as illustrated in
Once models 18 are trained and the committee of models 12 is defined, then an algorithm for fusing the model outputs is generated. Many different techniques can be utilized in connection with such fusion, and the present invention is not limited to any one particular fusion technique. Set forth below is one example fusion algorithm.
More particularly, and in the one embodiment with respect to probe 16, a fusion algorithm proceeds by retrieving neighbors/peers of the probe within the prediction inputs space. Local performance of the models is then computed, and multiple predictions are aggregated based on local model performance. Compensation is then performed with respect to the local performance of each model. Compensation may also be performed with respect to the global performance of each model. Such a global performance may be computed by relaxing the neighborhood range for a probe to the entire inputs space. A “fused” output is then generated.
Once the neighbors/peers of Probe Q are retrieved, then the local performance of each model for such neighbors/peers is evaluated, as shown in
Through aggregation and bias compensation, the outputs of the committee of models are fused to generate one output. Use of a committee of models facilitates boosting prediction performance. By decreasing uncertainty in predictions through use of a committee of models and fusion, an aggressive schedule can be deployed in an industrial application as compared to predictions based on one model only. In addition, use of a committee of models and fusion facilitates using a reduced amount of historical data as compared to the historical data used to train systems based on just one model, which facilitates accelerating system deployment.
While the invention has been described in terms of various specific embodiments, those skilled in the art will recognize that the invention can be practiced with modification within the spirit and scope of the claims.