This disclosure relates generally to the field of energy storage and distribution and, more specifically, to methods for predicting energy consumption demand for peak load shaving.
Meeting peak electric demand is a fundamental challenge that utilities and grid operators, faced with rising generation, transmission, and regulatory costs, must address in efficient and economical ways. Furthermore, utilities are under increasing legislative pressure that mandates increasing integration of high-variability renewables into their generation portfolios. All the while, regulated utilities must still fulfill the terms of their monopolies granted in exchange for guarantees to meet demand, and face stiff penalties for failure.
In order to mitigate the risks associated with meeting peak demand under heavier renewables integration requirements, utilities may invest in extra generation capacity that remains idle for all but a few extreme events in the year. Such an approach incurs very high capital and operational expenses. Another long-standing approach is forecasting demand several hours to several days ahead, and hedging against unexpected spikes in demand or generation failures by purchasing call option contracts securing the right but not the obligation to buy electricity from the wholesale market at set prices following a certain waiting period. This strategy carries its own risks: prices may fluctuate significantly as the forecast horizon decreases. A utility may lose value on its contracts if prices and/or demand drop, or the utility may need to make costly additional electricity purchases if demand spikes.
With the advent of the smart grid and associated advanced monitoring systems, utilities have increased ability to influence demand and mitigate costs by imposing variable pricing and demand charges corresponding to different periods in the day that relate to different expected loads. For instance, a utility may charge a higher per-unit price at 12:00 PM on a Wednesday in July than at 3:00 AM on a Sunday in October. Furthermore, a utility may levy a demand charge that corresponds to the peak load incurred by a customer over a given period, e.g., one month. Such charges in principle incentivize customers to reduce absolute peak usage, thereby reducing the cost to the utility of excessive reserve provisioning.
To cope with demand charges and energy efficiency goals, customers are increasingly turning to sophisticated building energy management systems (EMS). EMS are cyber-physical systems comprised of software and hardware that enable real-time monitoring, control, and optimization of electricity generation, transmission, storage, and usage. Together with stationary energy storage systems, an EMS enables a building manager to reduce or defer grid electricity consumption during periods of high demand charges. As used herein, the term “peak load shaving” refers to an energy management approach wherein grid electricity consumption is reduced during periods of peak demand. Such reductions are especially beneficial in the case of demand charges or inelastic demand that can be met by stored, dispatchable energy reserves. Consequently, improvements to EMSs that improve the effectiveness of stationary energy storage systems in providing peak shaving would be beneficial.
In one embodiment, a method for peak load shaving in an energy management system (EMS) has been developed. The method includes identifying with a controller an available energy capacity of an energy storage device in the EMS, estimating with the controller a level and duration of peak power consumption for a load connected to the EMS over a predetermined time period based on a feed-forward neural network trained with a history of peak power consumption measurements by the EMS, identifying with the controller a power consumption threshold for the load connected to the EMS with reference to the level and duration of peak power consumption estimated by the controller and the available energy capacity of the energy storage device, measuring with the controller a power consumption level of the load during the predetermined time period, and activating with the controller the energy storage device to provide energy to the load from the energy storage device in response to the measured power consumption level of the load exceeding the threshold.
In another embodiment, an EMS that performs peak load shaving has been developed. The EMS includes an energy storage device connected to a load and to an external electrical power source and a controller operatively connected to the energy storage device. The controller is configured to identify an available energy capacity of an energy storage device in the EMS, estimate a level and duration of peak power consumption for a load connected to the EMS over a predetermined time period based on a feed-forward neural network trained with a history of peak power consumption measurements by the EMS, identify a power consumption threshold for the load connected to the EMS with reference to the level and duration of peak power consumption estimated by the controller and the available energy capacity of the energy storage device, measure a power consumption level of the load during the predetermined time period, and activate the energy storage device to provide energy to the load from the energy storage device in response to the measured power consumption level of the load exceeding the threshold.
A method to assist in peak load shaving with an energy storage device includes generation of adaptive estimates of the load threshold for which the energy consumed by the load exceeding the threshold is equal to the effective capacity of the storage system. An energy management system (EMS) generates threshold predictions beginning during a period when demand is low, and are updated throughout the day using the observed load samples and previous threshold estimates as additional inputs. The EMS uses the estimates to control the stationary energy storage device to discharge whenever total load exceeds the current threshold estimate, and to charge to full capacity whenever total load falls below the current estimate.
There are three main benefits for generating predictions of the load threshold when compared to other methods. First, the predictions mitigate the uncertainty in predicting daily peak load or hourly load, which is often highly variable, by instead computing what amounts to an average over several hours. The threshold is a proxy for excess energy consumed, and the threshold can be computed by the product of the average instantaneous excess load multiplied by the number of hours during which the load exceeds the threshold. Second, predicting thresholds over a comparatively short period, such as one hour or a window of a few hours, reduces the computational complexity in predicting the load, which typically involves a far larger training data set and an increased number of models (corresponding to each horizon from 1 to 24 hours ahead) that decrease in accuracy as horizon increases. Instead, a single threshold suffices to convey the information that the controller 112 requires to characterize the load for a day. Third, the controller generates individual hourly models to forecast the threshold on the basis of information up to that hour, the controller 112 adaptively adjusts the estimate of the threshold and can more accurately capture surprise events that occur during the morning ramp up to peak load.
The systems and methods described herein enable peak shaving using threshold prediction. The prediction method makes a novel application of state of the art forecasting technology to quantify the threshold such that energy consumed by load in excess of the threshold equals a desired amount. One embodiment uses artificial neural networks for developing threshold predictions for the load profile of a school. The threshold prediction method is not limited to peak shaving since threshold prediction methods can also be used to determine other energy quantities related to daily load. The embodiments described herein are not model-dependent, and can be implemented using arbitrary nonlinear regression and training methods.
For the purposes of promoting an understanding of the principles of the embodiments disclosed herein, reference is now be made to the drawings and descriptions in the following written specification. No limitation to the scope of the subject matter is intended by the references. The present disclosure also includes any alterations and modifications to the illustrated embodiments and includes further applications of the principles of the disclosed embodiments as would normally occur to one skilled in the art to which this disclosure pertains.
The goal of threshold prediction is to quantify the threshold such that the total energy consumed by load exceeding that threshold is equal to a specified amount (e.g., 100 kWh).
The threshold prediction method makes use of pattern recognition and machine learning algorithms that find relationships within observed data. Given a load profile consisting of predictor-output pairs, with predictors, such as time of day/week/year, operating schedule, temperature, and previous loads, and associated outputs, such as measured loads, the controller 112 first compute thresholds for each day. The controller 112 uses the thresholds to create a new profile containing pairs consisting of predictors and daily thresholds. Note that while the initial load profile may have been sampled hourly or sub-hourly, a threshold profile consists of daily pairs.
The controller 112 uses statistical learning algorithms to build a discriminative model that estimates a functional relationship between predictors (inputs) and thresholds (outputs) using the training set of predictor-threshold pairs. Discriminative modeling frameworks include nonlinear regression models such as artificial neural networks, support vector machines, and kernel-smoothing regression, and enable the estimation of an unseen mean conditional on an observation. The controller 112 uses the trained model to predict unseen thresholds in the test set using predictor vectors.
The controller 112 generates a different model for each hour of the normal day shift (e.g., 8 AM to 3 PM). In addition to inputs that all models share such as the previous day peak load and threshold, each model uses the most recent measured load and estimated threshold as additional inputs. Thus, the EMS 104 operates with as many threshold profiles as there are models to be trained, and each daily threshold has a distinct input vector corresponding to each particular model.
By using multiple models, the prediction method adapts each individual day's estimated threshold each hour as new data becomes available. The hourly estimated threshold is used as an input to a controller that switches between charge and discharge modes depending on whether or not the load exceeds the current threshold estimate.
In one embodiment, the EMS 104 uses a neural network model to obtain threshold predictions for the load profile of a commercial customer. The neural network is an example of one embodiment of a prediction model. Alternative configurations of the EMS 104 use different predictors and modeling frameworks. Furthermore, alternative embodiments also apply the threshold prediction algorithm to other thresholds besides the daily peak excess energy, such as for instance the threshold over which the peak 100 kWh of morning ramp up energy usage falls. Neural networks are one modeling approach in the load forecasting literature to model the highly nonlinear relationship between predictors such as temperature and seasonality and historical load. Neural networks are particularly suited to learning curves for situations that are not well suited to development of parametric models or physics-based models, and have been successfully applied to numerous regression and classification problems.
The embodiment of
y
k(x)=hk(Σi=1mvigi(Σj=1nwijxj+∝i)+∝0)
In the neural network model, xj refers to the inputs, yk refers to the outputs, ∝i and ∝0 refer to bias terms, hk refer to output activation functions, gi refers to hidden layer activation functions, vi refers to weights for the hidden layer activation functions gi, and wijxj.
In one configuration, the neural network is trained using a maximum likelihood framework. In an embodiment that utilizes a Gaussian distribution of errors conditional on observed input data, the maximization of likelihood is equivalent to minimizing a least squares cost function equal to the sum of the squared difference between the outputs y of the neural network and the corresponding measured thresholds, or targets, t. Because the cost function includes non-convex parameters, the optimization problem may not have a unique global optimum, and nonlinear optimization algorithms can be used to train the network.
A major potential pitfall is overfitting, in which a nonlinear regression fits the training data very well, but performs poorly when predicting new data. Neural networks are susceptible to overfitting when the number of model parameters approaches or exceeds the number of data points. The controller 112 restricts the number of parameters in the model to be no more than 10% of the number of data points. The use of independent validation sets also help to obtain models with good generalization performance, and typically encourage selection of more parsimonious models.
Before beginning training, the controller 112 reserves a random subset of the training data for validation. During training, the controller 112 monitors the cost function on both the remaining training data as well as the set held out for validation, and stops training once the validation set error no longer decreases (even if the remaining training set error continues to decrease). Optimal network size often depends on the data, and the controller 112 selects the number of hidden units by training several network sizes several times, using a different validation set each time, and choosing the best performer on the basis of mean absolute error between training and target points on independent test sets not used during the training period. The controller 112 trains the neural network in a similar manner to k-fold cross validation, in which the training data is partitioned into k subsets and the network is trained k times, each time holding out one of the subsets for validation. Network performance is evaluated on the basis of overall performance on the validation subsets for each network size. The controller 112 uses Bayesian regularized gradient descent to determine parameters v, w, and α that minimize (perhaps locally) the least squares cost function. Bayesian regularization penalizes overfitting and maintains a parsimonious model by assigning parameter weights close to zero to inputs deemed irrelevant. Several other approaches to determine relevant inputs such as F tests and sensitivity analysis are viable alternatives.
The controller 112 uses a trapezoidal numerical integration algorithm to compute thresholds that are illustrated in
In one embodiment, the controller 112 performs ten rounds of training for each threshold model, and picks the best performer according to minimum training set error.
It will be appreciated that variants of the above-disclosed and other features and functions, or alternatives thereof, may be desirably combined into many other different systems, applications or methods. Various presently unforeseen or unanticipated alternatives, modifications, variations or improvements may be subsequently made by those skilled in the art that are also intended to be encompassed by the following claims.
This application claims priority to U.S. Provisional Application No. 62/095,455, which is entitled “Method for Adaptive Demand Charge Reduction,” and was filed on Dec. 22, 2014, the entire contents of which are hereby incorporated by reference herein. This application claims further priority to U.S. Provisional Application No. 62/095,810, which is entitled “Method for Adaptive Demand Charge Reduction,” and was filed on Dec. 23, 2014, the entire contents of which are hereby incorporated by reference herein.
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/US2015/067491 | 12/22/2015 | WO | 00 |
Number | Date | Country | |
---|---|---|---|
62095810 | Dec 2014 | US | |
62095455 | Dec 2014 | US |