The present disclosure relates, in general, to control of power distribution systems to achieve voltage regulation. Disclosed embodiments relate, in particular, to a methodology suitable for controlling voltage and reactive power flow in large scale power distribution systems using reinforcement learning.
Circuits in power distribution systems usually follow a radial topology, which may cause nodes that are far away from the substation (root node) to experience undervoltage. For example, voltage at the end of a feeder may drop below the acceptable range of ±5% of nominal the nominal voltage. Active control of voltage and reactive power flow may be desirable for maintaining healthy operation of power distribution systems.
Volt-var control refers to the control of voltage (Volt) and reactive power (Var) in power distribution systems. Volt-var control usually involves optimally dispatching controllable devices or actuators of a power distribution system to maintain voltage profile at the nodes as well as reduce power losses across the power distribution system.
The volt-var control problem may be subject to network constraints of the distribution system and device constraints of the controllable devices. The network constraint may bring nonconvexity and make it hard to apply traditional optimization algorithms. The device constraints may restrict the decision variables to be discrete, making the modeling and optimization for the decisions highly nontrivial. Hence, the volt-var control problem may be hard to scale to a system with thousands of buses, which is a typical size for many power distribution systems.
Briefly, aspects of the present disclosure provide a scalable technique suitable for volt-var control in large scale power distribution systems using reinforcement learning.
A first aspect of the disclosure provides a method for controlling a power distribution system comprising a number of controllable devices, wherein at least some of the controllable devices are discretely controllable devices operable in discrete switchable states. The method comprises acquiring observations via measurement signals communicated by a plurality of meters in the power distribution system to define a system state. The method further comprises processing the system state using a reinforcement learned volt-var control policy comprising a deep learning model to output a control action that includes respective integer actions for the discretely controllable devices. The integer actions are determined by: using learned parameters of the deep learning model to compute logits for a categorical distribution of predicted actions from the system state, wherein the predicted actions define switchable states of the discretely controllable devices, and processing the logits to reduce the categorical distribution of predicted actions for each discretely controllable device to an integer action for that discretely controllable device. The method comprises communicating the control action to the controllable devices for effecting a change of state of one or more of the controllable devices, to regulate voltage and reactive power flow in the power distribution system.
A further aspect of the disclosure provides a computer-implemented method for training a control policy for volt-var control of a power distribution system using reinforcement learning in a simulation environment, the power distribution system comprising a number of controllable devices, wherein at least some of the controllable devices are discretely controllable devices operable in discrete switchable states. The method comprises acquiring observations by reading state signals from the simulation environment to define a system state of the power distribution system. The method further comprises processing the system state using the control policy to output a control action that includes respective integer actions for the discretely controllable devices, wherein the control policy comprises a deep learning model. The integer actions are determined by using learnable parameters of the deep learning model to compute logits for a categorical distribution of predicted actions from the system state, wherein the predicted actions define switchable states of the discretely controllable devices; for each discretely controllable device, creating a discretized vector representation of the predicted actions based on the respective logits using a straight-through estimator; and reducing the discretized vector representation of the predicted actions to an integer action for the respective discretely controllable device using a linear transformation. The method further comprises updating the learnable parameters of the control policy by computing a policy loss based on the control action, the policy loss being dependent on evaluation of a reward function defined by a volt-var optimization objective.
Other aspects of the disclosure implement features of the above-described method in systems and computer program products for volt-var control in a power distribution system.
Additional technical features and benefits may be realized through the techniques of the present disclosure. Embodiments and aspects of the disclosure are described in detail herein and are considered a part of the claimed subject matter. For a better understanding, refer to the detailed description and to the drawings.
The foregoing and other aspects of the present disclosure are best understood from the following detailed description when read in connection with the accompanying drawings. To easily identify the discussion of any element or act, the most significant digit or digits in a reference number refer to the figure number in which the element or act is first introduced.
Volt-var control involves operating a power distribution system via controlling voltage and reactive power flow to maintain healthy operation of the power distribution system. In particular, volt-var control may be implemented to optimize the operation of controllable devices to minimize power delivery losses as well as maintain the voltage profile at each bus or node of the power distribution system.
In the context of the present description, a “controllable device” or “actuator” refers to a device or component of a power distribution system that is controllable to switch between multiple discrete or continuous states, to regulate voltage and/or reactive power flow in the power distribution system. The present description considers three types of controllable devices, namely, voltage regulators, capacitors and batteries. A voltage regulator may be considered as a switchable transformer operable in multiple states, defined by a tapping number. By changing the tapping number (i.e., by adjusting the output side voltage), a voltage difference between nodes connected to the input side and the output side of the voltage regulator may be adjusted. A capacitor can function as a storage for reactive power. Switching ON a capacitor from an OFF state may increase reactive power flow and bring up voltage profiles. A battery can be used for power management on the load side, for example, for compensating for large loads. Depending on the construction, a battery may have discrete or continuous switchable states.
A power distribution system may comprise any combination of the above-described types of controllable devices, among other types of devices, for implementing volt-var control. The operational change of any single device may potentially result in a change over the entire power distribution system. Thus, at the center of the volt-var control is an optimization for voltage profiles and power losses governed by physical constraints of the power distribution system.
The volt-var control problem can be formulated as an optimum power flow (OPF) problem that involves optimization of an objective function subject to the physical constraints. The objective function is referred to herein as a “volt-var optimization objective.” With the primary goal being voltage regulation, the volt-var optimization objective may typically include a cost for voltage violation at nodes. According to disclosed embodiments, the volt-var optimization objective may be defined by a combination of costs, represented below as:
In equation (1), ƒvolt is a cost for voltage violation at nodes, ƒpower is a cost for power losses, and ƒctrl(x) is a cost for control error pertaining to frequency of change of status of the controllable grid assets (actuators), to prevent actuator wear out by penalizing the actuator status from changing too frequently. In this problem, three different types of actuators are considered, namely voltage regulators (reg), capacitors (cap) and batteries (bat). The described actuators are considered to be discretely controllable, i.e., operable in discrete switchable states. A battery (bat) may be operable in switchable states defined by a discharge powers Pbat which can be discrete (i.e., a finite number of states) or continuous, a capacitor (cap) may be operable in switchable states Statuscap that can include binary ON/OFF states, and a voltage regulator (reg) may be operable in switchable states including a finite number of tapping numbers TapNumreg. In many embodiments, a power system may include a combination of continuously controllable actuators (such as certain types of batteries) and discretely controllable actuators (such as described above).
The volt-var optimization objective, for example as defined above, may be subject to a set of passive constraints that may be governed by the physics of power propagation in the network, as well as active constraints for controlling voltage. The power distribution system may be represented as a radial graph (N, ξ), where N is the set of nodes or buses and ξ is the set of edges defined by lines, transformers and voltage regulators. Denoting node i as j's parent (radial graph is a tree), the constraints may be defined as:
In the above equations, p, q denote active and reactive power consumed at buses (nodes) or power flow over lines (edges), v, l denote nodal voltage magnitude and squared current magnitude, and R, X denote resistance and reactance. All {Pbat, TapNumreg,Statuscap} need to be operating under their operational constraints captured by a set S. The top portion of equation (2c) defines an active constraint while the remaining constraint equations define passive constraints. Note that the volt-var control problem is a time-dependent problem, but for brevity, time t has been omitted in all the variables. The constraints in equations (2a) to (2e) include quadratic equalities, making any optimization upon it non-convex.
State-of-the-art methods have leveraged optimization solvers to solve the OPF problem for volt-var control. However, as seen above, due to the nature of an OPF problem, the resulting optimization problem may be non-convex and thus hard to solve. Together with many integer decision variables in controllable devices, the volt-var control problem can become extremely hard to scale to a system with thousands of buses, which is a typical size for power distribution systems.
The disclosed methodology attempts to solve the volt-var control problem by leveraging a control policy trained using reinforcement learning (RL). It is recognized that a power distribution system has no memory and the system's transition into the next state may be solely dependent on the control action and current state. Hence, according to the disclosed methodology, the volt-var control problem can be cast as a Markov Decision Process and solved using RL, where the volt-var optimization objective (e.g., see equation (1)) may be used to define the reward function in the RL framework.
The RL formulation can naturally satisfy the physical constraints of power flow in the distribution network, thus addressing at least some of the technical challenges of the OPF problem. However, the integer constraints of the controllable devices or actuators remain challenging, that can limit the application of state-of-the-art RL-based controllers only to small-scale power distribution systems.
Recent breakthroughs in RL have shown promising results in applications involving continuous control, such as in gaming, robotics, autonomous driving, among others. However, in a power distribution system, the decision variables for controlling the various actuators are often represented by integers (e.g., tapping number of voltage regulator, discharge power of battery, switching status of capacitor). Because integer variables are not readily differentiable, they can raise an issue with certain types of RL policies that are based on computing gradients over control actions, such as policies based on deep deterministic policy gradient (DDPG) or soft actor-critic (SAC) methods. While alternative methods exist that can obviate the integer action problem, most notably methods using REINFORCE estimators, they tend to have high variance, which can make them unsuitable for large-scale power distribution systems.
According to the disclosed methodology, the above problem can be addressed by leveraging reparameterization methods for discrete random variables (e.g., Gumbel-Softmax estimators), and combining them with straight-through estimators that can represent a discrete random variable in the forward pass (of a neural network) and maintain differentiability in the backward pass. Since integer variables admit a discrete structure, such a reparameterization technique using straight-through estimators can be adopted for the above-described integer action constraint in the volt-var control problem.
Furthermore, the disclosed methodology leverages a key insight that integer variables are fundamentally different from discrete or “categorical” variables. In general, integer variables are comparable while discrete variables are not. For example, the elements in the integer set {1, 2, 3} are comparable because 1<2<3, while the discrete set {dog, cat, frog} cannot assign such a relationship. Typically, a one-hot encoding is employed for representing a discrete sample due to its incomparability. An integer set has comparability, because of which an integer sample can be represented by an integer quantity instead of a one-hot vector. In light of this fact, the disclosed methodology employs a linear mapping from a discretized vector representation (e.g., a one-hot encoding) of categorical predicted actions defined by switchable states of an actuator, created using the straight-through estimator, to an integer action for the actuator.
The mapped integer action derived using the straight-through estimator can be random and differentiable with respect to learnable parameters of the policy. Furthermore, with the integer mapping, the effective dimension of the action space of the actuator can be essentially equal to one and independent of the fineness of discretization. With the differentiability and the effective dimension of the action space, the disclosed methodology may be readily compatible with continuous action-based RL methods such as DDPG and SAC, which have low variances and are hence scalable to large power distribution systems.
Turning now to the disclosed embodiments,
The power distribution system 100 may include measurement devices or meters associated with at least some of the nodes for acquiring observations pertaining to the respective nodes. These nodes are referred to as “measured nodes.” The meters can include, for example, smart metering infrastructure (SMI) devices, among others. The power distribution system 100 may also include one or multiple “unmeasured nodes” from which measurement signals are missing, for example, due to failure or unavailability of meters (e.g., nodes N3, N7 and N8).
Referring to
The nodal features may be assigned to every node of the power distribution system 100. According to disclosed embodiments, the nodal features may include the nodal voltage as well as capacitor, voltage regulator and battery status. Computationally, the nodal features for each node may be represented as a corresponding node vector. Nodes that do not have capacitors, voltage regulators or batteries may be padded with zeros indicative of “absent” status in the corresponding entries in the node vector. The measurement signals 202 may typically comprise time series data. The nodal features of each node may represent a snapshot or instantaneous data samples from the time series data acquired from that node. In many embodiments, the power distribution system 100 may include unmeasured nodes from which measurement signals are missing, for example, due to failure or unavailability of meters. In this scenario, the nodal features of the unmeasured nodes may consist of “zeros” in corresponding fields.
The nodal features may define an observation space which is a product of discrete and continuous variables. The discrete variables may be from the physical constraints of the actuators. For example, a capacitor may be operable only in an ON or an OFF state; a voltage regulator may be operable in a finite number of modes or tapping numbers (typical example is 33 tapping numbers); a discrete battery may be operable in a finite number of discharge powers. The continuous variables may include, for example, the measured nodal voltage/power, state-of-charge or discharge power of a continuous battery, etc.
According to a disclosed embodiment, the system state 206 may be defined by concatenating the nodal features of the nodes of the power distribution system 100 into a vectorized state representation. In another embodiment, the system state may be defined by a graph representation using the nodal features and topological information of the power distribution system 100, for example, as disclosed in the U.S. patent application Ser. No. 17/860,323.
The system state 206 may be sent as input to a volt-var control engine 208. The volt-var control engine 208 may process the system state 206 using a reinforcement learned control policy 210 comprising a deep learning model (such as a neural network) to output a control action for effecting a change of status of one or multiple actuators, to regulate voltage and reactive power flow in the power distribution system 100 based on the defined volt-var optimization objective. According to disclosed embodiments, at least some of the actuators of the power distribution system 100 may include discretely controllable actuators operable in discrete switchable states. The control action may include a combination of integer actions for the discretely controllable actuators.
Based on the control action output by the RL control policy 210, the volt-var control engine 208 may communicate control signals 212 to respective actuators C of the power distribution system 100 (e.g., as shown in
The control policy 210 may be trained via a process of reinforcement learning. The process can include, over a sequence of steps of trial, optimizing learnable parameters (e.g., weights) of the deep learning model of the control policy 210 to maximize a cumulative reward resulting from a sequence of control actions, based on a reward function r defined by a volt-var objective. The reward function r may depend on a current system state st, the action at taken in the current step and the next system state st+1. According to disclosed embodiments, the reward function r may be defined by:
r=−(rv+rc+rp) (3)
In equation (3), rv denotes the penalty or cost for voltage violation at nodes, rc denotes the penalty or cost for control error due to frequently changing the status of actuators and rp denotes the penalty or cost for power losses. The penalty terms in equation (3) map back to the costs in the volt-var optimization objective in equation (1). The terms and rv and rc may be conflicting in practice. Minimizing the voltage violation rv may require frequent operation of voltage regulators, capacitors, and batteries, which would subsequently increase the control error penalty rc, and vice versa. This may result in a multi-objective RL scenario. The reward function may include weights associated with the penalty terms. According to disclosed embodiments, the weights (e.g., wcap, wreg, wbat, wsoc and wpower) may be built into individual penalty terms, as described below.
The penalty rv for voltage violation may be determined, for example, as a sum of worst-case voltage violations among all phases across all nodes of the power distribution system. The upper/lower voltage violation thresholds (
In equation (4), (·)+ is a shorthand for max(·, 0). Thereby, the upper violation (maxp Vn,p−
The penalty rc for control error may be determined, for example, as a sum of the capacitors' and regulators' switching penalties (1st & 2nd rows of equation (5) respectively) and batteries' discharge penalty and state-of-charge (soc) penalty (3rd row of equation (5)). The penalty rc can discourage the control policy from making frequent changes and slow the actuators from wear out. The penalty rc may be thus determined as:
In equation (5),
represents a discharge error with
The penalty rp for power losses may be determined, for example, as a ratio of the overall power loss to the total power, given by:
According to disclosed embodiments, the volt-var control may be cast as an infinite-horizon MDP, with the objective of the RL being to determine an optimal control policy π that maximizes a cumulative reward or “return”, as given by:
In equation (7), E denotes an expectation, t denotes a time step, γ denotes a discount factor, where γ∈(0,1), and r denotes the reward function, for example, as defined by equations (3) to (6).
The control policy π may be parameterized by learnable parameters θ, as πθ. A typical way to optimize equation (7) by tuning over 0 may involve using a policy gradient which may depend on the Q function under πθ, which may be defined as:
Q
π
(s,a)=E[Σt=0∞γtr(st,at)|s0=s,a0=a,at˜π(·|st)] (8)
In equation (8), Qπ
In some embodiments with entropy-regularized RL, such as in a soft actor-critic method, an agent may be assigned a bonus reward at each time step t proportional to the entropy of the control policy πθ at that time step t. In these embodiments, the entropy bonus may be summed up with the volt-var control objective-based reward function r in equation (7). Correspondingly, the Q function Qπ
With Qπ
∇θJ(π0)=Es˜ρ
In equation (9), ρπ denotes an unnormalized discounted state distribution.
Since the expectation of equation (9) may not be easily computable, gradient estimation may become an important problem. Especially, since integer variables of the actions of the discreetly controllable actuators of the power distribution system are not readily differentiable, they can raise an issue in computing a policy gradient for tuning the learnable parameters θ of the control policy πθ. The above-described challenge may be addressed by the disclosed methodology of integer reparameterization.
According to disclosed embodiments, the training process may be implemented on a simulation environment using a simulation model of the power distribution system 100. Thus, the “system state” in the training process may refer to a simulated system state. The trained (and tested) control policy 300 may be subsequently deployed to a runtime system that can include a volt-var control engine for regulating voltage and reactive power flow in the physical power distribution system 100, as described in
Referring to
As described above, according to disclosed embodiments, the system state 302 may comprise a vector including a concatenation of nodal features of nodes of the power distribution system. The nodal features may include a measured electrical quantity (such as nodal voltage and/or power) and a status of actuators associated with the node. In another embodiment, the system state may be defined by a graph representation using the nodal features and topological information of the power distribution system 100, for example, as disclosed in the U.S. patent application Ser. No. 17/860,323.
The system state 302 may be provided as an input to the deep learning model 304. The deep learning model 304 may use the learnable parameters to compute logits 306 for a categorical distribution of predicted actions from the system state 302. In this case, the predicted actions may define switchable states of the discretely controllable actuators. In the illustration shown in
The term “logits” as used in this description, generally refers to a prediction or output of the deep learning model, such as defined by a final or output layer of a neural network. The logits may define a vector having a dimensionality as described above. According to the example embodiment disclosed herein, the logits 306 may define an unnormalized log probability vector.
The logits 306 generated by the deep learning model 304 based on the system state 302 may be processed using the disclosed integer reparameterization methodology to output a control action 318 that includes integer actions for the discretely controllable actuators. According to the disclosed methodology, for each discretely controllable actuator, the respective logits 306 may be used to create a discretized vector representation 314 of the predicted actions (of that actuator) using a straight-through estimator 312. A “straight-through estimator” refers to a computational unit that can use a non-differentiable threshold function in a forward pass (in this case, to compute the integer actions), and replace the non-differentiable function with a differentiable function in a backward pass or backpropagation (in this case, to compute a policy gradient). The discretized vector representation 314 of the predicted actions may be reduced to an integer action for the respective discretely controllable actuator using a linear transformer 316. An example implementation of the disclosed integer reparameterization methodology is described below.
Continuing with reference to
For each actuator, the biased samples 310 created using a Gumbel-Softmax estimator may be represented by a vector DGS having a dimension n equal to the number of discretely switchable states (predicted actions) of that actuator, where:
D
GS[i]=Softmax (logitθ
In equation (10), DGS[i] is the ith sample of the vector DGS corresponding to the ith predicted action for the actuator, logitθ
While the biased samples 310 may be differentiable, they may not be identical to the samples from the corresponding categorical distribution. The straight-through estimator 312 can address this issue by creating a discretized vector representation 314 (such as a one-hot vector encoding) of the categorical samples by correcting the bias in the forward pass, and using the differentiable biased samples 310 in the backward pass. In the context of the disclosed embodiment, the straight-through estimator may include a straight-through Gumbel-Softmax (STGS) estimator. According to disclosed embodiments, the STGS estimator 312 may correct the bias in the forward pass by introducing a constant shift of DGS, as given by:
D
STGS
=D
hot−[DGS]const+DGS (11)
where Dhot is a one-hot vector encoding of the biased samples, given by:
In equation (11), [DGS]const treats DGS as a constant by detaching dependency on θ during backpropagation (backward pass) of the neural network. Thus, in the backward pass, both Dhot and [DGS]const are constants for which the derivatives with respect to the learnable parameters θ is zero. In this manner, DSTGS is a one-hot vector 314 in the forward pass and a differentiable approximation 310 in the backward pass of the neural network. In
The learnable parameters θ may be updated by computing a policy loss based on the control action 318. The policy loss may be determined dependent on evaluation of a reward function (e.g., using equations (3) to (6)). The learnable parameters θ may be updated by computing a gradient of the policy loss. Using the STGS estimator 312, the policy loss may be defined by a differentiable objective function ƒ, where the policy gradient may be computed by estimating the gradient under DSTGS as:
In equation (13), θj denotes the jth element of the learnable parameters θ.
The above-described STGS estimator 312 may provide lower variance, prevent error in the forward pass and maintain sparsity. Although the STGS estimator 312 can provide a good reparameterization for discrete random variables, directly using the discretized vector representation 314 from the STGS estimator 312 to represent integer actions may result in high dimensionality, particularly in case of a large-scale power distribution system.
Integer decision variables usually appear in industrial settings, such as in power distribution systems, that involve many integer-only or binary actions. Despite the discreteness, integer variables share one thing in common with continuous variables, namely comparability, or more mathematically, the ability to form a totally ordered set. This is the key that makes integer variables distinct from discrete counterparts, which allows integer variables to be treated like continuous instead of discrete/categorical variables. The proposed integer reparameterization is based on the insight that, since integer actions do not necessarily require a one-hot encoding, once being generated, they may be treated like a continuous action in a subsequent computation. This can greatly reduce the computational overhead and can be a key in adapting continuous action-based RL algorithms, such as SAC and DDPG, for integer actions.
According to disclosed embodiments, the control policy 300 may further include a computational unit 316 that applies a linear transformation for each discretely controllable actuator, to reduce the discretized vector representation 314 to a respective integer action for that actuator. Recall that the STGS estimator 312 (see equation (11)) generates a differentiable one-hot random vector DSTGS. Then
the index/argument of 1 in DSTGS, is an integer random variable following the distribution formed by Softmax(logitθ
it is possible to have a differentiable random integer parameterized by θ. According to disclosed embodiments, such a differentiable mapping may be constructed using a linear map, as given by:
Indθ=[0,1, . . . ,n−1],DSTGS (14)
In equation (14), ·, · denotes an inner product and DSTGS is assumed to be n-dimensional. The linear transformation according to equation (14) thus results in a random integer Indθ, parameterized by θ, whose value equals
Still referring to
For a basic understanding of the SAC method, the reader is directed to the publication: Tuomas Haarnoja, Aurick Zhou, Pieter Abbeel, and Sergey Levine. Soft actor-critic: Off-policy maximum entropy deep reinforcement learning with a stochastic actor. In International conference on machine learning, pages 1861-1870. PMLR, 2018. The state-of-the-art also includes a number of variants of the above-mentioned approach.
Referring now to
The learnable parameters ϕk of the Q function approximators 402 may be updated using sampled state-action pairs (s, a) by regressing to a single shared target determined based on evaluation of the reward function (e.g., using equations (3) to (6)), for example via minimization of a mean-squared Bellman error (MSBE) loss function, as known in the art.
The learnable parameters θ of the control policy 300 may be updated in the following manner. The control policy 300 may receive, as input, a sampled system state s, and use its learnable parameters θ to output a control action aθ, which may include integer actions for each discretely controllable actuator, as described above. Each Q function approximator 402 may receive, as input the sampled system state s and the control action aθ and use its learnable parameters ϕk to respectively determine a scalar Qk (s, aθ). The learnable parameters θ of the control policy 300 may be updated based on minimization of a policy loss, which may be defined as a function of
averaged over the batch, for example, as a negative of the
The policy loss may be minimized by backpropagating the differentiable approximation DSTGS for aθ to estimate a gradient of the policy loss at each step, for example, using equation (13).
Since each integer action aθ is of effective dimension 1 (in a Euclidean space), the action dimension may be precisely equal to the number of integer action variables. Thereby, the input/output dimension of the control policy 300 and the Q function approximators 402 can be the same as their continuous action counterparts. The only difference may lie in the policy design, where the output is restricted to have an integer structure as described above, while the continuous action version does not have such a restriction.
Subsequent to the training process, the control policy 300 may be deployed to a runtime system as a reinforcement learned control policy for implementing volt-var control on a physical power distribution system, such as the RL control policy 210 shown in
Consistent with disclosed embodiments, the logits from the deep learning model may be processed by creating, for each discretely controllable actuator, a discretized vector representation of the predicted actions based on the respective logits. As described above, the discretized vector representation may be created by perturbing the logits with a random noise to create biased samples (e.g., using a Gumbel-Softmax estimator as defined in equation (10)), and computing a one-hot vector encoding of the biased samples (e.g., using equation (12)). An integer action for the respective actuator may then be determined from the discretized vector representation of the predicted actions using a linear transformation (e.g., using equation (14))
For validation of the disclosed methodology, a volt-var control policy was trained using a SAC algorithm based on the disclosed embodiments on four different power distribution system environments and compared to a volt-var control policy trained using a proximal policy optimization (PPO) algorithm without reparameterization. The experimental results are shown in
As shown in
The computing system 600 may execute instructions stored on the machine-readable medium 620 through the processor(s) 610. Executing the instructions (e.g., the state estimation instructions 622 and the volt-var control instructions 624) may cause the computing system 600 to perform any of the technical features described herein, including according to any of the features of the state estimation engine 204 and the volt-var control engine 208 described above.
The systems, methods, devices, and logic described above, including the state estimation engine 204 and the volt-var control engine 208, may be implemented in many different ways in many different combinations of hardware, logic, circuitry, and executable instructions stored on a machine-readable medium. For example, these engines may include circuitry in a controller, a microprocessor, or an application specific integrated circuit (ASIC), or may be implemented with discrete logic or components, or a combination of other types of analog or digital circuitry, combined on a single integrated circuit or distributed among multiple integrated circuits. A product, such as a computer program product, may include a storage medium and machine-readable instructions stored on the medium, which when executed in an endpoint, computer system, or other device, cause the device to perform operations according to any of the description above, including according to any features of the state estimation engine 204 and the volt-var control engine 208. Computer readable program instructions described herein can be downloaded to respective computing/processing devices from a computer readable storage medium or to an external computer or external storage device via a network, for example, the Internet, a local area network, a wide area network and/or a wireless network.
The processing capability of the systems, devices, and engines described herein, including the state estimation engine 204 and the volt-var control engine 208, may be distributed among multiple system components, such as among multiple processors and memories, optionally including multiple distributed processing systems or cloud/network elements. Parameters, databases, and other data structures may be separately stored and managed, may be incorporated into a single memory or database, may be logically and physically organized in many different ways, and may be implemented in many ways, including data structures such as linked lists, hash tables, or implicit storage mechanisms. Programs may be parts (e.g., subroutines) of a single program, separate programs, distributed across several memories and processors, or implemented in many different ways, such as in a library (e.g., a shared library).
Although this disclosure has been described with reference to particular embodiments, it is to be understood that the embodiments and variations shown and described herein are for illustration purposes only. Modifications to the current design may be implemented by those skilled in the art, without departing from the scope of the patent claims.
Number | Date | Country | |
---|---|---|---|
63242135 | Sep 2021 | US |