Embodiments of the invention pertain to the field of modeling and optimization.
Many theoretical and practical problems can be formulated as a general constrained nonlinear optimization problem of the form:
where the objective function ƒ(x), the equality constraints hi(x), i∈ε={1, . . . , nε} and the inequality constraints gi(x), j∈={1, . . . , n} can be nonlinear and are all twice differentiable, that is, they all belong to C2:Rn→R.
Solving the optimization problem (P.1) and finding the global optimal solution have many industrial applications in different areas. The optimal power flow (OPF) problem in electric power systems is one example. In solving the OPF problem, the objective function ƒ(x) can be the system total production cost or the system total power losses; and the equality constraints h(x) are the power balance equations that must be complied because of underlying physical laws for operating the power network; and the inequality constraints g(x) include the physical and operational limits imposed on the system equipment operating conditions or states to ensure a secure operation of the power network; and the optimization variables x are quantities associated with devices of the power network that can be adjusted, such as the power outputs by generators, voltage settings at system nodes, amount of shunt capacitors deployed, and tap positions of transformers. A tank design for a multi-product plant in chemical engineering is another example. In solving the tank design problem, the objective function ƒ(x) can be the sum of the production cost per ton per product produced; and the optimization variables x are quantities of products; and the equality constraints h(x) consist of balance of materials used in production; and the inequality constraints g(x) include lower and upper bounds on volume flows and inventories of products.
Because of the nonlinearity of the objective and constraint functions, a real world problem usually contains many local optimal solutions, and their objective values can differ significantly to each other; therefore, obtaining a global optimal solution to (P.1) is of primary importance in real applications. On the other hand, the constraint set of (P.1) defines the following feasible set:
S={x∈R
n
:h
i(x)=0,i∈ε,gj(x)≦0,j∈},
which can be any closed subset of Rn, and its structure can be very complex. The feasible set S is usually non-convex and disconnected; i.e., it is composed of several disjoint and path-connected feasible regions. As a result, the existence of multiple local optimal solutions, the number of which is typically unknown, to the optimization problem (P.1) can be due to not only the nonlinearity of the objective function, but also the complexity of the feasible regions. Since locating each connected feasible region of the set S is in itself a difficult task, it is a challenge to find the global optimal solution to the nonlinear optimization problem (P.1).
There has been a wealth of research efforts focused on developing effective and robust methods to solve the optimization problem (P.1). Existing optimization methods for solving (P.1) can be roughly categorized into two types. The first type is called local methods, such as trust-region methods, sequential quadratic programming (SQP), and interior point methods (IPM). These methods usually solve first-order necessary conditions numerically to find local optimal solutions to (P.1). They are generally deterministic and fast to compute a local optimal solution, but can be entrapped in the local optimal solution. The other type is called global methods, such as genetic algorithms (GA), particle swarm optimization (PSO) and simulated annealing (SA). These methods generally use stochastic heuristics to escape from a local optimal solution and directly search for an approximation to the global optimal solution to (P.1). Global methods are good at locating promising areas, but they are generally computationally demanding to find a good approximation to the global optimal solution.
Prior work related to the constrained global optimization is described in J. Lee and H. D. Chiang, A dynamical trajectory-based methodology for systematically computing multiple optimal solutions of general nonlinear programming problems, IEEE Trans. Autom. Control, vol. 49, pp. 888-899, 2004, the complete disclosure of which is hereby incorporated herein by reference. This publication presents a two-phase optimization procedure to tackle the optimization problem (P.1), where a particular nonlinear dynamical system is constructed in the first phase to locate multiple feasible components of (P.1) and another particular nonlinear dynamical system is constructed in the second phase for finding multiple local optimal solutions to (P.1) within each feasible component.
The idea of computing exit points of a class of nonlinear systems is described in U.S. Pat. No. 5,483,462, On-line Method for Determining Power System Transient Stability, issued to Hsiao-Dong Chiang on Jan. 9, 1996, the complete disclosure of which is hereby incorporated herein by reference.
The idea of computing all local optimal solutions to unconstrained continuous and discrete optimization problems is described in U.S. Pat. No. 7,050,953, Dynamical Methods for Solving Large-Scale Discrete and Continuous Optimization Problems, issued to Hsiao-Dong Chiang and Hua Li on May 23, 2006, the complete disclosure of which is hereby incorporated herein by reference.
The idea of computing all local optimal solutions to a constrained nonlinear optimization problem with a two-phase methodology is described in U.S. Pat. No. 7,277,832, Dynamical Method for Obtaining Global Optimal Solution of General Nonlinear Programming Problems, issued to Hsiao-Dong Chiang on Oct. 2, 2007, the complete disclosure of which is hereby incorporated herein by reference.
A method is provided for obtaining a global optimal solution to general constrained nonlinear optimization problems. The method includes the steps of: (i) finding, in a deterministic and tier-by-tier manner, all stable equilibrium points (SEPs) of a nonlinear dynamical system associated with the optimality conditions of the optimization problem; (ii) finding from the SEPs all local optimal solutions to the optimization problem; and (iii) finding from the local optimal solutions a global optimal solution to the optimization problem.
A practical numerical method is also provided for effectively and reliably computing all local optimal solutions, and finding from which a global optimal solution to large-scale constrained nonlinear optimization problems. The practical numerical method includes the steps of: (i) constructing a nonlinear dynamical system associated with the Karush-Kuhn-Tucker (KKT) conditions of the optimization problem; (ii) computing a comprehensive set of SEPs of the dynamical system through incorporating a local optimizer, namely, the interior point method (IPM), with an exit-point based multi-tier dynamical search; (iii) identifying a comprehensive set of local optimal solutions to the optimization problem among the set of stable equilibrium points; and (iv) identifying the global optimal solution among the set of local optimal solutions.
According to one embodiment of the invention, a method performed by an optimization system is provided for optimizing an objective function of an industrial system subject to a set of constraint functions. The method includes: modeling the objective function and the constraint functions as a constrained nonlinear optimization problem; and computing a global optimal solution to the constrained nonlinear optimization problem by performing the steps of: (a) constructing a dynamical system associated with optimality conditions of the constrained nonlinear optimization problem; (b) starting from an initial point, performing a deterministic, tier-by-tier dynamical search on the dynamical system and obtaining a complete set of SEPs of the dynamical system; (c) identifying a complete set of local optimal solutions to the constrained nonlinear optimization problem from the complete set of SEPs; and (d) identifying the global optimal solution to the constrained nonlinear optimization problem from the complete set of local optimal solutions.
In another embodiment, a non-transitory computer readable storage medium includes instructions that, when executed by a computer system, cause the computer system to perform the aforementioned method for optimizing an objective function of an industrial system subject to a set of constraint functions.
In yet another embodiment, a system is provided for optimizing an objective function of an industrial system subject to a set of constraint functions. The system includes: a memory to store a global optimizer module; and one or more processors coupled to the memory. The one or more processors are adapted to execute operations of the global optimizer module to: model the objective function and the constraint functions as a constrained nonlinear optimization problem; and compute a global optimal solution to the constrained nonlinear optimization problem. The one or more processor further are adapted to (a) construct a dynamical system associated with optimality conditions of the constrained nonlinear optimization problem; (b) start from an initial point, performing a deterministic, tier-by-tier dynamical search on the dynamical system and obtaining a complete set of SEPs of the dynamical system; (c) identify a complete set of local optimal solutions to the constrained nonlinear optimization problem from the complete set of SEPs; and (d) identify the global optimal solution to the constrained nonlinear optimization problem from the complete set of local optimal solutions.
Embodiments are illustrated by way of example and not limitation in the Figures of the accompanying drawings:
In the following description, numerous specific details are set forth. However, it is understood that embodiments of the invention may be practiced without these specific details. In other instances, well-known circuits, structures and techniques have not been shown in detail in order not to obscure the understanding of this description. It will be appreciated, however, by one skilled in the art, that the invention may be practiced without such specific details. Those of ordinary skill in the art, with the included descriptions, will be able to implement appropriate functionality without undue experimentation.
One reliable way to find the global optimal solution to (P.1) is to first find all local optimal solutions, and then find the global optimal solution from the local optimal solutions. To this end, it is desirable to develop a deterministic method that not only escapes from a local optimal solution, but also computes multiple local optimal solutions to the optimization problem (P.1).
Embodiments of the invention provide a systematic and deterministic methodology for finding all local optimal solutions to general constrained optimization problems of the form (P.1). For these optimization problems, one distinguishing feature of the methodology is that it systematically finds all of the local optimal solutions to a constrained nonlinear optimization problem whether or not its feasible region is disconnected. In particular, the methodology finds all of the local optimal solutions in a single phase, by constructing a nonlinear dynamical system using optimality conditions associated with the optimization problem. In this construction, all stationary points of the optimality conditions, which are critical points of the optimization problem and contain all its local optimal solutions, have a one-to-one mapping to the set of stable equilibrium points (SEPs) of the dynamical system. Therefore, by systematically finding all SEPs of the dynamical system, the methodology is able to find all local optimal solutions to the optimization problem (P.1). Then the global optimal solution can be found from these local optimal solutions.
Embodiments of the invention improve the existing methods in at least the following aspects:
1) A class of nonlinear dynamical systems is provided for solving general constrained nonlinear optimization problems. The trajectories of these dynamical systems can be exploited to develop numerical methods for finding a complete set of local optimal solutions and the global optimal solution. Moreover, this class of nonlinear dynamical systems and the trajectories of the systems can also be used to solve unconstrained nonlinear optimization problems.
2) A single-phase, dynamical trajectory-based method is provided. This method can incorporate any existing local optimizer to find a complete set of local optimal solutions and the global optimal solution to general constrained nonlinear optimization problems. Moreover, this method is a unified method that is applicable to both unconstrained and constrained nonlinear optimization problems.
TRUST-TECH Guided Methods for Constrained Nonlinear Global Optimization. As mentioned before, a reliable way to find the global optimal solution of the optimization problem (P.1) is to first find all of the local optimal solutions, and then find the global optimal solution from the local optimal solutions. This can be done through a procedure which conceptually includes the following two tasks:
Task 1: Start from an arbitrary point and compute a local optimal solution to the optimization problem (P.1).
Task 2: Move away from the local optimal solution and approach another local optimal solution of the optimization problem (P.1).
TRUST-TECH based methods accomplish these two tasks using the trajectories of a particular class of nonlinear dynamical systems. More specifically, TRUST-TECH based methods accomplish the tasks by the following steps:
i) Construct a dynamical system such that there is a one-to-one correspondence between the set of local optimal solutions to the optimization problem (P.1) and the set of SEPs of the dynamical system; in other words, for each local optimal solution to the problem (P.1), there is a distinct SEP of the dynamical system that correspond to it.
ii) Find all SEPs of the constructed dynamical system, and then find a complete set of local optimal solutions to the problem (P.1) among the complete set of SEPs.
iii) Find the global optimal solution from the complete set of local optimal solutions.
(a) Receive an initial point x0 (block 201).
(b) Construct a dynamical system associated with the optimality conditions of the optimization problem (P.1) (block 202).
(c) Apply a local optimizer using the initial point x0 to compute an initial SEP xs0 of the dynamical system
(d) Set i=0, Vs={xs0}, Vnewi={xs0}.
(e) Set Vnewi+1=Ø.
For each SEP in Vnewi, the method 200 further includes the following steps (f) through (l):
(f) Compute a set of search directions {Sij, j=1, 2, . . . , mi}, and set j=1.
(g) Search for a dynamic decomposition point xdj along the search direction Sij (block 203). If a dynamic decomposition point xdj is found, proceed to step (h); otherwise proceed to step (k). A dynamical decomposition point (DDP) is the first type-1 unstable equilibrium point whose stable manifold is hit by the search path at the exit point. A type-1 unstable equilibrium point is an equilibrium point at which the system Jacobian matrix has one and only one eigenvalue with a positive real part (all other eigenvalues have negative real parts).
(h) Set x0j=xdj+ε(xdj−xsi), where ε is a small number. Compute a new initial point xrj, which is in a next-tier stability region, by integrating the trajectory of the dynamical system for a few steps starting from x0j (block 204).
(i) Apply a local optimizer using the initial point xrj to compute an SEP xsj (block 204).
(j) Check whether xsj∈Vs. If xsj ∈Vs, then proceed to step (k); otherwise, set Vs=Vs ∪{xsj} and Vnewi+1=Vnewi+1∪{xsj} and proceed to step (k).
(k) Set j=j+1 and check whether j<=mi (i.e., whether all directions are searched) (block 205). If j<=mi (block 206), proceed to step (g); otherwise, proceed to step (l).
(l) Check if Vnewi+1 is non-empty (block 207). If Vnewi+1 is non-empty, then set i=i+1 (block 208) and proceed to step (e); otherwise, proceed to step (m).
(m) Output the complete set of SEPs of the dynamical system (block 209).
Interior Point Methods (IPMs). IPMs have been an important research area in optimization since the development of the Simplex method. In particular, these methods provide an attractive alternative to active set strategies in handling problems with large numbers of inequality constraints. Over the past decades, there has also been a better understanding of the convergence properties of IPMs. Efficient algorithms have been developed with desirable global and local convergence properties.
An IPM includes three basic elements: (i) a barrier method to handle inequality constraints; (ii) a Lagrangian method to handle equality constraints; and (iii) an improved Newton method to solve the set of nonlinear equations which originates from the Karush-Kuhn-Tucker (KKT) optimality conditions. In order to solve the optimization problem (P.1), an IPM firstly applies the Fiacco-McCormick barrier method and adds slack variables to transform the optimization problem (P.1) into the following equality-constrained optimization problem:
Then the following augmented Lagrangian function can be defined:
L
g=ƒ(x)−yTh(x)−wT(g(x)+u)−μln uj,
where, yi and wj are Lagrangian multipliers for equality and inequality constraints, respectively, and uj are slack variables. The first-order necessary conditions, namely the KKT conditions (also referred to as the first-order KKT optimality conditions), associated with the Lagrangian f unction Lg are given as follows:
L
x=∇ƒ(x)−∇xTh(x)y−∇xTg(x)w=0
L
y
=h(x)=0
L
w
=g(x)+u=0
L
u
=UWE+μE=0 (E.1)
where U=diag(u1, u2, . . . , un), W=diag(w1, w2, . . . , wn) and E=[1, 1, . . . , 1]T. The following two decomposed linear equations are obtained when applying the Newton method to solve the above nonlinear equations (E.1):
Observing equations (E.2) and (E.3), it is impossible to solve (E.2) and (E.3) directly because the right-hand side includes unknown high-order deviations ΔuΔw. In order to solve this problem, the predictor-corrector interior point method was proposed, in which a predictor step and a corrector step are needed at each iteration. Because the predictor step and the corrector step share the same coefficient matrix with two different right-hand sides, only one LU factorization is needed. The predictor-corrector interior point method is composed of the following steps:
1) Initialization: Set iteration number k=0, give the initial values of state variables x0, slack variables u0, Lagrange multipliers y0, w0.
2) Let μ=0, ΔuΔw=0, solve the linear equations (E.2) to obtain the affine directions Δxof and Δyof, then obtain Δuof and Δwof by back substitution of (E.3).
3) Compute step sizes αofp and αofd, and modify complementary gaps, GAP and GAPof, update the barrier parameter:
4) Set μ=μof, ΔuΔw=ΔuofΔwof, and resolve the linear equations (E.1) using the same LU factorization matrix obtained in step 2 to obtain centering-corrector directions Δx and Δy. Then obtain Δu and Δw by back substitution of (E.2), and update all the variables:
5) Compute complementary gap GAP=−uTw. If GAP and maximal absolute power flow mismatch are less than the given precision, or the maximal iteration is reached, then stop; otherwise, go to step 2.
Despite their nice numerical efficiency, IPMs are a class of local optimization methods. In other words, from a starting point, IPMs can only compute a single local optimal solution to the optimization problem (P.1). Same as other types of local methods, there is no mechanism available in IPMs to escape from a local optimal solution and to approach another local optimal solution.
The TRUST-TECH Guided IPM. According to embodiments of the invention, a TRUST-TECH guided IPM (TT-IPM) method is provided that integrates the concept of TRUST-TECH into IPMs. TT-IPM enables the computation of all local optimal solutions to the nonlinear optimization problem (P.1).
A principal task in developing TT-IPM is to design a nonlinear dynamical system whose trajectories can be explored to perform Task 1 and Task 2 mentioned above. The central idea in designing such a nonlinear dynamical system is that all the local optimal solutions to the optimization problem (P.1) corresponds with all the SEPs of the nonlinear dynamical system. In particular, every local optimal solution to the optimization (P.1) corresponds with an SEP of the nonlinear dynamical system.
As has been shown in previous section, IPM solves the optimization problem (P.1) by finding a stationary point of the penalized Lagrange function
L(x,y,z,s,μ)=ƒ(x)+yTh(x)+zT(g(x)+s)−μΣi=1 ln si. (E.4)
Specifically, a stationary point of (E.4) can be obtained by solving the KKT optimality conditions (i.e., first order necessary conditions), which can be represented as the following system of nonlinear equations:
L
x=∇ƒ(x)−∇xTh(x)y−∇xTg(x)z=0
L
y
=h(x)=0
L
z
=g(x)+s=0.
L
s
=Zs−μe=0 (E.5)
As μ→0, the solution of the system of nonlinear equations (E.5) approaches a critical point
The major difficulty in solving (E.5) is mainly related to the complementarity conditions Lz and Ls. Therefore, instead of solving the system of nonlinear equations (E.5) directly until μ=0, a nonlinear complementary approach can be introduced to handle the complementarity conditions Lz and Ls. In this construction, a critical point of the optimization problem (P.1) can be obtained by solving the following system of nonlinear equations
L
x=∇ƒ(x)−∇xTh(x)y−∇xTg(x)z=0
L
y
=h(x)=0
L
z
=g(x)+s=0
Ψs=ψ(z,s)=0 (E.6)
where the nonlinear complementarity function ψ:R2→R holds the property of
ψ(z,s)=0zs=0,z≧0,s≧0.
Examples of nonlinear complementarity functions include
In order to realize TT-IPM, in one embodiment of the present invention, a KKT dynamical system is constructed. The KKT dynamical system has a vector field associated with critical points of the optimization problem (P.1), and can be defined by the following formulation:
{dot over (w)}=F(w). (E.7)
In another embodiment of the present invention, a KKT dynamical system having a vector field associated with critical points of the optimization problem (P.1) is defined as follows:
{dot over (w)}=−∇
T
F(w)·F(w), (E.8)
where w is the vector of variables of the system of nonlinear equations, and includes all the primal and dual variables for the optimization problem (P.1). F(w) is the function of the system of nonlinear equations (E.5) or the modified system of nonlinear equations (E.6).
In these constructions, the state variables of the system are w=(x, y, z, s)T. In one embodiment, the nonlinear dynamical system equation (E.7) can be constructed using the gradient vector of the Lagrange function (E.4), that is, F(w)=(Lx, Ly, Lz, Ls)T. In another embodiment of the present invention, the nonlinear dynamical system equation (E.7) or (E.8) can be constructed using the nonlinear complementarity formulation (E.6), that is, F(w)=(Lx, Ly, Lz, Ψs)T.
Based on these constructions, the TT-IPM method can be numerically implemented via the following stages:
Stage 1: Approach an SEP of the KKT dynamical system, which is a critical point of the optimization problem (P.1).
Stage 2: Move from the SEP to a DDP (in order to escape from the critical point).
Stage 3: Approach another SEP of the KKT dynamical system, which is another critical point of the optimization problem (P.1), by moving along the unstable manifold of the DDP.
Stage 1 can be implemented by following the trajectory of the KKT dynamical system starting from any initial point. Stage 3 can be implemented by following the trajectory starting from an initial point, which is close to the DDP but outside the stability region, until it approaches another SEP of the KKT dynamical system, which is another critical point of the optimization problem (P.1).
Characterization of the KKT dynamical system. A trajectory of the KKT dynamical system (E.8) is used to systematically locate the set of critical points to the optimization problem (P.1) in a deterministic manner. The global dynamical behaviors of the KKT dynamical system (E.8) play a central role in finding the set of critical points of the optimization problem (P.1).
Assumption 1: For any bounded, closed interval [a,b] and for any relatively closed subset K of (L−1([a, b])\EL) in S, we have inf{∥∇TF(w)·F(w)∥x∈K}>0, where, EL={w∈Rm:∇T F(w)·F(w)=0} and m=n+2nε+2n.
It is noted that if each path-connected feasible component of M is compact, then Assumption 1 always holds. The constraint components of many practical optimization problems are compact, hence, Assumption 1 usually holds.
The global behavior of nonlinear dynamical systems can be very complicated. Their trajectories can behave unbounded or/and bounded but complicated such as almost periodic trajectory, or even chaotic motions. The global behavior of the KKT dynamical system (E.8) is very simple: every trajectory converges to an equilibrium point. There is no complicated behaviors such as closed orbit (i.e., limit cycles) and chaotic motion that can exist in the KKT dynamical system (E.8). Note that a nonlinear dynamical system is said to be completely stable if every trajectory of the system converges to one of its equilibrium points.
Theorem 1.1 (Completely Stable): If the optimization problem (P.1) satisfies Assumption 1, then the corresponding KKT dynamical system (E.8) is completely stable.
Theorem 1.2 (Equilibrium Points and Critical Points): If the optimization problem (P.1) satisfies Assumption 1 and all the equilibrium points of the KKT dynamical system (E.8) is hyperbolic and finite in number, then each critical points of the optimization problem (P.1) is an SEP of the KKT dynamical system (E.8). Conversely, if
Theorem 1.2 provides a theoretical basis for the TT-IPM method. It asserts that each critical point of the optimization problem (P.1) can be located via the stable equilibrium point of the KKT dynamical system (E.8).
Theorem 1.3 (Characterization of Quasi-Stability Boundary):
Let Ap(xs) be the quasi-stability region of xs of the KKT dynamical system (E.8) and let xi, i=1, 2, . . . , be all the DDPs of xs. If the system satisfies the following conditions: i) the equilibrium points on the quasi-stability boundary ∂Ap(xs) are hyperbolic and finite in number, and ii) the stable and unstable manifolds of the equilibrium points on ∂Ap(xs) satisfy the transversality condition; then the quasi-stability boundary ∂Ap(xs) is the union of the closure of the stable manifolds of xi. In other words, ∂Ap(xs)=∪x
Next, a dynamical relationship is derived between an SEP and a DDP lying on the quasi-stability boundary of the SEP in the following theorem.
Theorem 1.4 (DDPs and SEPs)
Let Ap(xs) be the quasi-stability region of xs of the KKT dynamical system (E.8) that satisfies the following conditions: i) the equilibrium points on quasi-stability boundaries are hyperbolic and finite in number, and ii) the stable and unstable manifolds of the equilibrium points on quasi-stability boundaries satisfy the transversality condition. If xd is a DDP on the quasi-stability boundary ∂Ap(xs), then there exists another one and only one SEP, says
Theorem 1.4 reveals a relationship between SEPs and DDPs. The unstable manifold of a DDP converges to two SEPs of the KKT dynamical system (E.8). In the context of optimization, Theorem 1.3 asserts that two neighboring critical points are connected by the unstable manifolds of the corresponding DDP. Note that the two conditions stated in Theorem 1.3 are generic properties for general nonlinear dynamical systems.
In the following, a numerical TT-IPM method is presented. Given a local optimal solution (e.g., xs0) and a set of search directions, The TT-IPM method computes tier-one local optimal solutions of xs0. Each tier-one local optimal solution is computed via computing the corresponding DDPs and then following the unstable manifold of each DDP and via carrying out IPMs.
TRUST-TECH Method for Computing Tier-One Local Optimal Solutions. Input: an initial local optimal solution xs0. Output: the set of tier-one local optimal solutions Vs of xs0. The following is the algorithm:
Step 1: Initialize the set of DDPs Vd= and the set of tier-one local optimal solutions Vs=.
Step 2: Define a set of search paths Si for xs0, i=1, 2, . . . , mj, and set i=1.
Step 3: For each search path Si starting from xs0, compute its corresponding local optimal solution using the following steps:
i) Compute a DDP to the KKT dynamical system to find the corresponding DDP. If a DDP is found, proceed to the next step; otherwise, go to step vi).
ii) Letting the found DDP be denoted as xdi, check if it belongs to the set Vd, i.e., xdi∈Vd. If it does, go to step vi); otherwise, set Vd=Vd ∪{xdi} and proceed to the next step.
iii) Set x0i=xs0+(1+ε)(xdi−xs0), where ε is a small number.
iv) Starting from x0i, conduct a transitional search by integrating the KKT dynamical system to obtain the corresponding system trajectory until it reaches an interface point, say xfi, at which IPMs outperforms the transitional search.
v) Starting from the interface point xfi chosen in step iv), apply an IPM to find the corresponding local optimal solution with respect to xdi, denoted as xsi. And set Vs=Vs∪{xsi}.
vi) Set i=i+1 and repeat steps i) through vi) until i>mj.
A TRUST-TECH Method for Computing All Local Optimal Solutions and the Global Optimal Solution.
The method 300 begins with receiving an initial point (block 301) and performing the following steps:
Step 0: Construct the KKT dynamical system associated with the constrained optimization problem (P.1) (block 302).
Step 1: Starting from x0, apply an IPM to find a local optimal solution, i.e., xs0, to the constrained optimization problem (P.1), which is also an SEP of the constructed KKT dynamical system (block 303).
Step 2: Set j=0. Initialize the set of local optimal solutions Vs={xs0}, the set of tier-j local optimal solutions Vnewj={xs0}, and the set of found DDPs Vd=.
Step 3: Initialize the set of tier-(j+1) local optimal solutions Vnewj+1=.
Step 4: For each local optimal solution in Vnewj (e.g., find all of its tier-one local optimal solutions. The method 300 further includes the following sub-steps for Step 4:
(i) Compute a set of search paths Sij for xsj, i=1, 2, . . . , mj and set i=1 (block 304). An exemplar search path can be a line originated from xsj. In one embodiment, the set of search paths Si={±v1, ±v2, . . . , ±vn} includes all eigenvectors v1, . . . , vn of the Jacobian matrix ∇F(xsi) of the KKT dynamical system computed at xsi. In another embodiment, the set of search directions Si={±v1, ±v2, . . . , ±vn} further includes random orthogonal directions, where each search path vi is a random vector satisfying the orthogonal conditions, that is, ∥vi∥=1 and viTvj=0, for all i≠j.
(ii) For each search path Sij starting from xsj, apply an effective method to nonlinear system (E.6) to find the corresponding exit point (block 305). An example of a method for finding the exit point can be the direct line search method. If an exit point is found, proceed to sub-step (iii); otherwise, go to sub-step (viii).
(iii) Denote the found DDP as xd,ij, check if it belongs to the set Vd, i.e., xd,ij∈Vd. If it does, go to sub-step (viii); otherwise, set Vd=Vd∪{xd,ij} and proceed to sub-step (iv).
(iv) Set x0,ij=xsj+(1+ε)(xd,ij−xsj), where ε is a small number (block 306).
(v) Starting from x0,ij, conduct a transitional search by integrating the KKT dynamical system to obtain the corresponding system trajectory until it reaches an interface point. xƒ,ij, at which IPMs outperform the transitional search.
(vi) Starting from the interface point xƒ,ij chosen in sub-step (v), apply an IPM to find the corresponding local optimal solution, denoted as xs,ji (block 307).
(vii) Check if xsji has been found before, i.e., xs,ji ∈Vs. If it is a new local optimal solution, set Vs=Vs ∪{xs,ji} and Vnewj+1=Vnewj+1∪{xs,ji}.
(viii) Set i=i+1 (block 309) and repeat sub-steps (ii) through (viii) until i>mjk (i.e., all directions have been searched) (block 308).
Step 5: If the set of all newly computed local optimal solutions, namely Vnewi, is empty (block 311), continue with Step 6; otherwise set j=j+1 (block 310) and go to Step 3.
Step 6 (block 312): Output the set of local optimal solutions Vs (block 313) and identify the best optimal solution from the set of Vs by comparing their corresponding objective function values and selecting the one with the smallest value. Output the smallest value as the global optimal solution (block 314).
Step 1: Receive the constrained nonlinear optimization problem (P.1) (block 401).
Step 2: Construct the penalized Lagrange function (E.4) with an augmented set of primal and dual variables (x, y, z, s, μ) (block 402).
Step 3: Construct the system of nonlinear equations (E.5) for the first-order KKT optimality conditions (block 403).
Step 4: Construct a KKT dynamical system associated with the optimality conditions of the constrained optimization problem (P.1) following the formulation (E.7) or the formulation (E.8) (block 404).
In another embodiment, referring to
Step 1: Receive the constrained nonlinear optimization problem (P.1) (block 411).
Step 2: Construct the penalized Lagrange function (E.4) with an augmented set of primal and dual variables (x, y, z, s, μ) (block 412).
Step 3: Construct the system of nonlinear equations (E.5) for the first-order KKT optimality conditions (block 413).
Step 4: Construct the modified system of nonlinear equations (E.6) using complementarity function for the first-order KKT optimality conditions (block 414).
Step 5: Construct the KKT dynamical system associated with the optimality conditions of the constrained optimization problem (P.1) following the formulation (E.7) or the formulation (E.8) (block 415).
It is noted that not all SEPs of the KKT dynamical system correspond to local optimal solutions to the optimization problem (P.1).
(a) Receive a complete set of SEPs, Vs={wsi=[xsi, λsi,ssi], i=1, . . . , K}, of the KKT dynamical system (E.6) associated with the optimization problem (P.1) (block 501).
(b) Initialize the set of local optimal solutions, Vx=, to the optimization problem (P.1), and set i=1.
(c) Compute the energy value E(wsi) at the SEP wsi and check the energy value of wsi (block 502).
(d) If E(wsi)<ε, where ε is a small number indicating the numerical tolerance for optimality (block 503), proceed to step (e); otherwise, the SEP wsi is not a local optimal solution to the optimization problem (P.1), proceed to step (g).
(e) Compute the Hessian matrix at the SEP wsi, and compute all eigenvalues of the Hessian matrix (block 504).
(f) If all eigenvalues of the Hessian matrix have positive real part (block 505), the SEP xsi is a local optimal solution, set Vx=Vx∪{xsi} and proceed to step (g). Otherwise, the SEP xsi is not a local optimal solution to the optimization problem (P.1), proceed to step (g).
(g) If i=K (block 508), proceed to step (h); otherwise, set i=i+1 (block 509) and proceed to step (c).
(h) Output the complete set of local optimal solutions Vx to the constrained nonlinear optimization problem (P.1) (block 510).
The following example presents a procedure for computing a complete set of local optimal solutions to a 5-dimension optimization problem:
The procedure is an example of applying the method(s) of
In this example, there are six local optimal solutions to (P.2), which are x1 through x6 illustrated in
1) Set the initial point x0=[0.0, 0.0, 0.0, 0.0, 0.0]T, the objective function is ƒ(x0)=1.0. This initial point is not feasible (the equality constraints are not satisfied).
2) Using x0 as the initial point, apply the IPM and attain a solution xs0=[4.2798, −1.3675, 0.4528, 2.4010, 0.4673]T. The KKT energy of this point is 31.7999 and its objective function ƒ(xs0)=65.0048. This point is not an actual local optimal solution, since the KKT energy is not close to 0 (
3) Tier-1 Search: Using xs0 as the central point, search along different directions on the KKT energy surface and find exit points. Using the exit points as the initial conditions, apply the IPM to solve problem (E.8). Two new solutions are obtained, which are: (i) xs1=[4.5696, −1.2522, 0.4718, 2.3032, 0.4377]T with KKT energy being 4.7217×10−10 and ƒ(xs1)=64.8719; (ii) xs2=[0.7280, −2.2452, 0.7795, 3.6813, 2.7472]T with KKT energy being 9.1276×10−13 and ƒ(xs2)=52.9067. This process is illustrated in
4) Tier-2 Search: Using xs1 and xs2 as the starting points, search along different directions on the KKT energy surface and find exit points. Using the exit points as the initial conditions, apply the IPM to solve problem (E.8). Two new solutions are obtained, which are: (i) xs3=[1.1166, 1.2204, 1.5378, 1.9728, 1.7911]T with KKT energy being 4.3324×10−15 and ƒ(xs3)=0.0293; (ii) xs4=[−2.7909, −3.0041, 0.2054, 3.8747, −0.7166]T with KKT energy being 1.2552×10−7 and f(xs4)=606.9965. This process is illustrated in
5) Tier-3 Search: Using xs3 and xs4 as the central points, search along different directions on the KKT energy surface and find exit points. Using the exit points as the initial conditions, apply the IPM to solve problem (E.8). Two new solutions are obtained, which are: (i) xs5=[−1.2731, 2.4104, 1.1949, −0.1542, −1.5710]T with KKT energy being 8.9774×10−14 and ƒ(xs5)=27.8730; (ii) xs6=[−0.7034, 2.6357, −0.0964, −1.7980, −2.8434]T with KKT energy being 1.7806×10−15 and ƒ(xs6)=44.0225. This process is illustrated in
The entire procedure of
The above example demonstrates the capability of the TT-IPM method and the following advantages of incorporating TRUST-TECH into the existing local optimization methods:
1) Local optimization methods, such as IPMs, can converge to a local optimal solution and can be entrapped in the local optimal solution.
2) Local optimization methods, such as IPMs, can sometimes converge to a point which is not a true local optimal solution.
3) The TT-IPM method integrates TRUST-TECH'S capability to locate multiple stability regions in the KKT dynamical system and IPM's capability to fast compute a local optimal solution, thus help local methods, such as IPMs, to escape from a local optimal solution and approach another local optimal solution.
TRUST-TECH based methods are valuable and can be of significant importance in the following sense: since a local method, such as IPMs, can only provide local optimal solutions in a restricted local region while a global method can only provide sparse approximated solutions within reasonable computational efforts, the TRUST-TECH based methods provide an enabling technology for:
(i) enhancing the functionality of local methods to find optimal solutions in a broader region, as described in the TT-IPM method;
(ii) enhancing the functionality of global methods to find global optimal solutions; and
(iii) providing flexibility in the integration of local and global methods to find global optimal solutions.
The method of
Embodiments of the methods disclosed herein may be implemented in hardware, software, firmware, or a combination of such implementation approaches. In one embodiment, the methods 200, 300, 400, 410, 500 and 1400 (
The computer system 1600 includes a processing device 1602. The processing device 1602 represents one or more general-purpose processors, or one or more special-purpose processors, or any combination of general-purpose and special-purpose processors. In one embodiment, the processing device 1602 is adapted to execute the operations of a global optimizer module 1650, which when executed causes the processing device 1602 to perform the functions of the global optimizer module 100 of
In one embodiment, the processing device 1602 is coupled, via one or more buses or interconnects 1630, to one or more memory devices such as: a main memory 1604 (e.g., read-only memory (ROM), flash memory, dynamic random access memory (DRAM), a secondary memory 1618 (e.g., a magnetic data storage device, an optical magnetic data storage device, etc.), and other forms of computer-readable media, which communicate with each other via a bus or interconnect. The memory devices may also different forms of read-only memories (ROMs), different forms of random access memories (RAMs), static random access memory (SRAM), or any type of media suitable for storing electronic instructions. In one embodiment, the memory devices may store the code and data of the global optimizer module 1650. In the embodiment of
The computer system 1600 may further include a network interface device 1608. A part or all of the data and code of the global optimizer 100 may be transmitted or received over a network 1620 via the network interface device 1608. Although not shown in
In one embodiment, the global optimizer module 1650 can be implemented using code and data stored and executed on one or more computer systems (e.g., the computer system 1600). Such computer systems store and transmit (internally and/or with other electronic devices over a network) code (composed of software instructions) and data using computer-readable media, such as non-transitory tangible computer-readable media (e.g., computer-readable storage media such as magnetic disks; optical disks; read only memory; flash memory devices as shown in
The operations of the methods and/or procedures described in connection with
While the invention has been described in terms of several embodiments, those skilled in the art will recognize that the invention is not limited to the embodiments described, and can be practiced with modification and alteration within the spirit and scope of the appended claims. The description is thus to be regarded as illustrative instead of limiting.