The present invention relates to methods, systems, and computer program products for modeling circuits and/or systems, and, more particularly, to systems, methods, and computer products for modeling time-invariant and time-varying nonlinear circuits and/or systems.
Over the past decade a relatively large body of work on model order reduction of IC interconnects has emerged from the design automation community. One purpose of model reduction is to generate models that are orders of magnitude smaller than the original system while more accurately approximating the input/output relationships. Compared to the success of model order reduction for linear time invariant (LTI) RLC networks, the problem of reducing nonlinear systems has generally been less understood and explored. There are numerous applications, however, where abstracting transistor-level circuit details that include important weakly nonlinear effects into a compact macromodel may be important. For instance, in RF communication IC design there is a growing interest in extracting efficient circuit-level models that are capable of capturing system nonlinear distortion. While circuit blocks in these applications may exhibit weak nonlinearities, the design specification for linearity may be important and/or stringent. As depicted in
A piecewise-linear approximation based nonlinear systems reduction has been proposed where a set of linearizations about the state trajectory due to a training input is used to model a nonlinear system and each linearization is reduced using Krylov projection. While having the potential capability of handling large nonlinearities, such as may happen in a micromachined device, a potential limitation of this approach is its training input dependency.
A different model reduction direction has been taken based on the Volterra functional series that have been used as a means for analyzing weakly nonlinear systems. The application of the Volterra series may make it possible to solve the circuit response of a weakly nonlinear system from low orders to high orders via a recursive procedure, commonly referred to as nonlinear current method. Frequency domain characterizations based on Volterra kernels or nonlinear transfer functions describe input-independent system nonlinear properties in a manner analogous to the use of linear transfer functions for linear system properties. Projection based nonlinear model order reduction frameworks have been proposed for making automated nonlinear system reduction possible by extending the popular implicit moment-matching projection techniques used for interconnects. Some approaches use the Taylor series expansion of nonlinear state equations to represent a weakly nonlinear system. Then the nonlinear system model inherited from the recursive nonlinear current method is used to view a weakly nonlinear system as a set of interconnected linear networks. A benefit of this perspective is that it may allow the use of existing linear system reduction techniques to serve as blackbox tools to reduce each individual linear network for achieving the goal of overall nonlinear system reduction. Converting a nonlinear reduction problem to several linear reduction problems, however, may leave some issues unaddressed. For example, how does the quality of each linear reduction problem impact the accuracy of the overall nonlinear model? What is the best strategy in carrying out each linear reduction for achieving good nonlinear model accuracy?
Another approach that may provide a theoretical foundation uses the bilinear form of a nonlinear system and explicitly matches the moments of nonlinear transfer functions analogously to Padé approximations of LTI systems. The use of bilinear form may allow the nonlinear transfer function moments to be expressed in a structured fashion such that a projection based moment-matching reduction scheme can be derived. By converting a standard state-equation form to its bilinear counterpart, however, a relatively large number of additional state-variables may be introduced. For example, the bilinear form of a system including up to the third order nonlinear coefficient matrices has O(N3) state variables, where N is the number of state variables in the original state equation form.
Under this framework of nonlinear model order reduction, the reduced model compactness facilitates effective model reduction. As the order of moment matching increases, the size of the reduced order model increases more rapidly. At the same time, as the size of the reduced model grows, it may become more costly to explicitly form the resulting high order system matrices of the reduced order model, thereby reducing the possible benefits of model reduction.
Background of Volterra Series
Time invariant Volterra series, which may be used for time-invariant systems will now be briefly discussed. For simplicity, consider a single-input multi-output system described by the following Modified Nodal Analysis (MNA) formulation:
where x is the vector of node voltages and branch currents, u is the input to the system,f() and q() are nonlinear functions relating currents of nonlinear resistors and nonlinear charges/fluxes with x; y is the vector of the output; and b and d are the input and output matrices respectively. When the input is small, only weakly nonlinear behavior is excited in the system. Assume the system is perturbated about the DC bias condition due to a small-signal input. Using a Taylor series to expand f() and q() at the bias point x0, and considering only small-signal quantities, we obtain
where {circle over (x)} is the Kronecker product operator, x(t) and u(t) are the small signal response and the input to the system, and
are the ith order conductance and capacitance matrices, respectively.
A nonlinear system can be analyzed using Volterra functional series under weakly nonlinear conditions. In a Volterra series, the response x(t) can be expressed as a sum of responses at different orders
where, xn is the n-th order response. The order of a response component specifies the cumulative number of multiplications of the input signal for generating the corresponding response. A converging Volterra series according to (3) may be truncated to a finite number of terms and each order of response may be solved recursively. The first order or linear response of the weakly nonlinear system in (2) may be obtained by retaining only first-order terms in (2).
Equation (4) represents a linear time-invariant (LTI) system. Higher order responses may be computed by solving the linearized system with different inputs. The second and the third order responses are given by the following two equations using vector Kronecker products of x:
In Equations (5) and (6), Gi and Ci are the ith order conductance and capacitance matrices respectively, and
Systems specified by Equations (4), (5), and (6) may be referred to as the linearized (first order), second order, and third order systems for Equation (2), respectively.
Alternatively, a weakly nonlinear system can be analyzed in the frequency domain, which is generally more suitable for a spectrum analysis. As an extension to the impulse response function of a LTI (linear time invariant) system, an nth order response can be related to a Volterra kernel of order nhn(τ1, . . . , τ2)
hn(τ1, . . . , τn) can also be transformed into the frequency domain via a Laplace transform
and referred to as the nonlinear transfer function of order N. The nth order response, yn, can be related to the input as
Based on Equations (7) and (9), it can be seen that Volterra kernels and nonlinear transfer functions specify the system response due to an arbitrary input. Thus, they are input-independent properties of the system, and may fully describe the nonlinear system behaviors. As such, if the original nonlinear transfer functions are approximated in a reduced order model using a method such as moment matching, then the reduced model may accurately model the original nonlinear system properties.
Existing Reduction Techniques
A projection based nonlinear model order reduction approach has been proposed in J. Roychowdhury, “Reduced-order modeling of time-varying systems,” IEEE Trans. Circuits and Systems II: Analog and Digital Signal Processing, vol. 46, no. 10,October, 1999 (hereinafter “Roychowdhury”), and J. Phillips, “Automated extraction of nonlinear circuit macromodels,” Proc. of CICC, 2000 (hereinafter “Phillips”), the disclosures of which are hereby incorporated herein by reference, where reduction techniques for LTI systems are extended to take the weakly nonlinear aspects of a system into consideration. For simplicity of description, consider the expanded state-space model of the nonlinear system in Equation (1)
Similar to Equations (4) through (6), consider the first order through the third order nonlinear responses
Equations (11) through (13) represent three LTI systems with inputs formed by ux1{circle over (x)}x1 and
respectively. Each of these LTI systems may be reduced to a smaller system using various projection methods. If the system of Equation (11) with n states is reduced to a system with q1 states through a Krylov subspace projection x1=Vq1{overscore (x)}1, then the inputs to Equation (12) can be approximated by Equation 14 below:
This suggests that Equation (12) now can be viewed as having q12 inputs instead of n2 inputs, and can be reduced by a Krylov projection x2=Vq2·{overscore (x)}2 to a smaller system with q2 states. Now the number of inputs to (13) is reduced from n2+n3 to q13+q1q2. Equation (13) can be reduced to a smaller LTI system with q3 states via projection x3=Vq3·{overscore (x)}3. The final reduced nonlinear model can either be in the form of a collection of reduced linear systems, or be described by a smaller set of nonlinear equations using variable embedding x=V·{overscore (x)} where V is an orthonormal basis of [Vq1,Vq2,Vq3]. For the latter case, the third order matrix can be reduced to a matrix of a smaller dimension, such as {overscore (A)}3=VA3(V{circle over (x)}V{circle over (x)}V). However, the reduced third order matrix is usually dense and has O(q4) entries, where q is the number of states of the reduced model. Therefore, achieving model compactness may be effective for model reduction.
An advantage of the above approach is that existing LTI techniques may be applied directly as a black-box tool. There are three major problems that may be associated with this direct extension to LTI based reductions. First, this approach may reduce a weakly nonlinear system by approximating underlying linear subsystems. The assumption is that overall nonlinear behavior of the original system will be closely replicated in the reduced system if the corresponding linear circuits are well approximated. Nevertheless, it may be difficult to assess the quality of the nonlinear system reduction. As a consequence, it remains unclear how the qualities of reduced linear subsystems interact with overall quality of the reduced nonlinear model. Furthermore, there may be a lack of guidance on how to optimize nonlinear model quality by choosing appropriate orders for the reduced linear blocks. Second, it can be seen that this method models a nonlinear system, potentially with a small number of inputs, as several linear systems with many more inputs. Intuitively, this increase in the degree of freedom may lead to unnecessarily large reduced models. Finally, as an extension to LTI based reduction, this approach may not explore flexible selections of expansion points for approximating nonlinear transfer functions. For instance, to capture accurately the third order intermodulation around the center frequency f0 of a narrow-band amplifier, it may be desirable to expand the third transfer function H(s1,s2,s3), at an expansion point (j2πf0,j2πf0,−j2πf0). This choice of expansion point is relatively difficult to incorporate in the aforementioned reduction approach.
According to some embodiments of the present invention, a nonlinear system may be modeled by obtaining a transfer function for the nonlinear system and generating a Taylor series expansion of the transfer function. The Taylor series expansion comprises a plurality of moments respectively corresponding to a plurality of coefficients of the Taylor series terms. At least one Krylov subspace is derived that matches at least one of the plurality of moments. The nonlinear system is modeled using the at least one Krylov subspace.
In other embodiments of the present invention, derivnig the at least one Krylov subspace comprises selecting an order k of the plurality of moments to be matched and deriving at least one Krylov subspace that matches at least one of the kth order moments of the plurality of moments.
In still other embodiments of the present invention, a plurality of sample points is selected and the k+1th order Transfer function is generated at each of the plurality of sample points. The contributions of non-linear elements in the k+1th order Transfer function generated at each of the plurality of sample points are determined. Selected non-linear elements from the k+1th order Transfer function that have lower contributions to the k+1th order Transfer function, for example, are discarded to obtain a pruned system that approximates the k+1th order Transfer function. The nonlinear system is modeled using the at least one Krylov subspace and the pruned system that approximates the k+1th order Transfer function.
In still other embodiments of the present invention, the Taylor series expansion is genearted by generating a plurality of Taylor series expansions of the Transfer function about a plurality of expansion points. Each of the plurality of Taylor series expansions comprises a plurality of moments. An order k of the plurality of moments to be matched is selected based on a number of the plurality of expansion points. At least one Krylov subspace is derived that matches at least one of the kth order moments of respective ones of the plurality of moments for the plurality of expansion points.
In still further embodiments of the present invention, the Taylor series expansion is generated by generating a plurality of Taylor series expansions of the Transfer function about a plurality of expansion points. Each of the plurality of Taylor series expansions comprising a plurality of moments. At least one Krylov subspace is derived that matches at least one of the plurality of moments for the plurality of expansion points.
In still further embodiments of the present invention, the transfer function comprises a single variable transfer function component and a multi-variable transfer function component.
In still furthe embodiments of the present invention, respective ones of the plurality of moments are matrices.
Other features of the present invention will be more readily understood from the following detailed description of specific embodiments thereof when read in conjunction with the accompanying drawings, in which:
While the invention is susceptible to various modifications and alternative forms, specific embodiments thereof are shown by way of example in the drawings and will herein be described in detail. It should be understood, however, that there is no intent to limit the invention to the particular forms disclosed, but on the contrary, the invention is to cover all modifications, equivalents, and alternatives falling within the spirit and scope of the invention as defined by the claims. Like numbers refer to like elements throughout the description of the figures. It should be further understood that the terms “comprises” and/or “comprising” when used in this specification is taken to specify the presence of stated features, integers, steps, operations, elements, and/or components, but does not preclude the presence or addition of one or more other features, integers, steps, operations, elements, components, and/or groups thereof. As used herein, the term “and/or” includes any and all combinations of one or more of the associated listed items.
While the invention is susceptible to various modifications and alternative forms, specific embodiments thereof are shown by way of example in the drawings and will herein be described in detail. It should be understood, however, that there is no intent to limit the invention to the particular forms disclosed, but on the contrary, the invention is to cover all modifications, equivalents, and alternatives falling within the spirit and scope of the invention as defined by the claims. Like reference numbers signify like elements throughout the description of the figures.
The present invention may be embodied as methods, data processing systems, and/or computer program products. Accordingly, the present invention may be embodied in hardware and/or in software (including firmware, resident software, micro-code, etc.). Furthermore, the present invention may take the form of a computer program product on a computer-usable or computer-readable storage medium having computer-usable or computer-readable program code embodied in the medium for use by or in connection with an instruction execution system. In the context of this document, a computer-usable or computer-readable medium may be any medium that can contain, store, communicate, propagate, or transport the program for use by or in connection with the instruction execution system, apparatus, or device.
The computer-usable or computer-readable medium may be, for example but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, device, or propagation medium. More specific examples (a nonexhaustive list) of the computer-readable medium would include the following: an electrical connection having one or more wires, a portable computer diskette, a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or Flash memory), an optical fiber, and a portable compact disc read-only memory (CD-ROM). Note that the computer-usable or computer-readable medium could even be paper or another suitable medium upon which the program is printed, as the program can be electronically captured, via, for instance, optical scanning of the paper or other medium, then compiled, interpreted, or otherwise processed in a suitable manner, if necessary, and then stored in a computer memory.
Referring now to
The processor 108 communicates with the memory 106 via an address/data bus. The processor 108 may be, for example, a commercially available or custom microprocessor. The memory 106 is representative of the overall hierarchy of memory devices containing the software and data used to model inductive effects in a circuit, in accordance with embodiments of the present invention. The memory 106 may include, but is not limited to, the following types of devices: cache, ROM, PROM, EPROM, EEPROM, flash, SRAM, and DRAM.
As shown in
Although
Computer program code for carrying out operations of the present invention may be written in an object-oriented programming language, such as Java, Smalltalk, or C++. Computer program code for carrying out operations of the present invention may also, however, be written in conventional procedural programming languages, such as the C programming language or compiled Basic (CBASIC). Furthermore, some modules or routines may be written in assembly language or even micro-code to enhance performance and/or memory usage. It will be further appreciated that the functionality of any or all of the program modules may also be implemented using discrete hardware components, a single application specific integrated circuit (ASIC), or a programmed digital signal processor or microcontroller.
The present invention is described hereinafter with reference to flowchart and/or block diagram illustrations of methods, systems, and computer program products in accordance with exemplary embodiments of the invention. It will be understood that each block of the flowchart and/or block diagram illustrations, and combinations of blocks in the flowchart and/or block diagram illustrations, may be implemented by computer program instructions and/or hardware operations. These computer program instructions may be provided to a processor of a general purpose computer, a special purpose computer, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions specified in the flowchart and/or block diagram block or blocks.
These computer program instructions may also be stored in a computer usable or computer-readable memory that may direct a computer or other programmable data processing apparatus to function in a particular manner, such that the instructions stored in the computer usable or computer-readable memory produce an article of manufacture including instructions that implement the function specified in the flowchart and/or block diagram block or blocks.
The computer program instructions may also be loaded onto a computer or other programmable data processing apparatus to cause a series of operational steps to be performed on the computer or other programmable apparatus to produce a computer implemented process such that the instructions that execute on the computer or other programmable apparatus provide steps for implementing the functions specified in the flowchart and/or block diagram block or blocks.
With reference to the flowchart of
Returning to
The flowchart of
A detailed mathematical analysis of methods, systems, and/or computer program products for modeling nonlinear systems and/or circuits, in accordance with some embodiments of the present invention, will now be provided.
Matrix-Form Nonlinear Transfer Functions and Moments
To derive the general matrix-form nonlinear transfer functions suitable for model order reduction, without loss of generality we consider Modified Nodal Analysis (MNA) formulation for a single input multiple output (SIMO) based weakly nonlinear system in Equation (1) and its expanded form in Equation (2). The nonlinear transfer functions presented here are the corresponding matrix-form expressions of the analytical nonlinear current method used for computing Volterra kernels. Before proceeding we first introduce the notations used throughout this detailed description.
For matrices in Equation (2) we define
and for an arbitrary matrix F define
We also define the Krylov subspace K1(A,p) corresponding to matrix A and vector (matrix) p as the space spanned by vectors {p, Ap, . . . A1p}.
For the systems in Equations (1) and (2), the first order transfer function for state-variables x is simply the transfer function of the linearized system
(G1+sC1)H1(s)−b, or H1(s)−(G1+sC1)−1b. (17)
Defining {overscore (s)}−s1+s2, it can be shown that the second transfer function is determined by
where
We similarly define
as the arithmetic average of terms of all possible permutations of frequency variables in the Kronecker product. Define {grave over (s)}−s1+s2+s3, then the third order nonlinear transfer function is given by the following equations
Without loss of generality, expand Equation (17) at the origin as a Mclaurin series
where M1,k−Akr1 is the kth order moment of the first order transfer function. Expand H2(s1,s1) at the origin (0,0), we have
where M2,k,1 is the kth order moment for the second order transfer function. To derive the expressions for the moments of H2(s1,s2), substituting Equation (21) into Equation (18) and expanding w.r.t. {overscore (s)}−s1+s2 yields
Comparing (22) with (23), we can express the moments of H2(S1,s2) as
Similarly, the third order transfer function can be expanded at the origin (0,0,0) as
where M3,k,1,m is a kth order moment given by
with
As an example, the first few moments of H2(s1,s2) are shown in Table 1. It can be clearly seen that significantly more moments have to be considered to match a nonlinear transfer function than a linear one while the order of moment matching is kept the same. This means nonlinear model order reduction may be intrinsically more complex and costly than that of LTI systems. Also note that in Equations (18) and (19), symmetric nonlinear transfer functions (symmetric w.r.t frequency variable permutations) are used, which is generally advantageous for model reduction as it reduces the number of moments that may need to be matched for H2 and H3 approximately by a factor of 2 and 6 respectively, when expanding them at a point with equal coordinates such as the origin.
NORM
In this section we present some embodiments of the invention, also referred to herein as NORM, but limit our discussion to SIMO time-invariant weakly nonlinear systems. Up to the third order nonlinear transfer functions are considered in the reduction. Extension to more general time-varying systems can be easily derived using time-varying Volterra series as a system description. To assess the model order reduction quality from a moment matching perspective, we first have the following definition:
To derive a set of minimum Krylov subspaces for the most compact order reduction, understanding the interaction between the moments of nonlinear transfer functions at different orders may be beneficial. For the moment matching of H2(s1,s2), this interaction is manifested in Equation (23). A particular term corresponding to s1ps2q in the expansion of H2, where p, q are integers, is a consequence of the two power series expansions: expansion of H1 in (17) and the expansion w.r.t. {overscore (s)}=s1+s2 of Equation (18). For instance, the kth order moment of H2 depends on moments of H1 with an order less than or equal to k. As such, the final expression for M2,l,m is in the form as shown in Equation (24). The expression for M3,l,m,n is derived similarly in a more complex form in Equation (25).
If we were to use a projection for order reduction, one issue is to find certain Krylov subspaces that contain the directions of all the moments to be matched. To minimize the number of projection vectors needed, nonlinear transfer function moments are considered explicitly. A close inspection of Equation (24) reveals that the Krylov subspaces of matrix A given in Table 2 are the desired Krylov subspaces of the minimum total dimension for constructing a kth order model in H2. For the last row, m=└k/2┘, n=k−└k/2┘, p=└(k−1)/2┘, and q=k−1−(k−1)/2. We denote the union of Krylov subspaces in Table 2 as
where k in the parenthesis indicates the order of H2 moments up to which contained in the subspaces.
In an analogous and somewhat more involved way, a set of minimum Krylov subspaces K3(k) for moment-matching of H3 up to the kth order can also be derived. More specifically, Equations 26 and 27 show that the subspace
consists of a collection of Krylov subspaces of A, where the starting vector vi takes one of the following forms set forth below in Equations (28) and (29):
with
Using these subspaces, Theorem 1 may be proved below as follows:
The complete single-point version of the NORM methodology (expanded at the origin) is shown in
Also note that in
Size of Reduced Order Models
Without causing ambiguity, we define the size of a state-equation model in terms of its number of states. Consider the size of the SIMO based reduced order models generated using the methods of Roychowdhury and Phillips under Definition 1. It can be shown that if the linear networks described by Equations (11)-(13) are each reduced to a system preserving up to k1, k2, and k3th order moments of the original system, respectively, then the overall reduced nonlinear model is a k1th order model in H1, a {overscore (k)}2th order model in H2, where {overscore (k)}2=min(k1,k2) and a {overscore (k)}3th order model in H3, where {overscore (k)}3=min(k1,k2,k3). For example, from Equation (23), the final moment expressions of H2 are a consequence of expanding with respect to s in Equation (17) and expanding with respect to {overscore (s)} in Equation (18). Therefore, the lesser of k1 and k2 determines the order up to which the moments of H2(s1,s2) are matched. This reveals that for this method, the most compact kth order model in H2 with a size in O(k3) is achieved by choosing k1=k2=k. In other words, choosing k2>k1=k does not necessarily increase the number of moments of H2 matched in the reduced model. On the other hand, if one would like to have k1>k2 to increase the accuracy of H1, one way is to only make use of the first k2·+1 moment directions of H1 for reducing H2 while the remaining H1 moments are included in the projection only for matching H1 itself, i.e., only a k2·+1th reduced model of the system in Equations (4) or (11) is computed when reducing the second order system in Equations (5) or (12). Similar strategies apply to reducing Equation (13) for the moment matching of H3 such as the optimal way to generate a kth order model in H3 is to choose k1=k2=k3=k with a resulting model size of O(k5). Note that these strategies are also used in the single point NORM algorithm. To compare with the above “optimal” model sizes achievable using the methods of Roychowdhury and Phillips, using the single-point NORM, the sizes of a kth order model in H2 and H3 are in O(k3) and O(k4), respectively. The exact model sizes for several values of k will be shown in the following section.
Multi-Point Expansions
To target a system's particular input frequency band of interest, particularly for RF circuits, it may be desirable to expand both linear and nonlinear transfer functions at points other than the origin such as along the imaginary axis. Under nonlinear context, a single-frequency expansion point along the imaginary axis may inherit multiple matrix factorizations due to the nonlinear frequency mixing effects. To this end, suppose that the in-band third order intermodulation of a nonlinear system around center frequency fo is important to model. To build the most compact model, one would opt to expand H1(s) at s=j2πfo, but to correctly perform moment matching for H2 and H3, the respective expansion points for high order transfer functions should be (j2πfo,j2πfo) and (j2πfo,j2πfo) for H2 , and (j2πfo,j2πfo,−j2πfo) for H3, respectively. Here the use of two expansion points for H2 takes care of matching second order mixing effects in terms of both sum and difference frequencies around the center frequency, and also ensures the moment matching of the third order in-band intermodulations. These choice of expansion points require the system matrix factorized at DC and j4πfo but not j2πfo Equations (12)-(13) or (18)-(19). Therefore, if every linear system in Equations (11 )-(13) is reduced by expanding at s=j2πfo using conventional methods, then there is no guarantee for matching any moments H2 and H3.
In addition to the benefit of using a specific expansion point, a multiple-point projection where several expansion points are used simultaneously has a unique efficiency advantage in terms of model compactness over the single-point method. Although multiple-point methods may not bring such an advantage for LTI system reduction, adopting multiple-point methods for nonlinear systems can lead to significantly smaller reduced models. To see this, notice that in Table 3 the size of the nonlinear reduced order model grows faster than the order of moment matching.
This stems from the fact as the order of moment matching proceeds, significantly more moments corresponding to various expansion terms emerge for multiple-variable high order transfer functions. However, it is much more revealing to recognize that the numbers of up to kth order moments of H2 and H3 are k(k+1)/2 and O(k3) respectively, i.e., the dimension of the subspaces used in the reduction grows even faster than the number of moments matched in the single-point expansion version of NORM. Using the bilinear form can produce models with a size proportional to the number of moments matched; however, this may be offset by the inflated problem size and the accuracy degradation for reducing a significantly larger system. In contrast, to preserve the value of a nonlinear transfer function at a specific point (a zeroth-order moment), for example, only one vector needs to be included in the projection assuming the dependency between transfer functions of different orders is resolved properly, i.e., the reduced model size of a zero-th order multiple-point method is the same as the number of moments matched. For the state-equation form of Equation (2), we compare the worst-case reduced model sizes generated by the methods listed in Table 3, where each method is used to match the moments of H2 or H2 & H3 up to kth order. In the table, the “optimal” strategies outlined in previous sections are used for the conventional methods. NORM(sp) denotes the single-point based NROM, and NORM(mp) is the “equivalent” zeroth-order multiple-point method preserving the same total number of moments. As clearly shown, using NORM the model compactness has been significantly improved.
In practice one would trade-off computational efficiency with the model compactness by using a multiple-point based approach where low-order moment-matching is performed at several expansion points. The added computational cost incurred by more matrix factorizations in this approach might be alleviated by exploiting the idea of recycled Krylov-subspace vectors for time-varying systems. It is also possible to use constant Jacobian iterations to iteratively solve the resulted linear problems for narrow band systems, such as certain RF applications. The LU factorization at one expansion point might be reused at another point not far apart as an approximated Jacobian in the iteration. One extra advantage of multiple-point methods is that they help to eliminate the potential numerical stability problem of the single-point methods for high order moment matching.
Numerical Results
Using various examples, NORM modeling of systems (single-point NORM (NORM-sp) and zeroth-order multiple-point NORM (NORM-mp)) has been compared with the modeling methodologies described in Roychowdhury and Phillips. For the Roychowdhury/Phillips method and single-point NORM method, the origin was chosen as the expansion point. Note that the strategies presented in the section entitled “Size Of Reduced Order Models” were applied to the Roychowdhury/Phillips method; otherwise, significantly larger models resulted with little accuracy improvement. For all three methods, an SVD procedure follows the computation of Krylov subspace vectors to deflate the subspaces as much as possible.
A Diode Circuit Example
We first consider a diode circuit as shown in
NORM-sp matches 5 moments of H1, 9 moments of H2, 4 moments of H3, resulting in a reduced model with 23 stages. NORM-mp produces a model with 18 states while matching 5 moments of H1, 10 moments of H2 and 4 moments of H3. As can be seen, NORM is able to generate a smaller model while matching more moments due to its improved selection of Krylov subspace vectors. The second order transfer functions of H2(j2πf1, j2πf2), where 0≦f1,f2≦50 MHz is plotted based on the full model in
A Double-Balanced Downconversion Mixer
For more realistic examples, next we consider a standard double-balanced mixer shown in
In the first simulation (
Due to the relatively large size of resulting third order matrices, the 60-state model of the Roychowdhury/Phillips method brought less than 5× runtime speedup over the full model. However, for the much smaller models produced by NORM-sp and NORM-mp, significant runtime speedups of 350˜380× and 730˜840× were achieved, respectively.
A 2.4 GHz Subharmonic Direct-Conversion Mixer
A 2.4 GHz subharmonic direct-conversion mixer used in WCDMA applications is shown in
The model produced by the Roychowdhury/Phillips method has 122 states and matches 4 moments of H1, 6 moments of H2, and 2 moments of H3. The maximum relative error for this model is about 14%. For both methods, the origin was used as the expansion point. The better accuracy obtained in the smaller model of NORM-sp can be explained by more moments being matched in the reduced model. We anticipate that both methods will generate more accurate models when the correct procedure is used to expand transfer functions at a point close to the center frequency. Lastly, NORM-mp generates a compact 22-state model with the smallest maximum relative error of 4% while matching 6 moments of H1, 12 moments of H2, and 4 moments of H3. In a two-tone harmonic balance simulation, we applied two RF sinusoidal tones around 2.4 GHz with 2 mv amplitude. Explicitly formed reduced models of NORM-sp and NORM-mp provided a runtime speedup of 237× and 1200× over the full model, respectively. Due to the problem of forming the large reduced third order system matrices, however, it may become inefficient to explicitly form the 106-state model of the Roychowdhury/Phillips method. Only the projection matrix was used to reduce the size of the linear problem solved at each simulation iteration. Consequently, the corresponding models did not provide a significant runtime speedup.
Hybrid Approach
Many circuits in RF and analog signal processing applications have sharp frequency selectivity (e.g. containing high-Q filtering). For these cases, as the nonlinear frequency domain system characteristics may vary dramatically within the band of interest, a full projection-based reduction often requires the use of a significant number of projection vectors for achieving sufficient accuracy. As a consequence, forming reduced high-order system matrices can be very expensive due to the large reduced order model size. Therefore, other types of model simplification may be needed. Although numerous nonlinearities can exist in a circuit, there is often a natural tendency for only a few of them to be dominant due to the specific circuit structure. Thus, identifying these dominant nonlinearities within the original circuit structure can be a very useful component of model generation. In the proposed hybrid approach, to cope with circuit problems such as high-Q circuits, the low-order (first and second) responses are matched using projection while high-order (third) response are approximated by exploiting both the circuit internal structure and projection-based model order reduction.
Approximation of Low-Order Responses
For the hybrid approach, we consider the more general scenario of the periodically time-varying weakly nonlinear systems. Instead of using a time-invariant Volterra description, these systems can be characterized by periodically time-varying nonlinear transfer functions Hn(t,·). As discussed above, H1(t,·) can be discretized using back-Euler based techniques on M sample points {overscore (H)}n=[Hn(t1,·)T,Hn(t2,·)T . . . , Hn(tM,·)T]T within a period of the system time variation T. A set of matrices {overscore (G)}1,{overscore (C)}1,{overscore (b)}1,{overscore (G)}2,{overscore (C)}2,{overscore (G)}3,{overscore (C)}3, can be formulated based on the discretization of the system nonlinear characteristics over a period T for periodically time-varying systems. Using these matrices, Equations 17 through 19 can be reformulated in terms of {overscore (H)}1,{overscore (H)}2,{overscore (H)}3. Each of these equations now has M×N unknowns and can be reduced directly using NORM, where N is the number of physical circuit unknowns of the system.
Assume that M=2K−1, where K is an integer, and consider the application of the NORM methodology to approximate the first and second nonlinear transfer functions. The corresponding projection matrix is denoted as V ε RNM×q, where q is the size of the reduced order model. The first and second order matrices generated by NORM are written as Equation (43) below:
Substituting Equation (43) into Equations (17) and (18), the first and second order transfer functions {tilde over (H)}1 and {tilde over (H)}2 of the reduced model may be determined. The time sampled first and second order transfer functions of the original system can be approximated as set forth in Equation (44) below:
As the reduced model of Equations (43) and (44) represents a time-invariant system, the approximation of the first and second order responses of the original periodically time-varying system involves the use of the discrete Fourier transform. It can be shown that the following qth order system is a time-domain realization of the corresponding reduced model:
where J=D·Γ·V·V,Γ is the DFT matrix converting M time samples to the corresponding Fourier coefficients, D=[e−jk
Approximation of the Third Order Response
In a full projection based approach, nonlinear model order reduction can be limited by the difficulty in explicitly forming the projected (dense) third order system matrices, especially for relatively large model sizes. Pruning based model simplification can be instead performed in the original system coordinates. Thus, the sparsity of the original circuit problem is not lost, but further enhanced. To approximate the third order transfer function based on the reduced model of Equations (43) through (45), Equation (43) maybe substituted into Equation (19) resulting in Equation (46):
where
Based on the Equations above, to compute the third order transfer function, one needs to solve a linear system in terms of {overscore (G)}1 and {overscore (C)}1, and form the input to the linear system in terms of g that is a function of high-order system matrices {overscore (G)}2,{overscore (C)}2,{overscore (G)}3,{overscore (C)}3, and low order transfer functions {overscore (H)}1,{overscore (H)}2. As {overscore (H)}1,{overscore (H)}2, are approximated using Equation (44), to further reduce Equation (46) involves reduction of the dimension of the linear problem in Equation (46) as well as the high-order system matrices {overscore (G)}2,{overscore (C)}2,{overscore (G)}3,{overscore (C)}3. This goal is accomplished by a combination of projection-based reduction of an adjoint network and direct matrix pruning.
In analog signal processing and RF applications, a circuit block usually has only one or at most a few output nodes. For periodically time-varying systems, typically only a small number of sidebands for the outputs are of interest. Thus, Equation (46) may be viewed as a system with potentially many inputs but few outputs, and can be reduced as an adjoint network. Without loss of generality, assume only the ith sideband of the voltage response at node p is of interest (e.g., i=0 pus or minus 1). Define l=[00 . . . 1 . . . 0]T as a N×1 vector that has a 1 at the pth location and zeros at all other locations, and
L=diag(lT,lT, . . . lT) ε RM×NM
d=[1e−j2πi/M . . . e−j2π(M−1)i/M]T/M, ba=LTd (48)
The ith sideband of the third order transfer function at node p, yp
We can apply Krylov subspace projection to reduce the linear adjoint network of Equation (49)
where Va is a orthonormal basis of the Krylov subspace colspan{ra,Aara,A2ara . . .} with Aa=−{overscore (G)}1−T{overscore (C)}1T,ra={overscore (G)}1−Tba. As the DFT vector d is absorbed into ba, to perform a real projection, ba can be split into real and imaginary parts before reduction. Based on Equations (49) and (50), yp
To speed up the computation of the third order transfer function or response, high order system matrices {overscore (G)}2,{overscore (C)}2,{overscore (G)}3,{overscore (C)}3 need to be reduced. To avoid forming dense reduced matrices (particularly for the third order ones) in a projection based approach, the internal structure of the problem may be exploited by applying a direct matrix pruning technique, where nonzero elements of {overscore (G)}2,{overscore (C)}2,{overscore (G)}3,{overscore (C)}3 are pruned according to their contributions to the third order transfer function at the output node of interest. Although a nonlinear circuit may contain many nonlinearities, a few of them tend to be dominant. Therefore, retaining the original circuit structural information by avoiding using projection and applying pruning to these high-order matrices in the original coordinates can be effective for model generation.
The matrix pruning process may be performed as follows in accordance with some embodiments of the present invention: A set of sampled frequency points Ω are selected within the frequency band of interest. For each of these frequency samples, the third order transfer function is computed as well as the individual contributions of nonzeros in {overscore (G)}2,{overscore (C)}2,{overscore (G)}3,{overscore (C)}3. These contributions are sorted by magnitude and non-dominant ones are discarded provided that a user specified error tolerance on H3 is not exceeded. A matrix entry is discarded if and only if its removal does not violate the model accuracy at all the frequency points. The end products of the process are a set of pruned matrices {overscore (G)}2p,{overscore (C)}2p,{overscore (G)}3p,{overscore (C)}3p that satisfy the error tolerance for all the sampled frequency points in Ω. These pruned high order system matrices retain the same dimensions as in the original system while their sparsity can be improved as a result of pruning. As the computation of H3 and individual contributions may take place on many sample points, matrix pruning may be relatively slow. To speed up the process, projection based model reduction may be incorporated into the pruning process.
To compute H3 at a frequency point H1 and H2 need to be computed at related points. According to Equation (47), to form vector g, the reduced order model of Equations (43) and (44) may be used to compute approximate first and second order transfer functions. Then, instead of solving a potentially large linear problem in Equation (46), the reduced adjoint network of Equation (51) can be used to approximate {circumflex over (z)}. Computation of each individual contribution term is based on Equations (47) and (51). For example, the contribution of the nonzero at the (m,n) location of {overscore (G)}3 can be computed as follows:
where {overscore (G)}3,m,n is the value of the nonzero, u1,n is the nth element of u1, and {circumflex over (z)}m is the mth element of {circumflex over (z)}. Because reduced order models are used, factorizing the original large system matrices at many sample points may be avoided. The cost for constructing projection-based reduced order models according to Equations (43) and (51) is generally dominated by the cost of a few matrix factorizations and is therefore bounded by o(kM3N3), where k is the number of matrix factorizations, M is the number of sampled points used to discretize the periodically time-varying transfer functions, and N is the number of physical circuit unknowns. The cost of the pruning process is dominated by evaluation and sorting of the different combinations, assuming that the use of reduced order models makes other costs generally less. The overall cost for model generation is, therefore, bounded by o(sE log(E)+kM3N3), where s is the number of frequency samples evaluated in the pruning, and E is the number of nonzeros in the high order matrices being pruned.
Using the pruned matrices {overscore (G)}2p,{overscore (C)}2p,{overscore (G)}3p,{overscore (C)}3p, Equation (47) can be further approximated as follows:
To derive a time-domain model for generating the desired third order response, Equation (53) is substituted into Equation (51) and both sides of the equation are transposed as follows:
Defining
it may be shown that the corresponding time domain model is
where {grave over (x)}1,{grave over (x)}2 are given in Equation (45), {grave over (y)}p,3(t) is the desired time-domain third order response. Because the ith harmonics of the third order transfer functions are considered the output, {grave over (y)}p,3(t) in the above equation is complex. To recover the corresponding real signal, the corresponding conjugate component may be added. The first and second order responses at node p, {grave over (y)}p,1(t), and e,gra yp,2(t) can be obtained from Equation (45) by selecting proper entries from {grave over (y)}. When more than one output or set of sidebands are of interest, multiple reduced adjoint networks can be incorporated into the model in a straightforward way.
Hybrid Model Results
Switched-capacitor filters are often found in RF receivers as channel-select filters. If the input signal is small, then these circuits can be characterized by periodically time-varying Volterra series. Due to the typical sharp transition between the passband and stopband, it can be very difficult to apply a full-projection based model reduction.
To demonstrate the proposed hybrid approach, consider a Butterworth lowpass switched-capacitor biquad shown in
To apply the hybrid approach, we first used the multi-point NORM algorithm to accurately capture the first and second transfer functions using a reduced order model with 27 states (SVD was used to deflate the Krylov subspaces), where 6 moments of H1 and 24 moments of H2 were matched. The adjoint network describing propagation of the third order nonlinear effects from various nonlinearities to the output was reduced to an 11th order model. Based on these reduced models, {overscore (G)}2,{overscore (C)}2,{overscore (G)}3,{overscore (C)}3 were significantly pruned in the original coordinates at 225 frequency points. Running on an IBM RS6000 workstation, it took 342 CPU seconds to complete the model generation. The overall runtime was dominated by the time spent in pruning. Therefore, the trade-off between runtime and accuracy can be made by selecting an appropriate number of frequency points used for pruning. The original second and third order system matrices have 45,221 and 84,521 nonzeros while the pruned matrices have only 196 and 430 nonzeros, respectively. The final hybrid model has a maximum relative modeling error less than 6% (or about 0.5 dB) for H3 as shown in
This hybrid model was also validated in a frequency-domain Volterra-like simulation using MATLAB, where 6 sinusoidals with various phases were selected from passband, transition band and stopband as input signals. The simulation result was compared against that of the full model, as shown in
Conclusion
The rapid growth of reduced order models for nonlinear time-varying systems, may make model order reduction more difficult than for the case of linear time invariant system order reduction. The new challenge originates because information about more complex nonlinear system effects may be encapsulated in the reduced order model in addition to the first order system properties. The proposed nonlinear system order reduction methodology, NORM, that is based on using a set of minimum Krylov subspaces for moment matching of nonlinear transfer functions, can produce relatively compact macromodels.
Thus, given that modeling of nonlinear systems may result in relatively large models that may be difficult to use and/or require extensive computational resources, some embodiments of the invention can provide more compact reduced order models for nonlinear systems. Some embodiments of the invention begin with a general matrix-form nonlinear transfer functions needed for model order reduction and derive the expressions for nonlinear transfer function moments. This development leads to a deeper understanding on the interaction between Krylov subspace projection and the moment matching under nonlinear context. From this development, embodiments of the invention can provide a new reduction scheme, NORM, which may reduce the size of reduced order models and copes with the model growth problem for nonlinear system representation by using a set of minimum Krylov subspaces.
Some embodiments of the invention can provide features for controlling the growth of reduced order models. First, the modeling accuracy for the nonlinear effect at each order can be selected individually for application-specific needs, while the overall reduced nonlinear model is constructed based on the interactions of moment matching for nonlinear transfer functions of different orders. This may allow the introduction of useless projection vectors for moment matching to be avoided. Second, to target the nonlinear system behavior within the circuit-specific frequency band of interest, the procedure for moment-matching with complex expansion points is accommodated. This may be particularly useful for certain narrow-band RF systems for which it may be beneficial to expand along an imaginary axis about the center frequency. Different from the reduction of an LTI system, however, it can be shown that moment matching of nonlinear transfer functions at a single frequency expansion point inherits multiple matrix factorizations. Therefore, by trading off computational cost with model compactness, some embodiments of the invention can provide a multi-point version of the NORM algorithm that further reduces the total dimension of the Krylov subspaces used in the projection, thereby producing a more compact model.
In concluding the detailed description, it should be noted that many variations and modifications can be made to the preferred embodiments without substantially departing from the principles of the present invention. All such variations and modifications are intended to be included herein within the scope of the present invention, as set forth in the following claims. The following claim is provided to ensure that the present application meets all statutory requirements as a priority application in all jurisdictions and shall not be construed as setting forth the scope of the present invention.
This application claims the benefit of and priority to U.S. Provisional Patent Application No. 60/475,392, filed Jun. 3, 2003, the disclosure of which is hereby incorporated herein by reference as if set forth in its entirety.
Number | Date | Country | |
---|---|---|---|
60475392 | Jun 2003 | US |