None
The invention is in the field of integrated circuit timing analysis.
As the size of the chip increases and the dimension of the transistors moves to nanometer range, process and environment variations become more and more important for the performance of the VLSI design. Variations are usually categorized into global variation (also known as chip-to-chip variation) and local variation (also known as on-chip variation, or OCV). In cutting-edge VLSI designs that are built-upon advanced manufacturing process node, the ratio of local on-chip variation over total variations keep increasing.
In static timing analysis, OCV derating techniques are widely used to model on-chip variations. One problem with OCV derating is that it uses a single number to model all on-chip variations, and consequently introduces more and more pessimism in the advanced technology nodes. To remove some pessimism, Location-based OCV (LOCV) and its derivatives were proposed to introduce logic-depth and location aware OCV techniques (see Hong, J.; Huang, K.; Pong, P.; Pan, J. D.; Kang, J.; Wu, K. C., “An LLC-OCV Methodology for Static Timing Analysis”, VLSI Design, Automation and Test, 2007.
On the other hand, a lot of research has also been done on statistical static timing analysis (SSTA) to address the variation impact (see Hongliang Chang, Sachin S. Sapatnekar: Statistical Timing Analysis Considering Spatial Correlations Using a Single Pert-like Traversal. ICCAD 2003:621-626; Yaping Zhan, Andrej J. Strojwas, Xin Li, Lawrence T. Pileggi, David Newmark, Mahesh Sharma: “Correlation-aware statistical timing analysis with non-Gaussian delay distributions” DAC 2005:77-82).
SSTA is a promising methodology for modeling variations. However, the high cost of its input data preparation, including statistical library characterization and statistical extraction, impedes a rapid, wide adoption of the technology.
Parametric on-chip variation (POCV) is a methodology that provides an on-chip variation model that is derived from statistical static timing methodologies, but that does not require expensive statistical library characterization and statistical RC parasitic extraction (see Ayhan Mutlu, Jiayong Le, Ruben Molina, Mustafa Celik, “A parametric approach for handling local variation effects in timing analysis”, IEEE/ACM International Conference on Computer Aided Design (ICCAD)“, November 2009). However, POCV as taught therein stopped short of addressing the excessive computing resources typically required when processing random variations. The need remains for a method of on-chip variation modeling in statistical timing analysis that is sufficiently low cost so as to encourage widespread and rapid adoption, where cost factors include input data preparation as well as computing resources.
The invention provides an improved methodology for performing timing analysis of integrated circuits. The invention further provides a method for performing statistical timing analysis that uses a novel on-chip variation model, referred to as Sensitivity-based Complex Statistical On-Chip Variation (SCS-OCV).
In particular, the inventive SCS-OCV provides at least these three innovations: First, the inventive SCS-OCV introduces complex variation concept to resolve the blocking technical issue of combining local random variations. This enables the accurate calculation of statistical variations with correlations, such as common-path pessimism removal (CPPR).
Secondly, the inventive SCS-OCV proposes practical statistical min/max operations for random variations that can guarantee pessimism at nominal and targeted N-sigma corner, and extended the method to handle complex variations. This enables the graph-based full arrival/required time propagation under variable compaction.
Thirdly, the inventive SCS-OCV provides a statistical corner evaluation method for complex random variables that can transform vector-based parametric timing information to the single-value corner-based timing report, and based on the method, derives equations to bridge POCV/SSTA with LOCV. This significantly reduces the learning curve and increases the usage of the technology, being more easily adopted by the industry.
The invention provides at least these three novel and innovative techniques, critical to providing an efficient and user-friendly SCS-OCV solution.
The following drawings are provided to aid in understanding the invention:
A method according to the invention is illustrated in
In static timing analysis, VLSI designs are described as a netlist that contains different kind of cells and their interconnection. Current approaches to statistical timing methodology include those discussed in Hongliang Chang, Sachin S. Sapatnekar: Statistical Timing Analysis Considering Spatial Correlations using a Single Pert-Like Traversal, ICCAD 2003: 621-626; and Yaping Zhan, Andrzej J. Strojwas, Xin Li, Lawrence T. Pileggi, David Newmark, Mahesh Sharma: “Correlation-aware Statistical Timing Analysis with Non-Gaussian Delay Distributions,” DAC 2005: 77-82.
As do the above cited current statistical static timing methodologies, the inventive SCS-OCV approach set forth herein, models local random variations as independent statistical random variables for each cell. And the individual cell delay can be expressed as a linear function of that specific variable in a parametric form.
For example, if we know the cell nominal delay is “a”, and the timing impact of local random variation has a standard deviation of “b”, then we can express the cell delay “d” with the following equation. In the equation, p is a standard normal random variable.
d=a+b·p (1)
During timing propagation, we can propagate these parametric delay expressions across the timing graph. Equation (2) calculates the arrival time via a path as the sum of all the cell delays d1, d2, d3, . . . , dn, represented as
where i represents a running variable that takes values from 1 to n.
arrivaltime=Σi=1ndi=Σi=1n(ai+bi·pi) (2)
Note that because individual delay variations are treated as statistical independent random variables, we cannot directly group different bi·pi by adding their coefficients. However, keeping all the individual bi·pi will significantly increase the computation cost and memory requirements, which is not feasible. A common technique is to group the multiple independent variables into single variable as in equation (3). In equation (3), the term
represents the sum of values a1, a2, a3, . . . , an, where i is a running variable that takes values from 1 to n and the term
represents the sum of values b12, b22, b32, . . . , bn2, where i is a running variable that takes values from 1 to n.
arrivaltime=Σi=1nai+√{square root over (Σi=1nbi2·p)} (3)
Compaction of multiple random variables into one variable resolves the memory problem. However, through compaction, one loses track of correlation, and correlation is particularly important when the circuit has common path on the clock tree.
arrival=d1+d2
require=d1+d3
Assume each cell has 100 pico second cell delay and 10% random delay variations, the arrival time and require time can then be expressed as
arrival=100+10p1+10p2
require=100+10p1+10p3
If we keep individual random variables, we can calculate the slack as required time minus arrival time
slack=10p3−10p2
slack3sigma=−30√{square root over (2)}
However, if we compact multiple random variables into one as described in Equation (3), the arrival time and require time we will end up having is
arrival=100+10√{square root over (2)}p
require=100+10√{square root over (2)}p
As you can see, we no longer know what part of the arrival and require time is from common cell. In other words, we lose the statistical correlation information. To guarantee pessimism, we have to assume that there is no common part of arrival time and require time.
slack=√{square root over (2·102+2·102)}p=20p
slack 3sigma=−60
It is clear that with compaction, there is a significant amount of pessimism introduced. Actually the same problem also exists in non-statistical STA when different OCV derating factors were applied to the same common clock cells in Min and Max mode. In STA, it is called common path pessimism. And people usually introduce some adjustment at the output of cell 1, and apply that during slack calculation. In full SSTA analysis that keeps delay sensitivities for different cells, the problem no long exists because the same cell variations will get canceled naturally. However, when we combine individual random variations, the problem comes back and we need some solution for that.
Using the same circuit in
d
1
=a
1
+b
1
p
1
d
2
=a
2
+b
2
p
2
d
3
=a
3
+b
3
p
3 (4)
If we do not compact independent random variables, we can calculate the slack of the circuit as
arr=(a1+a2)+b1p1+b2p2
req=(a1+a3)+b1p1+b3p3
slack=(a3−a2)+b3p3−b2p2 (5)
And if we do the compaction on the fly, we will only have one random variable for arrival time and require time, and when we calculate the slacks, we have to do a root sum squares (RSS) to calculate the slack variation. To remove the pessimism, we would like to introduce some adjustment on the variation part as well.
arr=(a1+a2)+√{square root over ((b12+b22))}p
req=(a1+a3)+√{square root over ((b12+b32))}p
slack=(a3−a2)+√{square root over ((2b12+b22+b32))}p+adjustment (6)
We would like to see whether the slack distribution after the adjustment can match the slack distribution we calculated without compaction. Since the nominal part already matches between equation (5) and equation (6), we only need to match the sigma part.
(2b12+b22+b32)+adjustment2=b22+b32
adjustment2=−2b12p2
As we can see, the adjustment we would like to have has a negative variation. To achieve that, we extend the definition of variation into the complex domain so that we can have negative variance. Equation (7) represents a complex variable var as a sum of a real component α and an imaginary component β where i represents √{square root over (−1)}.
var=α+βi (7)
When we add the individual delays to calculate the arrival time, or subtract the arrival time from require time to calculate the slack, we have to add and subtract these complex variables. Since we always assume the compacted variables are independent variables, addition and subtraction has same impact on variation part. The equations listed below gives the compaction rules while adding different complex variables.
Because we always assume the compacted variables are independent variables, addition and subtraction follow the same set of equations. Equation (8) represents addition of two individual delays represented as complex terms αi+βii and αj+βji.
From equation set (8) we can see that the compaction in addition of real variables always leads to real variables, and the compaction of addition of imaginary variables always leads to imaginary variables. Compacting a real variable with an imaginary variable during addition can lead to either a real variable or an imaginary variable.
Going back to our simple example, using the proposed complex variable concept, we can easily express the adjustment as an imaginary number as follows.
adjustment=(√{square root over (2)}bii)p (9)
The adjustment only contains b1, and is not a function of b2 and b3. Therefore, it can be pre-calculated without having the knowledge of complete paths. This makes it possible to combine the technique with the traditional common path pessimism removal for non-statistical static timing analysis.
SCS-OCV can be applied on path-based timing analysis as well as graph-based full-chip timing analysis. In graph-based timing analysis, when arrival times from two paths converge, the bigger one will be selected for further propagation (in max mode). When arrival time becomes a statistical distribution, the calculation of max of two arrival times now becomes more complicated.
A significant amount of research has been done on statistical max operations, as can be appreciated by referring to the references cited in paragraphs [05] through [07]. Proposed methods include preserving the statistical mean and standard deviation of max of two statistical arrival times, or using quadratic model for statistical max operations. Such max operations work well if Users are interested in predicting the moments such as mean and standard deviation of the final slack distributions. However, static timing analysis needs to make sure that the VLSI circuit works at specified corners. Therefore, guaranteed guard-bound at targeted corner point (e.g. 3 sigma corner or 6 sigma corner) is needed. Current statistical max operations do not have this characteristic. Another issue with the existing statistical max operation is that they do not obey Associative and Distributive rules due to the approximation errors in the operation.
Max(Max(A, B), C)≠Max(A, Max(B, C))
Max(Max(A, C), Max(B, C))≠Max(A, Max(B, C))
Because of this, statistical static timing analysis of the following two circuits may generate different timing results. As depicted in
In SCS-OCV, we derived our statistical max operation so that it preserves both nominal and N sigma corner delay values. The number N is a user specified value based on their confidence requirement.
Define dA and dB as two parametric delays:
d
A
=a
A
+b
A
p
d
B
=a
B
+b
B
p
Our statistical MAX of dA and dB can be calculated uses the following equation, where the non-capital max operation is the regular max operation on real numbers.
We can easily prove that the above statistical max operations preserve the pessimism at N sigma. In other words, dY is greater than dA and dB at N sigma.
Furthermore, it follows the Associative and Distributive rules.
Using the transform of min(x,y)=−(max(−x,−y), we can derive the statistical Min operation in similar manner
Now we have a MAX operation that can preserve worst corner and follow Associative and Distributive rules. Next we would like to extend the MAX operation to handle complex random variations with both real and imaginary components. Before we derive complex statistical MIN/MAX operation, it is useful to first check where such operations are needed. It turns out that the most common place for complex statistical MIN/MAX operation is the required time propagation on the data paths.
Consider the circuit in
Please note that the physical required time and CPPR (common path pessimism removal) adjustment are grouped together. Consequently, there is an imaginary variation part in required time.
arr=40+2p
slack=√{square root over (2)}p
req=40+(√{square root over (2)})p
To propagate required time back from Z pin to Y pin, we need to calculate the required time from Z to A and from Z to B and take the minimum.
reqz→A=40+(√{square root over (2)}i)p−(10+p)=20+ip
reqz→B=40+(√{square root over (2)}i)p−10=30+(√{square root over (2)}i)p
reqY=min(30+ip,30+(√{square root over (2)}i)p)
For this simple circuit in
(a+bp)N
(a+(bi)p)N
Given the corner definition in equation (12), we can derive the generalized statistical Max operations for complex statistical variables as in equation (13).
Statistical Min operations have similar form.
Given the formula above, we can easily finish the calculation for the example in paragraph 50
Graph-based Arrival Time and Required Time propagations are two main areas that statistical MIN/MAX operation can be applied. There are also many other places that require MIN/MAX operation to model variation correctly. One such place is Latch time-borrowing calculation.
Unlike edge-triggered flip-flop, Latch has a transparent period in which it is always on. In other words, signal can always propagate through when latch is on. This creates challenges for static timing analysis which relies on a clear definition of clock period and the isolation between different combination blocks. Please refer to
In static timing analysis, a technique called time borrowing is widely used to model the impact of latch's on-period. The basic idea is illustrated in
Basically, when the signal arrival is after the latch is open, but before the latch closes, the difference between the data arrival time and latch open time can be treated as the time this combinational block borrowed from next combinational block. If the signal arrival is after latch close time, then the whole open period is treated as borrowed time. In static timing analysis, the borrowed time is calculated using the following piece-wise functions:
When we have variations, however, expressing borrowed time as a simple piecewise linear function is no longer enough. Due to variation, there is no clear boundary at latch open and latch close points. As shown in
Instead of using a piecewise linear function to calculate the borrowed time, we need a better formula that can generate continuous model near the latch open and closing edges.
To help deriving the proper formula, we can plot the borrowed time as a function of arrival time in
borrowed time=Min(latch on period, Max(data arrival−latch open,0))
When there are no variations, the equation above will generate exact same plot as
SCS-OCV and SSTA calculate and propagate the statistical distribution of delays, arrival/required times, and slacks. In the end, the user needs to check their timing at their confidence level (e.g. three sigma). If they find a timing violation at an end point, they need to trace back the path, see how they can fix the violation.
In such cases, we not only need to calculate the N sigma corner of the slack, but also need to show that under which condition the N sigma corner will be reached. For example, if we know the end point slack is a function of multiple random variables, we can calculate the N sigma corner value of the slack.
However, we also need to calculate the specific value of individual variables under which the N sigma corner slack will be reached. If we link the above two equations into one, the problem is to find a set of Pi that satisfies the equation.
Since there are multiple random variables but only one equation, the above equation has infinite solutions. In SCS-OCV, we look for the solution that has highest probability as our statistical corner solution.
In equation (15), (p1, p2, pi . . . ) is a set of independent standard random variables. Therefore, if we plot the joint probability density function in the n-dimensional space spanned by (p1, p2, pi . . . ), it is easy to prove that the joint probability density function monotonically decreases when the distance between the evaluation point and the original point r increases.
On the other hand, equation (15) is actually a hyper-plane function in the same n-dimensional space. Therefore, the point we are looking for should be the point on the hyper-plane that is closest to the original point. The problem of finding the corner vector for our timing report is translated into the problem of finding the projection point of the original point to a given hyper-plane. The latter has a well defined solution:
Notice that the denominator of the equation (16) is actually the standard deviation of the slack. Therefore, although equation (16) is derived from un-compacted POCV/SSTA, the solution can be directly applied to compacted SCS-OCV and SSTA. And if bi is a complex number, the equation still holds
Equation (16) can also be used to derive the equivalent stage count and derate functions for LOCV. LOCV models derating as a function of stage count on a logic path.
Equation (18) is widely used to calculate the LOCV derating value for a cell on a path with K stages. From equation (16), we can derive stage count and derating functions (19) and (20) that makes path-based LOCV match POCV/SSTA results.
As can be appreciated from the description herein, the invention provides an efficient solution that is implementable in Parametric On-Chip Variation/SSTA, and useful in the most complicated design timing challenges.
This application is a continuation of U.S. patent application Ser. No. 13/784,701, filed Mar. 4, 2013, which is a continuation of U.S. patent application Ser. No. 13/199,222, filed Aug. 23, 2011, which claims the benefit of, and priority to U.S. provisional application 61/402,198 filed 25 Aug. 2010, the contents of all are incorporated by reference in their entirety.
Number | Date | Country | |
---|---|---|---|
61402198 | Aug 2010 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 13784701 | Mar 2013 | US |
Child | 13969297 | US | |
Parent | 13199222 | Aug 2011 | US |
Child | 13784701 | US |