The invention relates to a system and methods for monitoring a set of metrics. More particularly, the invention provides a system and methods for computing and displaying data distribution information for metrics.
Transactions are at the heart of e-business. Without fast, efficient transactions, orders dwindle and profits diminish. Today's e-business technology, for example, is providing businesses of all types with the ability to redefine transactions. There is a need, though, to optimize transaction performance and this requires the monitoring, careful analysis and management of transactions and other system performance metrics that may affect e-business.
Due to the complexity of modern e-business systems, it may be necessary to monitor thousands of performance metrics, ranging from relatively high-level metrics, such as transaction response time, throughput and availability, to low-level metrics, such as the amount of physical memory in use on each computer on a network, the amount of disk space available, or the number of threads executing on each processor on each computer. Metrics relating to the operation of database systems, operating systems, physical hardware, network performance, etc., all must be monitored, across networks that may include many computers, each executing numerous processes, so that problems can be detected and corrected when (or preferably before) they arise.
Due to the complexity of the problem and the number of metrics involved, it is useful to be able to quickly view information relating to one or more metrics across a period of time. In particular, viewing information on the frequency distribution of data may be useful. Such data distribution information generally may be viewed as a histogram, and many systems that perform statistical analysis are able to display histograms.
Unfortunately, there are a number of drawbacks to using conventional histograms to view time series data, such as the complex metrics discussed above. First, there is often a need to display multiple histograms on a screen when working with time series data, such as metrics. Each histogram typically requires a large amount of screen space, limiting the number of histograms that can be displayed at the same time.
Additionally, histograms are not always useful for discovering trends in data over time, since it may be difficult to see long-term trends in data by viewing a set of standard histograms side-by-side or stacked vertically. Some systems attempt to solve this problem by making the histograms small, and turning them on their side. Unfortunately, even when these steps are taken, it may be difficult to display more than five or six histograms on a single display. It would be impractical in such systems to display fifty or a hundred such histograms in a single display.
Another difficulty with using histograms to view data distribution information over a long period of time is the storage of histogram data. Typically, a large number of individual data samples are needed to construct a histogram. To display histograms for data over a long time period, there are two options available. First, a system can pre-compute the histogram for each required time interval, and save the histogram data. This approach requires many computations and storage of data that may never be used. Alternatively, a system can save all the individual data points over a long time period, so that histograms can be computed as they are needed. This approach requires a large amount of storage, and may require a large memory footprint and a large amount of computation when the histograms are generated. As a result, this approach may not be practical for long periods of time and large numbers of metrics.
Additionally, histograms are somewhat inflexible. For example, they cannot be effectively averaged or merged to condense the display of several time intervals into a single interval. Similarly, they cannot be effectively averaged or merged from multiple copies of the same metric collected from distinct but similar systems. Such data distribution information may be useful for viewing the health and status of an entire system using only a few displays or screens.
In view of the foregoing, there is a need for a system and methods for computing and displaying data distribution information for large sets of time-series data, such as metrics. Further, there is a need for a system and methods for efficiently storing, and merging such data distribution information.
In one aspect, the present invention provides a method for generating an approximate histogram of a data set. This is done by applying a quantile function on the data set to create a computational result, selecting a subset of the data set in response to the computational result, determining a condensed quantile function from the subset of the data set, and rendering the approximate histogram in response to the condensed quantile function.
In some embodiments, the subset of the data set includes at least one of a minimum value of the data set, a median value of the data set, and a maximum value of the data set.
In some embodiments, determining the condensed quantile function involves interpolating between values in the data set. In some such embodiments, the interpolation includes at least one of linear interpolation and polynomial interpolation.
In some embodiments, rendering the approximate histogram involves calculating an expected sample count based at least in part on an inverse of the condensed quantile function.
In some embodiments, the approximate histogram includes a plurality of bins. In these embodiments, the rendering step includes reducing the plurality of bins in the approximate histogram into a plurality of adjacent cells, and applying an indicium to each cell in response to the percentage of the data set within each cell. In some such embodiments, the indicium is visually discernable (e.g., a shade, texture, or color of the cell).
In another aspect, the invention provides a method of merging a plurality of data sets with reduced data storage requirements. This is accomplished by calculating a condensed quantile function for each data set, with each condensed quantile function supplying quantile values. Next, the quantile values are interleaved, and an inverse of each condensed quantile function is calculated at each interleaved quantile value. The average of the inverse of the condensed quantile functions at each interleaved quantile value are calculated, and a merged quantile function is defined as an inverse of the average of the inverse of the condensed quantile functions at each interleaved quantile value.
In some embodiments the plurality of data sets include data associated with a single variable from adjacent time intervals. In other embodiments, the plurality of data sets include data associated with a plurality of instances of a single variable from a single time interval.
Some embodiments further include a step of rendering a merged histogram in response to the merged quantile function. In some of these embodiments, rendering the merged histogram involves calculating an expected sample count based at least in part on an inverse of the merged quantile function. In some embodiments, the step of rendering the merged histogram includes reducing a plurality of bins in the merged histogram into a plurality of adjacent cells, and applying an indicium to each cell in response to a percentage of the data set within each cell. In some such embodiments, the indicium is visually discernable.
In a further aspect, the invention provides a method of optimizing performance in a distributed transaction system. This is accomplished by collecting data associated with at least one system performance metric, and applying a quantile function on the data to create a computational result. Next, a subset of the data is selected in response to the computational result, and a condensed quantile function is determined from the subset of the data. The method also includes rendering at least one approximate histogram in response to the condensed quantile function, identifying at least one trend in the approximate histogram, and adjusting, on an as-needed basis, operation of the distributed transaction system to modify the trend.
In some embodiments, the methods of the invention can be implemented in software, or in logic or other hardware. When the methods are implemented in software, the software may be made available to developers and end users online and through download vehicles. It may also be embodied in an article of manufacture that includes a program storage medium such as a computer disk or diskette, a CD, DVD, or computer memory device.
Other aspects, embodiments, and advantages of the present invention will become apparent from the following detailed description which, taken in conjunction with the accompanying drawings, illustrating the principles of the invention by way of example only.
The foregoing and other objects, features, and advantages of the present invention, as well as the invention itself, will be more fully understood from the following description of various embodiments, when read together with the accompanying drawings, in which:
In the drawings, like reference characters generally refer to the same parts throughout the different views. The drawings are not necessarily to scale, emphasis instead being placed on illustrating the principles of the invention.
As shown in the drawings for the purposes of illustration, the invention may be embodied in a system that collects, analyzes, and reports performance metrics for systems such as, for example, complex transaction-based structures typified by (but not limited to) e-commerce systems. In broad overview, the invention relates to the monitoring of a set of system performance metrics, such as transaction response time, throughput, availability, etc. The system receives metric data and generates and displays a set of “approximate histograms” from the quantile function of a time series segment. An approximate histogram is an estimation of the histogram of an original (i.e., complete) data set.
In addition to generating “approximate histograms,” an embodiment of the invention also displays a set of histograms for successive time intervals using a set of offset stacked bar charts. The resulting display provides a useful visual tool for tracking changes in the central values of a metric and for detecting subtle changes in its frequency distribution over long time scales. Due to the use of approximate histograms, this is accomplished without the need to store a large number of individual samples over long periods of time.
Based on the data monitored and displayed, an end user is able to identify trends and other indicia (e.g., threshold alarm notification) of system performance. The information displayed (e.g., response time, throughput, and availability) can indicate that certain transactions are not occurring as expected. Consequently, an end user is able to act to determine and remedy the root cause of transaction problems before they become critical.
The system and methods of the present invention are described herein as applying to software for use by a system manager, such as an e-business system manager, to assist, for example, in the achievement and maintenance of Service Level Agreements in terms of system performance. It will be understood that the system and methods of the present invention are not limited to this application, and can be applied to the storage and display of histogram data in most any system whose operation can be described through use of a set of system metrics.
In
Approximate histogram generation module 204 uses the methods detailed below to generate approximate histograms for the data. Using these techniques, it is possible to store histogram data over a long period of time without requiring large amounts of storage space. As will be described in detail below, this is done by generating and storing condensed quantile tables, from which approximate histograms for the data may be generated. Approximate histogram generation module 204 is able to aggregate histogram data by merging the condensed quantile tables.
Histogram display module 206 displays histogram data as “stack-bar” histograms, which will be described in detail below. Use of these stack-bar histograms permits multiple histograms to be displayed at once without requiring a large amount of screen space. Additionally scaled stack-bar histograms can be displayed for multiple time intervals, permitting an operator to view trends in the histogram data associated with a metric.
In operation, data collection module 202 collects N Data points for a metric over each time interval T. For example, N may be 900 points over a T=15 minute interval.
Once the data is collected, it is processed by an approximate histogram generation module, as described below.
Generating an Approximate Histogram
This section presents a method for generating an approximate histogram from the quantile function of a condensed data set. This is useful as a data compression technique because a small number of data points from the quantile function can retain much information about the distribution of a data set. As will be described, a histogram can be re-constructed from the condensed quantile points. Such a re-constructed histogram is referred to as an approximate histogram.
In general, four basic functions are involved in generating the approximate histogram: the histogram, the probability density function, the cumulative distribution function and the quantile function. The relationship between these functions provides a basis for the method, as described below.
Typically, all of the data points are needed to regenerate the histogram 400. Thus, unless histograms are pre-computed and saved, a large amount of data must be stored to produce a histogram on demand.
Referring now to
f(v)=0.000058556v9−0.00353074v8+0.14075v7−3.89408v6+76.2797v5−1057.71v4+10172.4v3−64611.8v2+243906.0v−414515
Note that an estimated histogram of the underlying data can be generated from the CDF. The probability that a value will fall between a lower bin limit, vl, and an upper bin limit, vh, may be computed using the CDF, F(v), as follows:
P(vl<v≦vh)=F(vh)−F(vl) (Eq. 2)
Thus, the expected value of the number of samples that will fall in a bin, j, is:
Nj=N*(F(vhj)−F(vlj)) (Eq. 3)
v=Q(p)v such that P(vk≦v)=p.
Q(p)=F−1(p) (Eq. 4)
A condensed quantile function may be generated by selecting a subset of the data points from the quantile function. For example, 13 points may be selected from the complete data set, including the minimum, median and maximum values. For the quantile function 800 of
As discussed above, an estimated histogram may be generated from a CDF, which is the inverse of the quantile function. Thus, an approximate histogram may be generated based on the inverse of the condensed estimated quantile function.
Note that the histogram 1100 in
Use of a condensed estimated quantile function, as discussed above, can provide numerous benefits. First, because far fewer points are needed to generate histograms from condensed estimated quantile functions than from the original data, the amount of memory and the number of calculations that are needed to produce histograms are greatly reduced. Because of the reduced number of calculations that are needed, it is possible to quickly compute numerous histograms, representing large amounts of data. Additionally, use of condensed estimated quantile functions can provide substantial savings of storage space. For example, to store 24 hours of raw, one-second samples would require 24 hr.×3600 sec./hr., or 86,400 storage locations. In contrast, the storage of 96 condensed quantile functions of thirteen points each, providing one such quantile function for each fifteen minute interval, would requires only 96×13=1248 storage locations.
The functions outlined in the above example provide the basis for generating an approximate histogram. As will be shown below, in an embodiment of the invention, the process is implemented in an efficient manner, in which a quantile function is directly estimated from the data.
Referring to
At step 1202, the system generates an estimated quantile function from the samples of a time interval. In general, the estimated quantile function associates a value (v), with the portion (p) of all values that are less than v. The portion (p) is an estimate of the probability that a randomly chosen value (vk) will be less than the value (v). The quantile function takes the probability or portion (p) as input and returns the value (v).
v=Q(P)v such that P(vk≦v)=p.
Given a set of N values (i.e., N samples from a time interval), the estimated quantile function may be constructed, for example, by sorting the set of values (vi) into ascending order. This may be done incrementally as the data is collected using, for example, an insertion sort. If the data is already available it may be sorted efficiently with any number of standard techniques including Quicksort or with a merge sort, depending on the sorting performance and stability sought.
Next, for each sorted value vi (where i is in the range 0 to N−1), a pseudo-probability pi is assigned as follows:
pi=(i+0.5)/N (Eq. 5)
This process will produce a set of N (pi, vi) pairs, where the pi values are evenly spaced. These (pi, vi) pairs form an estimated quantile function for the data.
In step 1204, to reduce storage requirements, in accordance with and embodiment of the invention, only some of the sample quantile function points are stored as a condensed quantile table. To maintain the maximum amount of information, these points, or sub-samples could be more closely spaced where the quantile function has high curvature, and farther apart where there is less curvature. An algorithm to select such sub-samples will be described in detail below, with reference to
A strategy for selecting sub-samples that will work for many applications is to evenly space the sub-samples. This also eliminates the need to store the probabilities, since they are implicit (i.e., they are evenly spaced).
Given N samples, M evenly spaced sub-samples, psk may be chosen. For suitable values of N and M, the indices of the sub-samples may be chosen as follows:
To ensure that the median is captured when M is odd, the value of the middle quantile point may be set to Q(0.5). When M is odd, the middle index is (M−1)/2.
In general, the sub-sample probabilities (psk) will not correspond to the original sample probabilities (pi). For example if ps equals 0.5 and there are an even number of samples N, then Q(0.5) will be the average of sample vN/2−1 and vN/2. So Q(ps) must be calculated by interpolation for any (ps) that falls between two points in the estimated quantile function table.
As an example, let N=900 and M=11, then the probabilities and corresponding quantile points that make up the condensed quantile table would be as follows:
The value of the quantile function Q(ps) can be calculated from the estimated quantile table (i.e., the (pi, vi) pairs) by interpolation. First, the system finds the index in the table that has the probability value closest to (ps). This can be done by inverting the pseudo probability function. Given that
pi=(i+0.5)/N
The indices (i0,i1) that correspond to the interval containing (ps) can be found as follows:
i0=IntegerPart(ps*N−0.5) (Eq. 6)
i1=i0+1 (Eq. 7)
The value of Q(ps) is then computed using known linear interpolation techniques to interpolate a result between vi0 and vi1. If required, polynomial interpolation could be used for higher precision.
If N is large, the difference between p0, ps and p1 will be small. For example, if N=1000, the difference between p0 and p1 will be 0.001, and ps will be within 0.0005 of either p0 or p1. This leads to an even more accurate estimate of Q(ps) when N is large.
Thus, when the number of sorted samples (N) is large, the condensed quantile table can be constructed with little loss of accuracy by choosing samples of vi corresponding to the required probabilities pi. Given that
pi=(i+0.5)/N
the index of the value closest to Q(ps) is:
i=Round(ps*N−0.5)
Thus, when N is large, Q(ps )≈vi for all required values of ps. The probability error (ep) introduced by this method will be less than or equal to one-half of the difference between successive probabilities, i.e.:
ep<=0.5/N.
The corresponding quantile value error introduced is less than or equal to one-half of the maximum difference between two successive values, vi, vi+1.
If the values of M and N are such that the fractional part of [pi*N−0.5] is close to 0.5, then an improvement in accuracy may be gained by averaging adjacent samples. For example, if N=900 and M=11, then the probabilities and corresponding quantile points that make up the condensed quantile table would be as follows (indices are given by equations 6 and 7):
The estimated quantile function, condensed or not condensed, represents a set of samples that are an estimate of the inverse of the Cumulative Probability Distribution Function F(v) of the underlying data set. In step 1206, as discussed above, the approximate histogram is computed. This is done by finding the inverse of the condensed estimated quantile function, which is represented by the condensed quantile table, and using the inverse (which is an approximate CDF) to produce the approximate histogram.
To estimate the number of samples that would fall in a histogram bin, the inverse of the quantile function is used, evaluated at both bin limits. (The inverse of the quantile function is the probability that a randomly chosen value Vk will be less than V.)
F(v)=Q−1(v)=P(vk≦v).
The probability that a value will fall between vl and vh is computed as follows:
P(vl<v≦vh)=Q−1(vh)−Q−1(vl)
The expected value of the number of samples to fall in the same range is:
Nj=N*(Q−1(vhj)−Q−1(vlj))
The histogram can be formed as the set of Nj computed using this formula. The set bin limits (vl, vh) are established at equal intervals over a reasonable range of values. For example, these could be from the lowest quantile point to the highest, or a certain number of inter-quantile ranges (IQR's) around the median. The resulting set of Nj will approximate the original histogram of the data set.
The inverse of the quantile function can be computed efficiently with a binary search method. This is because the quantile function is monotonic. Given a table of quantile function values (psj, vsj), and a value (v) it is possible to find Q−1(v) using the following procedure (provided as pseudocode):
Having computed an estimate for the CDF (i.e., Q−1(v)), the methods described above are used to determine the expected number of samples in each bin of the approximate histogram.
Generally, the adaptive fit algorithm 1300 takes an estimated quantile table (Q), which includes a set of pairs (pi, vi), where i is from 1 to N. A goal is to generate a condensed quantile table, including a set of K pairs (psj, vsj), where j is from 1 to K, and K is much less than N. Preferably, the condensed quantile function will closely approximate Q, so that the condensed quantile table retains most of the information in the larger estimated quantile table.
In step 1302, a line is created between the first point in the quantile table and the last point, using the p and v values as x and y coordinates, respectively.
Next, in step 1304, the system computes the perpendicular distance between each quantile point and the line that was created in step 1302 (i.e., the distance between the line and the point along a linear path that runs perpendicular to the line). The line and the point that has the maximum perpendicular distance from the line are recorded in a segment list. Additionally, the “current” segment is set to this line.
In step 1306, the system creates two line segments from the “current” segment. The first of these line segments extends from the first point of the “current” segment to P, the point along the “current” segment having the maximum perpendicular distance from the “current” segment. The second segment extends from P to the last point of the “current” segment.
In step 1308, for each of the two line segments created in step 1306, the system computes the perpendicular distance between each point on that segment and the line between the first point of the segment and the last point of the segment. The system records the location of the point that has the maximum distance from the line segment.
In step 1310, the system removes the “current” segment from the segment list, and inserts in its place the two new segments, each of which consists of a line, and a most distant point.
Next, in step 1312, the system finds the segment in the segment list that has the largest maximum point distance, and makes this segment the “current” segment. The system then repeats steps 1306–1312 until the segment list has K−1 segments (step 1314).
Finally, the system generates a K point condensed quantile table by taking from the segment list the first point of the first segment, followed by the last points of the K−1 segments in the segment list, in order. As discussed above, this condensed quantile table can be used to generate an approximate histogram.
The approximate histograms that are generated by the above-described methods may be used for many of the same purposes for which a normal histogram may be used. In addition, because the amount of information needed to generate an approximate histogram is considerably less than what is needed to generate a regular histogram, and approximate histograms may be quickly generated from condensed quantile tables, use of approximate histograms may be used in instances where use of a regular histogram would be impractical.
Stack-Bar Histogram Charts
In accordance with an illustrative embodiment of the invention, approximate histograms may be used to create stack-bar histograms, which permit numerous histograms to be displayed on a single screen. The ability to display numerous histograms on a single screen permits a viewer to see trends in the data that would not have been readily ascertainable without such a display.
Stack bar histogram 1406 represents a data distribution using shades, textures, or colors, rather than using the height of bars to show the number of items in each bin. Advantageously, since such stack bar histograms are more compact than a regular representation of a histogram, numerous such stack bar histograms may be arranged in a single display.
Displaying numerous stack-bar histograms side-by-side, as shown in display 1502, makes it easy to track shifts in the mean, and in the shape of the distribution. For example, in display 1502, it is easy to see that the mean of the metric tended upward at first, and then leveled off. It can also be seen that the distribution varied from one where most of the values were evenly distributed about the median to others where most of the values concentrated near the lower or higher end of the range.
Display 1702 also includes statistics area 1708, in which the minimum, maximum, median, and 5th and 95th percentiles for a currently selected interval are shown. In Limits area 1710, upper and lower dynamic limits, such as those described in commonly owned, co-pending U.S. patent application Ser. No. 10/198,689, filed on Jul. 18, 2002, are shown. Histogram area 1712 displays a regular histogram for the currently selected area, and permits a user to view information including the range and count in each bin of the histogram by placing the cursor over the bars that represent the bins of the histogram.
It will be understood that display 1702 is an example of one display screen that uses stack-bar histograms, and that many variations are possible. Generally, such stack-bar histograms may be used in any system in which it is desirable to simultaneously display data distribution information over numerous related sets of data.
Merging Sets of Quantile Functions
When building displays based on stack-bar histograms, as shown in the foregoing figures, it may be useful to be able to aggregate multiple quantile functions, either to combine multiple time intervals, or to represent an aggregation of multiple instances of the same variable at the same time interval. Since direct quantile averaging methods generally do not preserve the overall range and distribution information from the individual quantile functions, it may be desirable to use a merging technique for quantile functions, as described hereinbelow.
For example, given two condensed quantile function tables (Qa, Qb) representing the quantile functions for a single metric for two adjacent time intervals, a single merged quantile function may be produced. When two quantile functions are merged, the resulting quantile function is approximately the same as that which would have been created if the two original data sets were merged and the quantile function of the resulting data set was computed. When this is done, the total range and distribution of the two data sets is preserved.
The merge is accomplished by interleaving of the quantile values of the quantile functions to be merged, and then computing the average of Qa−1 and Qb−1 at each interleaved value. This has the effect of determining at each value (vi), what percentage of the values from the data associated with Qa would be less than vi, and what percentage of those associated with Qb would be less than vi. Given that an equal number of total samples are drawn from each data set, the average percentage Qm−1(vi) is the percentage of the total that would be less than the quantile value vi at the given point. This may be expressed as:
Qm−1(vi)=(Qa−1(vi)+Qb−1(vi))/2
The algorithm operates on a set of K quantile tables, and produces a merged quantile table.
At step 2102, an array of K quantile table indices (indexArray) are created, and each index is set to 1.
Next, at step 2104, the smallest minimum (setMin) and the largest maximum (setMax) are found over all the quantile tables that are being merged.
In step 2106, the maximum number of points for the merged quantile table is set to a “reasonable” value, M. Typical values of M may be the sum of the number of points in each quantile table, the average number of points in each quantile table, or a value between these two. In step 2108, the system creates an empty array of M values to hold the result of merging the quantile tables (mergedQuantile).
In step 2110, the system sets a variable (deltaV) representing the change in value for each step to the difference between setMax and setMin, divided by the number of merged quantile points, M. In step 2112, the system sets the current value (currValue) to setMin plus delta V. Next, in step 2114, the system sets the value of a variable, probSum, to zero, to begin accumulation for an average.
In step 2116, for each quantile table (i) in the set of quantile tables being merged, the system increments indexArray[j] until currValue is between the indexArray[j] and indexArray[j]+1 entries in the quantile table. The system then linearly interpolates on this interval to estimate the probability (p) associated with the value currValue, given the interval endpoints. The result, p, is an estimate of the inverse quantile function Q−1(currValue). The probability, p, is then accumulated into probSum.
In step 2118, the average cumulative probability, cp, is computed as probSum divided by K.
In step 2120, the system saves the pair {cp, currValue} as the next point in the merged quantile table, mergedQuantile.
In step 2122, deltaV is added to currValue, and steps 2114 through 2122 are repeated until currValue reaches setMax−deltaV (step 2124).
When the process is complete, in step 2126, the table mergedQuantile holds the resulting merged quantile table.
Note that because
This application claims priority to and the benefit of, and incorporates herein by reference, in its entirety, provisional U.S. patent application Ser. No. 60/322,021, filed Sep. 13, 2001. This application also incorporates herein by reference, in their entirety, U.S. patent application Ser. No. 10/186,401, filed Jul. 1, 2002, and Ser. No. 10/198,689, filed Jul. 18, 2002.
Number | Name | Date | Kind |
---|---|---|---|
4821220 | Duisberg | Apr 1989 | A |
5062055 | Chinnaswamy et al. | Oct 1991 | A |
5101348 | Arrowood et al. | Mar 1992 | A |
5161116 | Schneider et al. | Nov 1992 | A |
5307499 | Yin | Apr 1994 | A |
5343538 | Kasdan | Aug 1994 | A |
5367473 | Chu et al. | Nov 1994 | A |
5375070 | Hershey et al. | Dec 1994 | A |
5432932 | Chen et al. | Jul 1995 | A |
5450406 | Esaki et al. | Sep 1995 | A |
5459837 | Caccavale | Oct 1995 | A |
5459868 | Fong | Oct 1995 | A |
5475625 | Glaschick | Dec 1995 | A |
5475843 | Halviatti et al. | Dec 1995 | A |
5483468 | Chen et al. | Jan 1996 | A |
5491791 | Glowny et al. | Feb 1996 | A |
5504921 | Dev et al. | Apr 1996 | A |
5506955 | Chen et al. | Apr 1996 | A |
5513126 | Harkins et al. | Apr 1996 | A |
5553235 | Chen et al. | Sep 1996 | A |
5555191 | Hripcsak | Sep 1996 | A |
5572672 | Dewitt et al. | Nov 1996 | A |
5581482 | Wiedenman et al. | Dec 1996 | A |
5581696 | Kolawa et al. | Dec 1996 | A |
5586254 | Kondo et al. | Dec 1996 | A |
5598429 | Marshall | Jan 1997 | A |
5615135 | Waclawsky et al. | Mar 1997 | A |
5615323 | Engel et al. | Mar 1997 | A |
5616904 | Fernadez | Apr 1997 | A |
5619656 | Graf | Apr 1997 | A |
5621663 | Skagerling | Apr 1997 | A |
5634009 | Iddon et al. | May 1997 | A |
5636344 | Lewis | Jun 1997 | A |
5655081 | Bonnell et al. | Aug 1997 | A |
5661668 | Yemini et al. | Aug 1997 | A |
5668944 | Berry | Sep 1997 | A |
5675510 | Coffey et al. | Oct 1997 | A |
5684945 | Chen et al. | Nov 1997 | A |
5696486 | Poliquin et al. | Dec 1997 | A |
5696701 | Burgess et al. | Dec 1997 | A |
5699403 | Ronnen | Dec 1997 | A |
5732273 | Srivastava et al. | Mar 1998 | A |
5740441 | Yellin et al. | Apr 1998 | A |
5748098 | Grace | May 1998 | A |
5752062 | Gover et al. | May 1998 | A |
5758071 | Burgess et al. | May 1998 | A |
5778230 | Wimble et al. | Jul 1998 | A |
5796663 | Park et al. | Aug 1998 | A |
5799154 | Kuriyan | Aug 1998 | A |
5802256 | Heckerman et al. | Sep 1998 | A |
5815718 | Tock | Sep 1998 | A |
5819028 | Manghirmalani et al. | Oct 1998 | A |
5845125 | Nishimura et al. | Dec 1998 | A |
5850388 | Anderson et al. | Dec 1998 | A |
5870540 | Wang et al. | Feb 1999 | A |
5896536 | Lindsey | Apr 1999 | A |
5901315 | Edwards et al. | May 1999 | A |
5923834 | Thieret et al. | Jul 1999 | A |
5948113 | Johnson et al. | Sep 1999 | A |
5956662 | Hemker et al. | Sep 1999 | A |
5974237 | Shurmer et al. | Oct 1999 | A |
5974457 | Waclawsky et al. | Oct 1999 | A |
5978594 | Bonnell et al. | Nov 1999 | A |
5987249 | Grossman et al. | Nov 1999 | A |
5991881 | Conklin et al. | Nov 1999 | A |
6021437 | Chen et al. | Feb 2000 | A |
6026234 | Hanson et al. | Feb 2000 | A |
6026237 | Berry et al. | Feb 2000 | A |
6028842 | Chapman et al. | Feb 2000 | A |
6042614 | Davidson et al. | Mar 2000 | A |
6044335 | Ksendzov | Mar 2000 | A |
6058102 | Drysdale et al. | May 2000 | A |
6058393 | Meier et al. | May 2000 | A |
6061518 | Hoffman | May 2000 | A |
6067412 | Blake et al. | May 2000 | A |
6072777 | Bencheck et al. | Jun 2000 | A |
6073089 | Baker et al. | Jun 2000 | A |
6076083 | Baker | Jun 2000 | A |
6081797 | Hittt | Jun 2000 | A |
6085029 | Kolawa et al. | Jul 2000 | A |
6106572 | Halpern | Aug 2000 | A |
6108658 | Lindsay et al. | Aug 2000 | A |
6108800 | Asawa | Aug 2000 | A |
6115393 | Engel et al. | Sep 2000 | A |
6118936 | Lauer et al. | Sep 2000 | A |
6118940 | Alexander, III et al. | Sep 2000 | A |
6119074 | Sarangapani | Sep 2000 | A |
6119247 | House et al. | Sep 2000 | A |
6138122 | Smith et al. | Oct 2000 | A |
6139198 | Danforth et al. | Oct 2000 | A |
6145121 | Levy et al. | Nov 2000 | A |
6149318 | Chase et al. | Nov 2000 | A |
6151701 | Humphreys et al. | Nov 2000 | A |
6154736 | Chickering et al. | Nov 2000 | A |
6161200 | Rees et al. | Dec 2000 | A |
6167398 | Wyard et al. | Dec 2000 | A |
6182022 | Mayle et al. | Jan 2001 | B1 |
6182157 | Schlener et al. | Jan 2001 | B1 |
6216119 | Jannarone | Apr 2001 | B1 |
6222652 | Roberts | Apr 2001 | B1 |
6249755 | Yemini et al. | Jun 2001 | B1 |
6263298 | Kerman et al. | Jul 2001 | B1 |
6269401 | Fletcher et al. | Jul 2001 | B1 |
6320585 | Engel et al. | Nov 2001 | B1 |
6327550 | Vinberg et al. | Dec 2001 | B1 |
6327677 | Garg et al. | Dec 2001 | B1 |
6343288 | Lindsay et al. | Jan 2002 | B1 |
6359976 | Kalyanpur et al. | Mar 2002 | B1 |
6377907 | Waclawski | Apr 2002 | B1 |
6381306 | Lawson et al. | Apr 2002 | B1 |
6405250 | Lin et al. | Jun 2002 | B1 |
6446058 | Brown | Sep 2002 | B1 |
6453346 | Garg et al. | Sep 2002 | B1 |
6466929 | Brown et al. | Oct 2002 | B1 |
6470464 | Bertram et al. | Oct 2002 | B2 |
6484129 | Klein et al. | Nov 2002 | B2 |
6505246 | Land et al. | Jan 2003 | B1 |
6513065 | Hafez et al. | Jan 2003 | B1 |
6553366 | Miller et al. | Apr 2003 | B1 |
6865295 | Trajkovic | Mar 2005 | B2 |
20010052087 | Garg et al. | Dec 2001 | A1 |
20020012011 | Roytman et al. | Jan 2002 | A1 |
20020019870 | Chirashnya et al. | Feb 2002 | A1 |
20020049838 | Sylor et al. | Apr 2002 | A1 |
20020054169 | Richardson | May 2002 | A1 |
20020090134 | Van Zon | Jul 2002 | A1 |
20020095661 | Angel et al. | Jul 2002 | A1 |
20020111755 | Valadarsky et al. | Aug 2002 | A1 |
20020123865 | Whitney et al. | Sep 2002 | A1 |
20020133757 | Bertram et al. | Sep 2002 | A1 |
20020152185 | Satish et al. | Oct 2002 | A1 |
20020152304 | Collazo | Oct 2002 | A1 |
20020152305 | Jackson et al. | Oct 2002 | A1 |
20020158918 | Feibush et al. | Oct 2002 | A1 |
20020159386 | Grosdidier et al. | Oct 2002 | A1 |
20020165892 | Grumann et al. | Nov 2002 | A1 |
20020169870 | Vossler et al. | Nov 2002 | A1 |
20020173997 | Menard et al. | Nov 2002 | A1 |
20020174174 | Ramraj et al. | Nov 2002 | A1 |
20020183972 | Enck et al. | Dec 2002 | A1 |
20020184065 | Menard et al. | Dec 2002 | A1 |
20020184615 | Sumner et al. | Dec 2002 | A1 |
20020198879 | Schwarcz | Dec 2002 | A1 |
20020198985 | Fraenkel et al. | Dec 2002 | A1 |
20030005362 | Miller et al. | Jan 2003 | A1 |
20030009507 | Shum | Jan 2003 | A1 |
20030014692 | James et al. | Jan 2003 | A1 |
20030018241 | Mannheimer | Jan 2003 | A1 |
20030018494 | Bronstein et al. | Jan 2003 | A1 |
20030023712 | Zhao et al. | Jan 2003 | A1 |
20030028631 | Rhodes | Feb 2003 | A1 |
20030033404 | Richardson | Feb 2003 | A1 |
Number | Date | Country |
---|---|---|
296 00 609 U 1 | Mar 1997 | DE |
0 508 386 | Oct 1992 | EP |
2 783 620 | Mar 2000 | FR |
WO 9612224 | Apr 1996 | WO |
WO 9913427 | Mar 1999 | WO |
Number | Date | Country | |
---|---|---|---|
20030088542 A1 | May 2003 | US |
Number | Date | Country | |
---|---|---|---|
60322021 | Sep 2001 | US |