A general goal of computer program verification is to discover invariants that are strong enough to verify given assertions in a program in order to prove program correctness and/or find bugs. Traditionally, iterative fixed-point computation based techniques like data-flow analyses, abstract interpretation or model checking have been used for discovering these invariants.
An alternative is to use a constraint-based invariant generation technique. Constraint-based techniques offer advantages over fixed-point computation based techniques, including that they are goal-directed and thus have the potential to be more efficient. Another advantage is that they do not require the use of widening heuristics that are used by fixed-point based techniques which lead to loss of precision that is often hard to control. However, current constraint-based techniques are limited in a number of ways.
This Summary is provided to introduce a selection of representative concepts in a simplified form that are further described below in the Detailed Description. This Summary is not intended to identify key features or essential features of the claimed subject matter, nor is it intended to be used in any way that would limit the scope of the claimed subject matter.
Briefly, various aspects of the subject matter described herein are directed towards a technology by which program analysis uses rich invariant templates that may specify an arbitrary Boolean combination of linear inequalities for program verification to determine constraints. Also described is choosing a cut-set that identifies program locations, each of which is associated with an invariant template.
The verification analysis generates second-order constraints, converts second-order logic formula based on those constraints into first-order logic formula, then converts the first-order logic formula into a quantifier-free formula, which is then converted into a Boolean satisfiability formula. Off-the-shelf constraint solvers may then be applied to the Boolean satisfiability formula to generate program analysis results. Various templates may be used to convert the second-order logic formula into the first-order logic formula.
The verification may perform interprocedural analysis, including by computing procedure summaries for relations between procedure inputs and outputs structured as sets of precondition and postcondition pairs. A weakest precondition may be generated using a binary search-based algorithm and/or using an algorithm that generates locally-weakest preconditions with respect to a given neighborhood structure. Analysis may determine a worst-case bound on the running time of a procedure, including by computing bounds on loop iterations or on a number of recursive procedure call invocations, or by computing bounds on both loop iterations and on a number of recursive procedure call invocations. Further, most-general counter-examples to termination may be determined, along with most-general counter-examples to safety assertions.
Other advantages may become apparent from the following detailed description when taken in conjunction with the drawings.
The present invention is illustrated by way of example and not limited in the accompanying figures in which like reference numerals indicate similar elements and in which:
Various aspects of the technology described herein are generally directed towards an program analysis approach that translates (the second-order constraints represented by) a program into (boolean) satisfiability constraints that can be solved using off-the-shelf boolean SAT solvers. In one aspect, there is described the choice of a cut-set that affects the precision and efficiency of the constraint processing algorithm. Further described are interprocedural analysis and the determination of weakest precondition and strongest postcondition examples.
It should be understood that the examples described herein are non-limiting examples. As such, the present invention is not limited to any particular embodiments, aspects, concepts, structures, functionalities or examples described herein. Rather, any of the embodiments, aspects, concepts, structures, functionalities or examples described herein are non-limiting, and the present invention may be used various ways that provide benefits and advantages in computing and program analysis in general.
Turning to
As described herein, the exemplified program analyzer 102 operates by generating second-order constraints 112. Conversion logic 114 uses various templates 116 as described herein to convert second-order logic formula based on those constraints into first-order logic formula followed by using Farkas lemma to convert that first-order logic formula into a quantifier-free formula followed by using bit-vector modeling to generate a SAT formula. Existing constraint solvers 118 (e.g., SAT solvers) may then be used (e.g., as selected by a results generator 120) to generate the results 106.
In one implementation, uniform constraint-based techniques are provided as a core framework, including for three classical program analysis problems, namely program verification, weakest precondition generation and strongest postcondition generation over the abstraction of linear arithmetic. Using this core framework of analyses other examples apply to bounds analysis and finding most-general counterexamples to safety and termination properties. One feature of the analyzer 102 is that it can uniformly handle a large variety of challenging examples that otherwise require many different specialized techniques for analysis.
Moreover, the constraint-based technique can generate linear arithmetic invariants with bounded Boolean structure, which also allows the approach to extend to a context-sensitive interprocedural setting. In one aspect, the scheme for reducing second-order constraints to SAT constraints may be used in solving a special class of second-order formulas.
Another aspect is directed towards an appropriate choice of cut-set. The analyzer 102 can verify assertions (safety properties) in benchmark programs (e.g., used by alternative state-of-the-art techniques) that require disjunctive invariants and sophisticated procedure summaries. Constraint-based invariant generation can be applied to verifying termination properties as well as to the difficult problem of bounds analysis.
A goal of weakest precondition generation is to infer the weakest precondition that ensures validity of all assertions in a given program. Described is a constraint-based technique for discovering some form of weakest preconditions. The analyzer 102 can generate weakest preconditions of safety as well as termination properties for a wide variety of programs that cannot be analyzed uniformly by any other technique.
Also described is an interesting application of weakest precondition generation, namely generating most-general counterexamples for both safety and termination properties. By generating most-general counterexamples (as opposed to generating any counterexample), counterexamples are characterized in a succinct specification that provides better intuition to the programmer. For example, if a program has a bug when (n≧200+y) Λ (9>y>0), then this information is more useful than simply generating any particular counterexample, say n=356 Λ y=7 (as described below with reference to
A goal of strongest postcondition generation is to infer precise invariants in a given program so as to precisely characterize the set of reachable states of the program. Note that current constraint-based invariant generation techniques work well only when the program enforces the constraint that the invariant be strong enough to verify the assertions. However, in absence of assertions in programs, there is no guarantee about the precision of invariants. Described is a constraint-based technique that can be used to discover some form of strongest invariants.
In the area of fixed-point computation based techniques, the problem of generating precise invariants has led to development of several widening heuristics that are tailored to specific classes of programs. The analyzer 102 can uniformly discover precise invariants for such programs.
Given a program with some assertions, the program verification problem is to verify whether or not the assertions are valid. One challenge in program verification is to discover the appropriate invariants at different program points, especially inductive loop invariants that can be used to prove the validity of the given assertions. Note that the issue of discovering counterexamples, in case the assertions are not valid, is described below.
In one implementation, programs are considered that have linear assignments, i.e., assignments of the form x:=e, or nondeterministic assignments x:=?. Assume and assert statements of the form assume(p) and assert(p) are also allowed, where p is some Boolean combination of linear inequalities e≧0. Here x denotes some program variable that takes integral values, and e denotes some linear arithmetic expression. Since assume statements are allowed, conditionals in the program are assumed to be non-deterministic.
One problem in program verification can be reduced to finding solutions to a second-order constraint. The second-order unknowns in this constraint are the unknown program invariants that are inductive and strong enough to prove the desired assertions. To illustrate the process of constraint generation for an example program, consider the program 220 in
To generate such constraints in a more general setting of any arbitrary procedure, a first step is to choose a cut-set. A cut-set is a set of one or more program locations (called cut-points) such that each cycle in the control flow graph passes through some program location in the cut-set. These may be placed anywhere, not necessarily just at loop headers, and thus are considered herein to be arbitrarily located. One simple way to choose a cut-set is to include all targets of back-edges in any depth first traversal of the control-flow graph. (In case of structured programs, where all loops are natural loops, this corresponds to choosing the header node of each loop.) However, as described below, some other cut-set choices may be more desirable from an efficiency/precision viewpoint. For notational convenience, assume that the cut-set always includes the program entry location πentry and exit location πexit.
Each cut-point π is then associated with a relation Iπ over program variables that are live at π. The relations Iπentry and Iπexit at program's entry and exit locations, respectively, are set to true, while the relations at other cut-points are unknown relations that the analyzer seeks to discover. Two cut-points are adjacent if there is a path in the control flow graph from one to the other that does not pass through any other cut-point. Constraints between the relations at adjacent cut-points π1 and π2 are as follows: let Paths(π1, π2) denote the set of paths between π1 and π2 that do not pass through any other cut-point. The notation VC(π1, π2) is used to denote the constraint that the relations Iπ1 and Iπ2 at adjacent cut-points π1 and π2, respectively, are consistent with respect to each other:
Above, X denotes the set of program and fresh variables that occur in Iπ1 and ω(τ,Iπ2). The notation ω(τ,I) denotes the weakest precondition of path π (which is a sequence of program instructions) with respect to I and is as defined below:
where r is some fresh variable and the notation [e/x] denotes substitution of x by e and may not be eagerly carried out across unknown relations. Let π1, π2 range over pairs of adjacent cut-points. Then any solution to the unknown relations Iπ in the following (verification) constraint (which also may have substitutions), yields a valid proof of correctness:
Note that this constraint is universally quantified over the program variables and is a function of {right arrow over (I)}, the vector of relations Iπ at all cut-points (including Iπentry, Iπexit). It is written as the verification constraint ∀X.φ({right arrow over (I)}).
For program verification Iπentry and Iπexit are set to true. Returning to the above example, the second order constraints corresponding to the program in
Turning to constraint solving, and to solving the second-order constraint from Equation 1 that represents the verification condition of unknown relations at cut-points, one way to solve these constraints for discovering the unknown invariants Iπ is to use fixed-point based techniques like abstract interpretation. Another (significantly manual) approach is to require the programmer to provide the invariants at the cut-points, which can then be verified using a theorem prover. Instead, the approach described herein is directed towards reducing the second-order constraint into a Boolean formula such that a satisfying assignment to the formula maps to a satisfying assignment for the second-order constraint. The reduction over the constraints may be illustrated via
This constraint-solving approach involves three main steps. First, some invariant templates 116 (possibly disjunctive) are assumed to reduce the second-order constraints to first-order constraints over the unknown parameters of the templates 116. Farkas' lemma is then used to translate the first-order constraints (with universal quantification) into an existentially quantified multi-linear quadratic constraint. These constraints are then translated into a SAT formula using bit-vector modeling (instead of solving them using specialized mathematical solvers, for example).
To convert second-order unknowns to first-order unknowns, instead of searching for a solution to unknown relations (which are second-order entities) from an arbitrary domain, the search is restricted to a template that is some Boolean combination of linear inequalities among program variables. For example, an unknown relation can have the template:
where ai, bi, ci, di are all unknown integer constants and xi are the program variables. The template can be provided by the user (for example, by specifying the maximum number of conjuncts and disjuncts in DNF representation of any unknown relation), or an iterative scheme may be used that progressively increases the size of the template until a solution is found. Given such templates, the unknown relations in the constraint in Equation 1 are replaced by the templates, thereafter applying any pending substitutions to obtain a first-order logic formula with unknowns that range over integers.
For the example in
Next, first-order universal quantification is translated to first-order existential quantification using Farkas' lemma (at the cost of doing away with some integral reasoning). Farkas' lemma implies that a conjunction of linear inequalities ei≧0 (with integral coefficients) is unsatisfiable over rationals iff (meaning if and only if as used herein) some nonnegative (integral) linear combination of ei yields a negative quantity, i.e.:
The reverse direction of the above lemma is straightforward, since it is not possible for a non-negative linear combination of non-negative expressions ei to yield a negative quantity. The forward direction also holds since the only way to reason about linear inequalities over rationals is to add them, multiply them by a non-negative quantity or add a non-negative quantity. The universal quantification on the right hand side of the above equivalence is over a polynomial equality, and hence can be removed equating the coefficients of the program variables X on both sides of the polynomial equality. Any universally quantified linear arithmetic formula ∀X(φ) may be converted into an existentially quantified formula using Farkas' lemma as follows: convert φ in conjunctive normal form
where each conjunct φi is a disjunctions of inequalities
Note that
and that φi can be rewritten as
whereby Farkas' lemma, as stated above, can be applied to each ∀X(φi).
A next step converts the first-order existentially quantified (or quantifier-free) formula obtained from the previous step into a SAT formula. The formula obtained from that previous step is a conjunction of (multilinear quadratic polynomials) over integer variables. Such a formula may be converted into a SAT formula by modeling integer variables as bit-vectors and encoding integer operations like arithmetic, multiplication, and comparison as Boolean operations over bit-vectors. This approach to constraint solving provides that any satisfying solution to the SAT formula yields a valid proof of correctness. Note that this constraint solving approach is complete if the unknown invariants are instances of given templates, and if checking consistency of invariants at adjacent cut-points does not require integral reasoning.
Turning to the choice of a cut-set, the choice of a cut-set affects the precision and efficiency of any constraint-based technique. Described herein is the choosing of a cut-set that strikes a good balance between precision and efficiency. Based on the definition of a cut-set, it follows that some program locations from each loop need to be included in the cut-set. One strategy is to include all header nodes (or targets of back-edges) as cut-points; such a choice of cut-set necessitates searching/solving for unknown relations over disjunctive relations when the proof of correctness involves a disjunctive loop invariant. Note that for several programs that require disjunctive loop invariants, there is another choice for cut-set that requires searching for unknown relations over only conjunctive domains. Furthermore, the number of conjuncts required is less compared to those required when the header nodes are chosen to be cut-points. This choice for cut-set corresponds to choosing one cut-point on each path inside the loop. If there are multiple sequential conditionals present inside a loop, this requires expanding the control-flow inside the loop into disjoint paths and choosing a cut-point anywhere on each disjoint path. Indeed, this choice for cut-set leads to the greatest precision in the following sense:
Various examples show that the reverse direction in Theorem 1 is not true (i.e., there exists a solution to the unknown relations corresponding to cut-set C, but there is no solution to unknown relations corresponding to some other choice of cut-set). This is illustrated by the example in
Turning to interprocedural analysis, the computation described above is applicable only in an intraprocedural setting. The constraint-based method may be extended to perform a precise (i.e., context sensitive) interprocedural analysis. Precise interprocedural analysis is challenging because the behavior of the procedures needs to be analyzed in a potentially unbounded number of calling contexts. One way to do precise interprocedural analysis is to compute procedure summaries, which are relations between procedure inputs and outputs. These summaries are usually structured as sets of pre/postcondition pairs (Ai,Bi), where Ai is some relation over procedure inputs and Bi is some relation over procedure inputs and outputs. The pre/postcondition pair (Ai,Bi) denotes that whenever the procedure is invoked in a calling context that satisfies constraint Ai, the procedure ensures that the outputs will satisfy the constraint Bi.
As described herein, a constraint-based approach is particularly suited to discovering such useful pre/postcondition (Ai,Bi) pairs. To this end, note that the desired behavior of most procedures can be captured by a small number of such (unknown) pre/postcondition pairs. Procedure calls may be replaced by these unknown behaviors and assert that the procedure, in fact, has such behaviors. More particularly, assume that a procedure does not read/modify any global variables; instead all global variables that are read by the procedure are passed in as inputs, and all global variables that are modified by the procedure are returned as outputs. Suppose there are q interesting pre/postcondition pairs for procedure P(x){S; return y;} with the vector of formal arguments x and vector of return values y. In practice, the value of q can be iteratively increased until invariants are found that make the constraint system satisfiable. The behavior of procedure P may be summarized using q tuples (Ai,Bi) for 1≦i≦q, where Ai is some relation over procedure inputs x, while Bi is some relation over procedure inputs and outputs x and y.
This is indeed the case by generating constraints for each i as below and asserting their conjunction:
assume(Ai); S; assert(Bi); (2)
Procedure calls v:=P(u) on any simple path are compiled away by replacing them with the following code fragment:
Note that in one approach, there is no need, in theory, to have q different pre/postcondition pairs. In fact, the summary of a procedure can also be represented as some formula φ(x, y) (with arbitrary Boolean structure) that represents a relation between procedure inputs x and outputs y. In such a case, it is asserted that φ indeed is the summary of procedure P by generating a constraint for {S; assert(φ(x, y));}, and a procedure call v:=P(u) is compiled away by replacing it by the code fragment v:=?; assume(φ[u/x, v/y]). However, this approach of maintaining multiple symbolic pre/postcondition pairs (which is also inspired by the data structures used by the traditional fixed-point computation algorithms) is more efficient because it enforces more structure on the assume-guarantee proof and leads to lesser unknown quantities and simpler constraints.
Consider the example 440 shown in
Given a program with some assertions, the problem of weakest precondition generation is to infer the weakest precondition Iπentry that ensures that whenever the program is run in a state that satisfies Iπentry, the assertions in the program hold. As described below, a solution to this problem can be a useful tool for a wide range of applications. There is thus described herein a constraint-based approach to inferring weakest preconditions under a given template. Because a precise solution to this problem is undecidable, a relaxed notion of weakest precondition is used, namely for a given template structure, A is considered a weakest precondition if A is a precondition that fits the template and involves constants whose absolute value is at most c (where c is some given constant such that the solutions of interest are those that involve constants whose absolute value is at most c) and there does not exist a weaker precondition than A with similar properties.
A first step in a constraint-based approach to weakest precondition generation is to treat the precondition Iπentry as an unknown relation in Equation 1, unlike in program verification where Iπentry is set to be true. However, this small change merely encodes that any consistent assignment to Iπentry is a valid precondition, not necessarily the weakest one. In fact, when running the analyzer 102 with this small change for any example, it returns false as a solution for Iπentry. Note that false is always a valid precondition, but not necessarily the weakest one. One simple approach to finding the weakest precondition may be to search for a precondition that is weaker than the current solution (which can be easily enforced by adding another constraint to Equation 1), and to iterate until none exists. However, this approach can have a very slow progress. When analyzing the example in
To encode that Iπentry is a weakest precondition, the verification constraint in Equation 1 can be regarded as function of two arguments Iπentry and Ir, where Ir denotes the relations at all cut-points except at the program entry location, and can thus be written as ∀X.φ(Iπentry, Ir). For any other relation I′ that is strictly weaker than Iπentry, it should not be the case that I′ is a valid precondition. This can be stated as the following constraint.
∀X.φ(Iπentry,Ir) Λ ∀I′,I′r(weaker(I′, Iπentry)∀X.φ(I′,I′r))
where weaker
Farkas' lemma cannot be applied here because there is existential quantification nested inside universal quantification.
By way of example, for the procedure 440 in
Described is a binary search strategy, noting that the weakest precondition to be discovered is a conjunctive invariant. This is because the disjunctive weakest precondition can be obtained as disjunctions of disjoint weakest conjunctive solutions.
be some non-false precondition. For any n×(n+1) matrix D of non-negative constants, let I(D) denote the formula
Theorem 2 suggests a binary search-based algorithm 660 (shown in
Note that the preconditions in line 7 and line 11 of
Another algorithm for generating a weakest precondition may be even more efficient. More particularly, assume that each non-trivial maximally strongly connected component in the control flow graph has exactly one cut-point (an assumption that can also be ensured by simple transformations); note however that this will work without this assumption. An algorithm 770 for generating a weakest precondition is described in
The performance of the algorithm depends on the choice of the input neighborhood structure, which affects the number of iterations of the loop in Line 3. A denser neighborhood structure may result in a lesser number of iterations of the while loop (i.e., a lesser number of queries to the constraint solver), but a larger-sized query as a result of the condition in Line 7.
The following neighborhood structure N may be quite efficient in practice. The set of relations that are in the neighborhood N of a conjunctive relation is described below:
The neighborhood structure set forth above has a geometric interpretation. The neighbors of a convex region
are obtained by slightly moving any of the hyper-planes ej≧0 parallel to itself, or by slightly rotating any of the hyper-planes ej≧0 along its intersection with any other hyper-plane el≧0. The neighborhood structure defined above may be extended to relations in disjunctive normal form as:
Note that the above choice of the neighborhood structure helps avoid the repeated iteration over the preconditions i≧j+127, i≧j+126, . . . , i≧j to obtain the weakest precondition i≧j for the example in
However, in general, a locally pointwise-weakest relation with respect to the neighborhood structure defined above may not be a pointwise-weakest relation. For example, consider the program 880 in
It should be noted that the use of a binary search algorithm for weakest precondition generation may be combined with the neighborhood-structure approach for weakest precondition generation. The binary search algorithm provides good theoretical results, while the neighborhood-structure approach provides good practical results.
The problem of strongest postcondition is to generate the most precise invariants at a given cut-point. As in the weakest precondition case, a relaxed notion of strongest postcondition is used. The technique for generating strongest postcondition is very similar to the weakest precondition inference technique described above, and uses the algorithm 990 described in
By way of examples of a more general version of the procedure in
The above-described constraint-based techniques for verification of safety properties may be applied for finding counterexamples to safety properties, verification of termination (which is a liveness property), and finding counterexamples to termination. The termination problem involves checking whether the given procedure terminates under all inputs. The constraint-based approach may be used to solve a more difficult problem, namely bounds analysis. The problem of bounds analysis is to find a worst-case bound on the running time of a procedure (e.g., in terms of the number of instructions executed) expressed in terms of its inputs. One method builds on the techniques that reduce the bounds analysis problem to discovering invariants of a specific kind. In general, this is directed towards computing bounds on loop iterations and number of recursive procedure call invocations. Each of these can be bounded by appropriately instrumenting counter variables and estimating bounds on counter variables. In particular, the number of loop iterations of a while loop “while c do S” can be bounded by computing an upper bound on the instrumented variable i inside the loop in the following code fragment: “i:=0; while c do {i++; S;}”. The number of recursive procedure call invocations of a procedure “P(x){S}” can be bounded similarly by computing an upper bound on global variable i inside procedure P in the following code fragment: “P(x){i:=0; P′(x);}; P′(x′){i++; S[x′/x];}”.
In general, let P be a given program and let P′ be the transformed program obtained after instrumenting counters that keep track of loop iterations and recursive call invocations and introducing partial assertions at appropriate locations that assert that the counters are bounded above by some function of the inputs. The program P terminates iff the assert statements in P′ are satisfied. While invariant generation tools such as abstract interpretation can be used to compute bounds on the counter variables, a constraint-based approach is particularly suited for discovering these invariants since they have a specified form and involve linear arithmetic. To this end, assert statements are used with templates
(at the counter increment i++ site in case of loops and at the end of the procedure in case of recursive procedures) for bounding the counter variable. Note that in addition to the counter instrumentation strategy mentioned above, other counter instrumentation strategies can be used to compute non-linear bounds as a composition of linear bounds on multiple instrumentation counters. Such techniques can also be used herein to compute non-linear bounds using linear templates. Additionally, the constraint-based approach solves an even more difficult problem, namely inferring preconditions under which the procedure terminates and inferring a bound under that precondition.
For this purpose, the analyzer tool is run in the weakest precondition inference mode. This is particularly significant when procedures are expected to be called under certain preconditions and would not otherwise terminate under all inputs. By way of example, in
Because program analysis is an undecidable problem, there are no tools that can prove correctness of all correct programs or find bugs in all incorrect programs. To maximize the practical success rate of verification tools, it is desirable to search for both proofs of correctness as well as counterexamples in parallel. To find most-general counterexamples to safety properties, the problem of generating a most-general counterexample for a given set of (safety) assertions involves finding the most general characterization of inputs that leads to the violation of some reachable safety assertion.
To find such a characterization using the techniques described above, a general aspect is directed towards reducing the problem to that of finding the weakest precondition for an assertion. This reduction involves constructing another program from the given program P using the following transformations:
In general, let P be a program with some safety assertions. Let P′ be the program obtained from program P by using the transformation described above. Then, P has an assertion violation iff the assertions in program P′ hold; this does hold, meaning that the weakest precondition inference on the transformed program may be used to discover the most-general characterization of inputs under which there is a safety violation in the original program.
By way of example, the program 1102 shown in
The problem of generating a most-general counterexample for program termination involves finding the most-general characterization of inputs that leads to non-termination of the program. Assume that the program has at most one exit point, that is, let P be a given program with a single exit point. Let P′ be the program obtained from P by adding the assert statement “assert(false)” at the end of the program. Then, P is nonterminating iff the assert statement in P′ is satisfied.
As a result the weakest precondition inference may be used on the transformed program to discover most-general characterization of inputs under which the original program is non-terminating. By way of example, consider the example program 1202 shown in
As can be seen, there is thus provided a constraint-based approach to invariant generation in programs that translates a program into constraints that are solved using existing constraint solvers to yield desired program invariants. The generated constraints are Boolean combinations of quadratic inequalities over integer variables, which may be reduced to SAT formulae using bit-vector modeling, with existing SAT solvers used to solve them.
In one implementation, the constraint-based approach can be used to model a wide spectrum of program analyses in an expressive domain containing disjunctions and conjunctions of linear inequalities. Further described is modeling the problem of context-sensitive interprocedural program verification, as well as a constraint-based approach to weakest precondition and strongest postcondition inference. Also described are applications of the above analyses, including bounds analysis and generation of most-general counter-examples for both safety and termination properties.
Exemplary Operating Environment
The invention is operational with numerous other general purpose or special purpose computing system environments or configurations. Examples of well known computing systems, environments, and/or configurations that may be suitable for use with the invention include, but are not limited to: personal computers, server computers, hand-held or laptop devices, tablet devices, multiprocessor systems, microprocessor-based systems, set top boxes, programmable consumer electronics, network PCs, minicomputers, mainframe computers, distributed computing environments that include any of the above systems or devices, and the like.
The invention may be described in the general context of computer-executable instructions, such as program modules, being executed by a computer. Generally, program modules include routines, programs, objects, components, data structures, and so forth, which perform particular tasks or implement particular abstract data types. The invention may also be practiced in distributed computing environments where tasks are performed by remote processing devices that are linked through a communications network. In a distributed computing environment, program modules may be located in local and/or remote computer storage media including memory storage devices.
With reference to
The computer 1310 typically includes a variety of computer-readable media. Computer-readable media can be any available media that can be accessed by the computer 1310 and includes both volatile and nonvolatile media, and removable and non-removable media. By way of example, and not limitation, computer-readable media may comprise computer storage media and communication media. Computer storage media includes volatile and nonvolatile, removable and non-removable media implemented in any method or technology for storage of information such as computer-readable instructions, data structures, program modules or other data. Computer storage media includes, but is not limited to, RAM, ROM, EEPROM, flash memory or other memory technology, CD-ROM, digital versatile disks (DVD) or other optical disk storage, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices, or any other medium which can be used to store the desired information and which can accessed by the computer 1310. Communication media typically embodies computer-readable instructions, data structures, program modules or other data in a modulated data signal such as a carrier wave or other transport mechanism and includes any information delivery media. The term “modulated data signal” means a signal that has one or more of its characteristics set or changed in such a manner as to encode information in the signal. By way of example, and not limitation, communication media includes wired media such as a wired network or direct-wired connection, and wireless media such as acoustic, RF, infrared and other wireless media. Combinations of the any of the above may also be included within the scope of computer-readable media.
The system memory 1330 includes computer storage media in the form of volatile and/or nonvolatile memory such as read only memory (ROM) 1331 and random access memory (RAM) 1332. A basic input/output system 1333 (BIOS), containing the basic routines that help to transfer information between elements within computer 1310, such as during start-up, is typically stored in ROM 1331. RAM 1332 typically contains data and/or program modules that are immediately accessible to and/or presently being operated on by processing unit 1320. By way of example, and not limitation,
The computer 1310 may also include other removable/non-removable, volatile/nonvolatile computer storage media. By way of example only,
The drives and their associated computer storage media, described above and illustrated in
The computer 1310 may operate in a networked environment using logical connections to one or more remote computers, such as a remote computer 1380. The remote computer 1380 may be a personal computer, a server, a router, a network PC, a peer device or other common network node, and typically includes many or all of the elements described above relative to the computer 1310, although only a memory storage device 1381 has been illustrated in
When used in a LAN networking environment, the computer 1310 is connected to the LAN 1371 through a network interface or adapter 1370. When used in a WAN networking environment, the computer 1310 typically includes a modem 1372 or other means for establishing communications over the WAN 1373, such as the Internet. The modem 1372, which may be internal or external, may be connected to the system bus 1321 via the user input interface 1360 or other appropriate mechanism. A wireless networking component 1374 such as comprising an interface and antenna may be coupled through a suitable device such as an access point or peer computer to a WAN or LAN. In a networked environment, program modules depicted relative to the computer 1310, or portions thereof, may be stored in the remote memory storage device. By way of example, and not limitation,
An auxiliary subsystem 1399 (e.g., for auxiliary display of content) may be connected via the user interface 1360 to allow data such as program content, system status and event notifications to be provided to the user, even if the main portions of the computer system are in a low power state. The auxiliary subsystem 1399 may be connected to the modem 1372 and/or network interface 1370 to allow communication between these systems while the main processing unit 1320 is in a low power state.
While the invention is susceptible to various modifications and alternative constructions, certain illustrated embodiments thereof are shown in the drawings and have been described above in detail. It should be understood, however, that there is no intention to limit the invention to the specific forms disclosed, but on the contrary, the intention is to cover all modifications, alternative constructions, and equivalents falling within the spirit and scope of the invention.
Number | Name | Date | Kind |
---|---|---|---|
5867649 | Larson | Feb 1999 | A |
6553362 | Saxe et al. | Apr 2003 | B2 |
7024661 | Leino et al. | Apr 2006 | B2 |
7089542 | Brand et al. | Aug 2006 | B2 |
7120902 | Flanagan et al. | Oct 2006 | B2 |
7283945 | Rajan | Oct 2007 | B2 |
7454399 | Matichuk | Nov 2008 | B2 |
7574695 | Chander et al. | Aug 2009 | B2 |
7797687 | Tillmann et al. | Sep 2010 | B2 |
8046746 | Tillmann et al. | Oct 2011 | B2 |
20020046393 | Leino et al. | Apr 2002 | A1 |
20020112201 | Flanagan et al. | Aug 2002 | A1 |
20020133806 | Flanagan et al. | Sep 2002 | A1 |
20040019468 | De Moura et al. | Jan 2004 | A1 |
20050010907 | Namjoshi | Jan 2005 | A1 |
20050027495 | Matichuk | Feb 2005 | A1 |
20050166095 | Chander et al. | Jul 2005 | A1 |
20050166167 | Ivancic et al. | Jul 2005 | A1 |
20050229044 | Ball | Oct 2005 | A1 |
20060247907 | Qadeer et al. | Nov 2006 | A1 |
20060282807 | Ivancic et al. | Dec 2006 | A1 |
20070033576 | Tillmann et al. | Feb 2007 | A1 |
20070088740 | Davies et al. | Apr 2007 | A1 |
20070168988 | Eisner et al. | Jul 2007 | A1 |
20070277163 | Avresky | Nov 2007 | A1 |
20070299793 | Musuvathi et al. | Dec 2007 | A1 |
20080320486 | Bose et al. | Dec 2008 | A1 |
20090007038 | Wang et al. | Jan 2009 | A1 |
20100281306 | Sinha | Nov 2010 | A1 |
Entry |
---|
Dirk Beyer, Thomas A. Henzinger, Rupak Majumdar, and Andrey Rybalchenko. 2007. Path invariants. SIGPLAN Not. 42, 6 (Jun. 2007), 300-309. DOI=10.1145/1273442.1250769 http://doi.acm.org/10.1145/1273442.1250769. |
Randal E. Bryant, Daniel Kroening, J. Ouaknine, Sanjit A. Seshia, Ofer Strichman, and Bryan Brady. 2007. Deciding bit-vector arithmetic with abstraction. In Proceedings of the 13th international conference on Tools and algorithms for the construction and analysis of systems (TACAS'07), Orna Grumberg and Michael Huth (Eds.). Springer-Verlag, Berlin. |
Colón, Michael; Sankaranarayanan, Sriram; Sipma, Henny “Linear Invariant Generation Using Non-linear Constraint Solving” Computer Aided Verification Lecture Notes in Computer Science 2003. Springer Berlin / Heidelberg p. 420-432. |
Sumit Gulwani and Ashish Tiwari. 2007. Computing procedure summaries for interprocedural analysis. In Proceedings of the 16th European conference on Programming (ESOP'07), Rocco De Nicola (Ed.). Springer-Verlag, Berlin, Heidelberg, 253-267. |
Wei-Ngan Chin, Siau-Cheng Khoo, and Dana N. Xu. 2001. Deriving Pre-conditions for Array Bound Check Elimination. In Proceedings of the Second Symposium on Programs as Data Objects (PADO '01), Olivier Danvy and Andrzej Filinski (Eds.). Springer-Verlag, London, UK, 2-24. |
Fabian Bannwart and Peter Mueller. 2005. A Program Logic for Bytecode. Electron. Notes Theor. Comput. Sci. 141, 1 (Dec. 2005), 255-273. DOI=10.1016/j.entcs.2005.02.026 http://dx.doi.org/10.1016/j.entcs.2005.02.026. |
B. Lisper. Fully Automatic, Parametric Worst-Case Execution Time Analysis. In Proc. WCET '03: Intl. Workshop on Worst-Case Execution Time Analysis, pp. 99-102, 2003.). |
Aaron R. Bradley and Zohar Manna. 2007. Checking Safety by Inductive Generalization of Counterexamples to Induction. In Proceedings of the Formal Methods in Computer Aided Design (FMCAD '07). IEEE Computer Society, Washington, DC, USA, 173-180. |
Anirban Majumdar, Clark Thomborson, and Stephen Drape. 2006. A survey of control-flow obfuscations. In Proceedings of the Second international conference on Information Systems Security (ICISS'06), Aditya Bagchi and Vijayalakshmi Atluri (Eds.). Springer-Verlag, Berlin, Heidelberg, 353-356. |
Ball, Thomas Formalizing Counterexample-Driven Refinement with Weakest Preconditions Engineering Theories of Software Intensive Systems. NATO Science Series 2005 Springer Netherlands p. 121-139 vol. 195 <http://www.springerlink.com/content/hk552416k67w4560/>. |
Csallner, et al., “DySy: Dynamic Symbolic Execution for Invariant Inference”, Microsoft Research Technical Report, 2007, pp. 1-10. |
Xie, et al., “Path Sensitive Program Analysis Using Boolean Satisfiability”, 2002, ACM, pp. 10. |
Chauhan, et al., “Automated Abstraction Refinement for Model Checking Large State Spaces using SAT based Conflict Analysis”, Proceedings of the 4th International Conference on Formal Methods in Computer-Aided Design, vol. 2517, 2002, pp. 1-18. |
Jackson, et al., “Finding Bugs with a Constraint Solver”, ACM SIGSOFT Software Engineering Notes, vol. 25, No. 5, 2000, pp. 12. |
Balaban, et al., “Ranking Abstraction of Recursive Programs”, Springer-Verlag Berlin Heidelberg, 2006, pp. 267-281. |
Berdine, et al., “Variance Analyses from Invariance Analyses”, Principles of Programming Languages, 2007, ACM, pp. 211-224. |
Beyer, et al., “Invariant Synthesis for Combined Theories”, In Proceedings of the Eighth International Conference on Verification, Model Checking, and Abstract Interpretation (VMCAI 2007, Jan. 14-16), LNCS 4349, pp. 378-394, 2007. Springer-Verlag, pp. 17. |
Beyer, et al., “Path Invariants”, Technical Report, Models and Theory of Computation Laboratory, Dec. 22, 2006, pp. 16. |
Bradley, et al., “Verification Constraint Problems with Strengthening”, Springer Berlin / Heidelberg, vol. 4281, 2006, pp. 15, Jun. 12, 2012. |
Bradley, et al., “Linear Ranking with Reachability”, Springer Berlin / Heidelberg, vol. 3576, 2005, pp. 1-55. |
Colon, et al., “Linear Invariant Generation Using Non-linear Constraint Solving”, Springer-Verlag Berlin Heidelberg 2003, pp. 420-432. |
Cousot, et al., “A Unified Lattice Model for Static Analysis of Programs by Construction or Approximation of Fixpoints”, 1977, ACM, pp. 238-252. |
Gonnord, et al., “Combining Widening and Acceleration in Linear Relation Analysis”, Springer Berlin / Heidelberg, vol. 4134, 2006, pp. 1-17. |
Gulavani, et al., “Automatically Refining Abstract Interpretations”, Jul. 2007, Technical Report, pp. 20. |
Gulavani, et al., “Counterexample Driven Refinement for Abstract Interpretation”, LNCS.3920, Springer-Verlag Berlin Heidelberg, 2006, pp. 474-488. |
Gupta, et al., “Proving Non-Termination”, 2008, ACM, pp. 12. |
Kapur, “Automatically Generating Loop Invariants Using Quantifier Elimination”, Applications of Computer Algebra (ACA 2004), ACA & Lamar University, 2004, pp. 1-17. |
Kildall, “A Unified Approach to Global Program Optimization”, Proceedings of the 1st annual ACM SIGACT-SIGPLAN Symposium on Principles of Programming Languages, ACM, 1973, pp. 194-206. |
Manna, et al., “Formalization of Properties of Functional Programs”, Journal of the Association for Computing Machinery, vol. 17, No. 3, Jul. 1970, pp. 555-569. |
Muller-Olm, et al., “Precise Interprocedural Analysis through Linear Algebra”, ACM, 2004, pp. 12. |
Muller-Olm, et al., “Interprocedural Analysis (Almost) for Free”, 2004, pp. 1-15. |
Podelski, et al., “A Complete Method for the Synthesis of Linear Ranking Functions”, Springer Berlin / Heidelberg, vol. 2937, 2004, pp. 239-251. |
Sankaranarayanan, et al., “Non-linear Loop Invariant Generation using Gröbner Bases”, 2004, ACM, pp. 318-329. |
Sankaranarayanan, et al., “Scalable Analysis of Linear Systems Using Mathematical Programming”, vol. 3385, Springer-Verlag Berlin Heidelberg, 2005, pp. 25-41. |
Seidl, et al., “Interprocedurally Analysing Linear Inequality Relations”, vol. 4421, Springer-Verlag Berlin Heidelberg, 2007, pp. 284-299. |
Wang, et al., “Using Counterexamples for Improving the Precision of Reachability Computation with Polyhedra”, Springer Berlin / Heidelberg, vol. 4590, 2007, pp. 1-14. |
Number | Date | Country | |
---|---|---|---|
20090326907 A1 | Dec 2009 | US |