This disclosure relates generally to quantum interior point methods (QIPMs), and more particularly to implementing quantum interior point methods (QIPMs).
The practical utility of finding optimal solutions to well-posed optimization problems has been known since the days of antiquity. With the advent of the quantum era, there has been great interest in developing quantum algorithms that solve optimization problems with provable speedups over classical algorithms. Unfortunately, it can be difficult to implement these quantum algorithms and evaluate whether these quantum algorithms will be practically useful.
In some aspects, the techniques described herein relate to a quantum interior point method (QIPM) for solving a second-order cone program (SOCP) instance using a quantum computing system, the method including: receiving the SOCP instance; defining a Newton system for the SOCP instance by constructing matrix G and vector h, where matrix G and vector h describe constrains for a linear system Gu=h based on the SOCP instance; preconditioning matrix G and vector h via row normalization to reduce a condition number of matrix G; iteratively determining u until a predetermined iteration condition is met, the iterations including: causing the quantum computing system to apply matrix G and vector h to a quantum linear system solver (QLSS) to generate a quantum state; causing the quantum computing system to perform quantum state tomography on the quantum state; and updating a value of u based on a current value of u and the output of the quantum state tomography; and determining a solution to the SOCP instance based on the updated value of u.
Other aspects include components, devices, systems, improvements, methods, processes, applications, computer readable mediums, and other technologies related to any of the above.
Embodiments of the disclosure have advantages and features which will be more readily apparent from the following detailed description and the appended claims, when taken in conjunction with the examples in the accompanying drawings, in which:
The figures and the following description relate to preferred embodiments by way of illustration only. It should be noted that from the following discussion, alternative embodiments of the structures and methods disclosed herein will be readily recognized as viable alternatives that may be employed without departing from the principles of what is claimed.
This disclosure studies quantum interior point methods (QIPMs) for second-order cone programming (SOCP), guided by the example use case of portfolio optimization (PO). This disclosure provides a complete quantum circuit-level description of the algorithm from problem input to problem output, making several improvements to the implementation of the QIPM. This disclosure reports the number of logical qubits and the quantity/depth of non-Clifford T-gates used to run the algorithm, including constant factors. The determined resource counts depend on instance-specific parameters, such as the condition number of certain linear systems within the problem. To determine the size of these parameters, numerical simulations of small PO instances are performed, which lead to concrete resource estimates for the PO use case. The numerical results do not probe large enough instance sizes to make conclusive statements about the asymptotic scaling of the algorithm. However, already at small instance sizes, the analysis suggests that, due primarily to large constant pre-factors, poorly conditioned linear systems, and a fundamental reliance on costly quantum state tomography, fundamental improvements to the QIPM are desired for it to lead to practical quantum advantage.
The practical utility of determining optimal solutions to well-posed optimization problems has been known since the days of antiquity, with Euclid considering the minimal distance between two points using a line. In the modern era, optimization algorithms for business and financial use cases continue to be ubiquitous. Partly as a result of this utility, algorithmic techniques for optimization problems have been well studied since even before the invention of the computer, including a famous dispute between Legendre and Gauss on who was responsible for the invention of least squares fitting. With the advent of the quantum era, there has been great interest in developing quantum algorithms that solve optimization problems with provable speedups over classical algorithms.
Unfortunately, it can be difficult to evaluate whether these quantum algorithms will be practically useful. In some cases, the algorithms are heuristic, and their performance can only be measured empirically once it is possible to run them on actual quantum hardware. In other cases, the difficulty in evaluating practicality stems from the inherent complexity of combining many distinct ingredients, each with their own caveats and bottlenecks. To make an apples-to-apples comparison and quantify advantages of a quantum algorithm, an end-to-end resource analysis that accounts for all costs from problem input to problem output may be performed.
Such an end-to-end analysis for a quantum interior point method (QIPM) was performed for solving second-order cone programs (SOCPs). In particular, this disclosure focuses on a concrete use case with very broad applications, but of interest in the financial services sector: portfolio optimization (PO). In general, PO is the task of determining the optimal resource allocation to a collection of possible classes to optimize a given objective. In finance, one seeks to determine the optimal allocation of funds across a set of possible assets that maximizes returns and minimizes risk, subject to constraints. Noteworthy, many variants of the PO problem can be cast as a SOCP and subsequently solved with a classical or quantum interior point method. Indeed, classical interior point methods (CIPMs) are efficient not only in theory, but also in practice; they are the method of choice within fast numerical solvers for SOCPs and other conic programs, which encompass a large variety of optimization problems that appear in industry. Notably, QIPMs structurally mirror CIPMs, and seek improvements by replacing certain subroutines with quantum primitives. Thus, compared to other proposed quantum algorithms for conic programs not based on widely used classical techniques (e.g., solvers that leverage the multiplicative weights update method), QIPMs are uniquely positioned to provide not only a theoretical asymptotic advantage, but also a practical quantum solution for this common class of problem.
However, the QIPM is a complex algorithm that delicately combines some purely classical steps with multiple distinct quantum subroutines. The runtime of the QIPM is stated in terms of several parameters that can only be evaluated once a particular use case has been specified; depending on how these parameters scale, an asymptotic speedup may or may not be achievable. Additionally, any speedup is contingent on access to a large quantum random access memory (QRAM), an ingredient that in prior asymptotic-focused analyses has typically been assumed to exist without much further justification or cost analysis.
The resource analysis is detailed and takes care to study aspects of the end-to-end pipeline, including the QRAM component. This disclosure reports results in terms of relevant problem parameters, and then describes numerical experiments to determine the size and scaling of these parameters for actual randomly chosen instances of the PO problem, based on historical stock data. This approach allows us to estimate the exact resource cost of the QIPM for an example PO problem, including a detailed breakdown of costs by various subroutines. This estimate incorporates several optimizations to the underlying subroutines, and technical improvements to how they are integrated into the QIPM. Consequently, our analysis allows us to evaluate the prospect that the algorithm may exhibit a practical quantum advantage, and it reveals the computational bottlenecks within the algorithm that are most in need of further improvement.
While this disclosure focuses on the QIPM and its application to the PO problem, this disclosure has more general applications and more general takeaways for quantum algorithms and for quantum computing applications. Firstly, the results emphasize the importance of end-to-end analysis when evaluating a proposed application. Furthermore, the modular treatment of the underlying algorithmic primitives produces quantitative and qualitative takeaways that are relevant for end-to-end treatments of a large number of other algorithms that also rely on these subroutines, especially those in the area of machine learning, where data access via QRAM and quantum linear algebra techniques are often used.
The resource analysis focuses on three central quantities that determine the overall cost of algorithms implemented on fault-tolerant quantum computers: the number of logical qubits, the total number of T gates (“T-count”), and the number of parallel layers of T gates (“T-depth”) useed to construct quantum circuits for solving the problem. The T-depth acts as a proxy for the overall runtime of the algorithm, whereas the T-count and number of logical qubits are helpful for determining how many physical qubits may be used for a full, fault-tolerant implementation. We justify the focus on T gates by pointing out that, in many prominent approaches to fault-tolerant quantum computation, quantum circuits are decomposed into Clifford gates and T gates, and the cost of implementing the circuit is dominated by the number and depth of the T gates. The fault-tolerant Clifford gates can be performed transversally or even in software, whereas the T gates use the expensive process of magic state distillation. This disclosure stops short of a full analysis of the algorithm at the physical level, as the logical analysis seems to suffice to evaluate the overall outlook for the algorithm and identify its main bottlenecks.
At the core of any interior point method (IPM) is the solving of a linear system of equations. The QIPM performs this step using a quantum linear system solver (QLSS) together with pure state quantum tomography. The cost of QLSS depends on a parameter κF, the Frobenius condition number ∥G∥F∥G−1∥ of the matrix G that is inverted (where ∥⋅∥F denotes the Frobenius norm, and ∥⋅∥ denotes the spectral norm), while the cost of tomography depends on a parameter ξ, a precision parameter. These parameters are evaluated empirically by simulating the QIPM on small instances of the PO problem.
Table I reports a summary of overall resource calculation, in which the asymptotically leading term is shown (along with its constant prefactor) in terms of parameters κF and ξ, as well as n, the number of assets in the PO instance, and ∈, the desired precision to which the portfolio should be optimized. It is determined (numerically) that κF grows with n, and that ξ shrinks with n; it is estimated that, at n=100 and ∈=10−7, the implementation of the QIPM may use 8×106 qubits and 8×1029 total T gates spread out over 2×1024 layers. These resource counts are decidedly out of reach both in the near and far term for quantum hardware, even for a problem of modest size by classical standards. Even if quantum computers one day match the gigahertz-level clock-speeds of modern classical computers, 1024 layers of T gates would take millions of years to execute. By contrast, the PO problem can be easily solved in a matter of seconds on a laptop for n=100 stocks.
This disclosure cautions that the numbers reported should not be interpreted as the final word on the cost of the QIPM for PO. Further examination of the algorithm may uncover many improvements and optimizations that may reduce the costs compared to the current calculations. On the other hand, the results do already incorporate several innovations made to reduce the resource cost, including preconditioning the linear system.
Besides the main resource calculation, this disclosure makes several additional contributions and observations:
The outline for the remainder of this disclosure is as follows. Section II describes and defines the portfolio optimization problem in terms of Markowitz portfolio theory. Section III describes Second Order Cone Programming (SOCP) problems, illustrate how portfolio optimization can be represented as an instance of SOCP, and discuss how IPMs can be used for solving SOCPs. Section IV review the quantum ingredients used to turn an IPM into a QIPM. In particular, this disclosure reviews quantum linear system solvers, block-encoding for data loading, and quantum state tomography for data read out. This disclosure also presents better bounds on the tomography procedure than were previously known. Section V describes the implementation of using QIPM and quantum algorithms for SOCP for the portfolio optimization problem, including a detailed resource estimate for the end-to-end problem. Section VI shows numerical results from simulations of the full problem, and section VII reflects on the calculations performed, identifying the main bottlenecks and drawing conclusions about the outlook for quantum advantage with QIPM.
The QIPM has many moving parts using several mathematical symbols. While all symbols are defined as they are introduced in the text, this disclosure also provides a full list of symbols for the reader's reference in the section Additional Information A. Throughout the paper, all vectors are denoted in bold lowercase letters to contrast with scalar quantities (unbolded lowercase) and matrices (unbolded uppercase). The only exception to this rule will be the symbols N, K, and L, which are positive integers (despite being uppercase), and denote the number of rows or columns in certain matrices related to an SOCP instance.
Portfolio optimization is the process widely used, for example, by financial analysts to assign allocations of capital across a set of assets within a portfolio, given optimization criteria such as maximizing the expected return and minimizing the financial risk. The creation of the mathematical framework for modern portfolio theory (MPT) is credited to Harry Markowitz, for which he received the 1990 Alfred Nobel Memorial Prize in Economic Sciences. Markowitz describes the process of selecting a portfolio in two stages, where the first stage starts with “observation and experience” and ends with “beliefs about the future performances of available securities.” The second stage starts with “the relevant beliefs about future performances” and ends with “the choice of portfolio.” The theory is also known as mean-variance analysis.
Typically, portfolio optimization strategies include diversification, which is the practice of investing in a wide array of asset types and classes as a risk mitigation strategy. Some popular asset classes are stocks, bonds, real estate, commodities, and cash. After building a portfolio, one may expect a return (or profit) after a specific period of time. Risk is defined as the fluctuations of the asset value. MPT describes how high variance assets can be combined with other uncorrelated assets through diversification to create portfolios with low variance on their return. Naturally, among equal-risk portfolios, investors prefer those with higher expected return, and among equal-return portfolios, they prefer those with lower risk.
Within a portfolio, wi represents the amount of an asset i being held over some period of time. Often, this amount is given as the asset's price in dollars at the start of the period. When the price is positive (wi>0), this is referred to as a long position; and when the price is negative (wi<0), this is referred to as a short position with an obligation to buy this asset at the end of the period. Typically, investment banks hold long positions, while hedge funds build portfolios with short positions that have higher risk due to the uncertainty of the price to buy the asset at the end of the period. The optimization variable in the portfolio optimization problem is the vector of n assets w∈ in the portfolio.
The price of each asset i varies over time. ui is defined to be the relative change (positive or negative) during the period of interest. Then, the return of the portfolio for that period is defined as
To capture realistic problem formulations, one or more mathematical constraints may be added to the optimization problem corresponding to the problem-specific considerations. For example, two common constraints in portfolio optimization problems are no short positions (wi≥0 for all i, denoted by w≥0) and that the total investment budget is limited (1Tw=1, where 1 denotes the vector of ones). This forms the classical portfolio optimization problem from Markowitz's mean-variance theory:
This formulation is a quadratic optimization problem where risk is minimized, while achieving a target return of at least
This optimization problem is no longer a QO problem, but it can be mapped to a conic problem, as described later in section III B. Depending on the problem, additional constraints can be added. For instance, constraints can be added to allow short positions, component-wise short sale limits, or a total short sale limit. Another variant of this is a constraint for a collateralization requirement, which limits the total of short positions to a fraction of the total long positions. Often, buying or selling an asset results in a transaction fee that is proportional to the amount of asset that is bought or sold. Linear transaction costs or maximum transaction amounts are often included as constraints in portfolio optimization. Diversification constraints can limit portfolio risk by limiting the exposure to individual positions and groups of assets within particular sectors. To illustrate the flexibility of this analysis, a maximum transaction constraint is included and use the following problem formulation is used in the analysis in the rest of the disclosure:
where
Second-order cone programming (SOCP) is a type of convex optimization that allows for a richer set of constraints than linear programming (LP), without many of the complications of semidefinite programming (SDP). Indeed, SOCP is a subset of SDP, but SOCP admits interior point methods (IPMs) that may be just as efficient as IPMs for LP. Many real-world problems can be cast as SOCP, including the example portfolio optimization problem of interest.
For any k-dimensional vector v, the following may be used v=(v0; {tilde over (v)}), where v0 is the first entry of v, and {tilde over (v)} contains the remaining k−1 entries.
Definition 1. A k-dimensional second-order cone (for k≥2) is the convex set
={(x0;{tilde over (x)})∈|x0≥∥{tilde over (x)}∥}, (4)
where ∥⋅∥ denotes the vector two-norm (standard Euclidean norm). For k=1, ={x0∈|x0≥0}.
Definition 2. In general, a second-order cone problem is formulated as
where =× . . . × is a Cartesian product of r second-order cones of combined dimension N=N1+ . . . +Nr, and A is a full-rank K×N matrix encoding K linear equality constraints, with K≤N.
Note that the special case of linear programming is immediately recovered if Ni=1 for all i. We say that a point x is primal feasible whenever Ax=b and x∈. It is strictly primal feasible if additionally it lies in the interior of .
The dual to problem in eq. (5) is a maximization problem over a variable y∈, given as follows:
We say that a pair (s; y) is dual feasible whenever ATy+S=c and s∈. For any point (x; y; s) with x, s∈, the duality gap may be defined as
where r is the number of cones, as in definition 2, and the second equality holds under the additional assumption that the point is primal and dual feasible. The fact that x, s∈ implies that μ(x, s)≥0. Moreover, assuming that both the primal and dual problems have a strictly feasible point, the optimal primal solution x* and the optimal dual solution (y*; s*) are guaranteed to exist and satisfy cTx*=bTy*, and hence
Thus, the primal-dual condition of optimality can be expressed by the system
Ax=b
A
T
y+s=c
x
T
s=0
x∈
,s∈
. (8)
The portfolio optimization problem can be solved by reduction to SOCP, and this reduction is often made in practice. Here this disclosure describes one way of translating the portfolio optimization problem, as given in eq. (3) into a second-order cone program.
The objective function in eq. (3) has a non-linear term q√{square root over (wTΣw)}, which may be linearized by introducing a new scalar variable t, and a new constraint t≥√{square root over (wTΣw)}. This results in the equivalent optimization problem
A goal now is to express the constraints in eq. (9) as second-order cone constraints. Given an m×n matrix M for which Σ=MTM, the constraint on t can be expressed by introducing an m-dimensional variable η subject to the equality constraint η=Mw and the second-order cone constraint (t; η) ∈.
The matrix M can be determined from Σ via a Cholesky decomposition, although for large matrices Σ, this computation may be costly. Alternatively, if Σ and {circumflex over (μ)} are calculated from stock return vectors u(1), . . . , u(m) during m independent time epochs (e.g. returns for each of m days or each of m months), then a valid matrix MT is given by (u(1)−û, . . . , u(m)−û), i.e. the columns of MT are given by the deviation of the returns from the mean in each epoch. This is the approach taken in the numerical experiments herein (presented later).
The absolute value constraints are handled by introducing a pair of n-dimensional variables ϕ and ρ, subject to equality constraints ϕ=ζ−(w−
In summary, the portfolio optimization problem from eq. (3) may be described as the following SOCP that minimizes over the variable x=(w; ϕ; ρ; t; η) ∈
where I denotes an identity block, 0 denotes a submatrix of all 0s, 0 is a vector of all 0s, 1 is a vector of all 1s, and the size of each block of A can be inferred from its location in the matrix. Thus, the total number of cones is r=3n+1, and the combined dimension is N=3n+m+1. Note that r=2n+1 cones if the absolute value constraints are represented using dimension-2 cones. The SOCP constraint matrix A is a K×N matrix, with K=2n+m+1.
Notice that many of the rows of the K×N matrix A are sparse and contain only one or two nonzero entries. However, the final m rows of the matrix A will be dense and contain n+1 nonzero entries due to the appearance of the matrix M containing historical stock data; in total a constant fraction of the matrix entries will be nonzero, so sparse matrix techniques will provide only limited benefit.
Additionally, note that the primal SOCP in eq. (10) has an interior feasible point as long as (has strictly positive entries. To better understand this, choose w to be any strictly positive vector that satisfies |w−
Interior point methods (IPMs) are a class of efficient algorithms for solving convex optimization problems including LPs, SOCPs, and SDPs, where (in contrast to the simplex method) intermediate points generated by the method lie in the interior of the convex set, and they are guaranteed to approach the optimal point after a polynomial number of iterations of the method. Each iteration involves forming a linear system of equations that depends on the current intermediate point. The solution to this linear system determines the search direction, and the next intermediate point is formed by taking a small step in that direction. This disclosure considers path-following primal-dual IPMs, where, if the step size is sufficiently small, the intermediate points are guaranteed to approximately follow the central path, which ends at the optimal point for the convex optimization problem.
To define the central path, first establish some notation related to the algebraic properties of the second-order cone. Let the product u∘v of two vectors u=(u0; ũ), v=(v0; {tilde over (v)}) ∈ be defined as
u∘v=(uTV;u0{tilde over (v)}+v0ũ) (11)
and the identity element for this product is denoted by the vector e=(1; 0)∈. For the Cartesian product =× . . . × of multiple second-order cones, the vector e is defined as the concatenation of the identity element for each cone, and the circle product of two vectors is given by the concatenation of the circle product of each constituent. A consequence of this definition is the that eTe is equal to the number of cones r.
Now, for the SOCP problem of eq. (5), the central path (x(v); y(v); s(v)) is the one-dimensional set of central points, parameterized by v∈[0, ∞), which satisfies the conditions:
Ax(v)=b
A
T
y(v)+s(v)=c
x(v)∘s(v)=ve
x(v)∈,s(v)∈. (12)
Note that the central path point (x(v); y(v); s(v)) has a duality gap that satisfies μ(x(v), s(v))=v, and that when v=0, eq. (12) recovers eq. (8).
Path-following primal-dual interior point methods determine the optimal point by beginning at a central point with v>0 and following the central path to a very small value of v, which is taken to be a good approximation of the optimal point. For a given SOCP, determining an initial point on the central path is non-trivial and, in general, can be just as hard as solving the SOCP itself. One solution to this problem is the homogeneous self-dual embedding, a slightly larger self-dual SOCP is formed with the properties that (i) the optimal point for the original SOCP can be determined from the optimal point for the self-dual SOCP and (ii) the self-dual SOCP has a trivial central point that can be used to initialize the IPM.
To do this, introduce new scalar variables τ, θ, and κ, which are used to give more flexibility to the constraints. Previously, Ax=b was used. In the larger program, this constraint is relaxed to read Ax=bτ−(b−Ae)θ, such that the original constraint is recovered when τ=1 and θ=0, but x=e is a trivial solution when τ=1 and θ=1. Similarly, the constraint ATy+s=c is relaxed to read ATy+s=cτ−(c−e)θ, which has the trivial solution y=0, s=e when τ=θ=1. These can be complemented with two additional linear constraints to form the program:
where
Note that if the point (x; y; τ; θ; s; K) is feasible, i.e., if it satisfies the four linear constraints in eq. (13), then the following identity is determined:
where the first, second, third, and fourth rows of eq. (13) are invoked above in lines one, two, three, and four, respectively. This equality justifies the redefinition in eq. (14): noting that the primal objective function in eq. (13) is (r+1)θ, and (since the program is self-dual) the associated dual objective function is −(r+1)θ, note that the gap between primal and dual objective functions, divided by the number of conic constraints (2r+2), is exactly equal to θ.
The central path for the augmented SOCP in eq. (13) is defined by the feasibility conditions for the SOCP combined with the relaxed complementarity conditions x∘s=ve and κτ=v. Thus, the point (x=e; y=0; τ=1; θ=1; s=e; κ=1) is not only a feasible point for the SOCP in eq. (13), but also a central point with v=1.
Finally, a noteworthy property of the self-dual SOCP in eq. (13) is that the optimal point for the original SOCP in eq. (5) can be derived from the optimal point for the SOCP in eq. (13). Specifically, let (xsd*; ysd*; τ*; θ*; ssd*; κ*) be the optimal point for eq. (13) (it can be shown that θ*=0). Then if
is an optimal primal-dual point for eqs. (5) and (6). If τ*=0, then at least one of the original primal SOCP in eq. (5) and the original dual SOCP in eq. (6) must be infeasible. As previously demonstrated, the specific SOCP for portfolio optimization in eq. (10) is primal and dual feasible, so τ*≠0 for that example.
What if there is only a point that is approximately optimal for the self-dual SOCP? An approximately optimal point for the original SOCP can still be determined. Suppose a feasible point for which μ (x, τ, s, κ)=∈. The point (x/τ; y/τ; s/τ) is (∈) close ti feasible for the original SOCP in the sense that the equality constraints are satisfied up to (∈) error
Moreover, since κ>0 and θ=∈, assert using the third row of eq. (13) that the difference in objective function achieved by the primal and dual solutions is also (∈), that is
In summary, by using the self-dual SOCP of eq. (13), a trivial point is obtained from which to start the IPM, and given an (approximately) optimal point, the following is obtained: either an (approximately) optimal point to the original SOCP or a certificate that the original SOCP was not feasible to begin with.
Each iteration of the IPM takes as input an intermediate point (x; y; τ; θ; s; κ) that is feasible (or in some formulations, nearly feasible), has duality gap
equal to μ, and is close to the central path with parameter v=μ. The output of the iteration is a new intermediate point (x+Δx; y+Δy; τ+ΔT; θ+Δθ; s+Δs; κ+Δκ) that is also feasible and close to the central path, with a reduced value of the duality gap. Thus, many iterations lead to a solution with duality gap arbitrarily close to zero.
One additional input is the step size, governed by a parameter σ<1. The IPM iteration aims to bring the next intermediate point onto the central path with parameter v=σμ. This is accomplished by taking one step using Newton's method, where the vector (Δx; Δy; Δτ; Δθ; Δs; Δκ) is uniquely determined by solving a linear system of equations called the Newton system. The first part of the Newton system is the conditions to be met for the new point to be feasible, given in the following system of N+K+2 linear equations:
Note that if the point is already feasible, the right-hand-side is equal to zero.
The second part of the Newton system is the linearized conditions for arriving at the point on the central path with duality gap σμ. That is, aim for (x+Δx)∘(s+Δs)=σμe and (κ+Δκ)(τ+Δτ)=σμ. By ignoring second order terms (i.e., the (Δx∘Δs) and (ΔκΔτ) terms), these become
x∘Δs+s∘Δx=σμe−x∘s
κΔτ+τΔκ=σμ−κτ. (20)
The expression above can be rewritten as a matrix equation by first defining the arrowhead matrix U for a vector u=(u0; ũ)∈Qk as
When u∈ lies in the direct product of multiple second-order cones, the arrowhead matrix is formed by placing the appropriate matrices of the above form on the block diagonal. The arrowhead matrix has the property that for any vector v, Uv=u∘v.
Using this notation, the Newton equations in eq. (20) can be written as
where X and S are the arrowhead matrices for vectors x and s.
Equations (19) and (22) together form the Newton system. It is observered that there are 2N+K+3 constraints to match the 2N+K+3 variables in the vector (Δx; Δy; Δτ; Δθ; Δs; Δκ). As long as the duality gap is positive and (x; y; τ; θ; s; κ) is not too far from the central path (which will be the case as long as a is chosen sufficiently close to 1 in every iteration), the Newton system has a single unique solution. Note that one can choose different search directions than the one that arises from solving the Newton system presented here; this includes first applying a scaling transformation to the product of second-order cones, then forming and solving the Newton system that results, and finally applying the inverse scaling transformation. Alternate search directions are explained in section Additional Information D, but in the main text the basic search direction illustrated above is maintained, since in the numerical simulations the simple search direction gave equal or better results than more complex alternatives, and it enjoys the same theoretical guarantee of convergence.
The Newton system formed by combining eqs. (19) and (22) is an L×L linear system of the form Gu=h, where L=2N+K+3. Classically this can be solved exactly a number of ways, the most straightforward being Gaussian elimination, which scales as (L3). Using Strassen-like tricks, this can be asymptotically accelerated to (Lω) where ω<2.38, although practically the runtime is closer to (L3). Meanwhile, the linear system can be approximately solved using a variety of iterative solvers, such as conjugate gradient descent or the randomized Kaczmarz method. The complexity of these approaches depends on the condition number of the Newton matrix. Section IV discusses quantum approaches to solving the Newton system.
It is noteworthy to distinguish methods that exactly solve the Newton system, and methods that solve it inexactly, because inexact solutions typically lead to infeasible intermediate points. As presented above, the Newton system in eqs. (19) and (22) can tolerate infeasible intermediate points; the main consequence is that the right-hand-side of eq. (19) becomes non-zero. As discussed in section IV, exact feasibility is difficult to maintain in quantum IPMs, since the Newton system cannot be solved exactly.
The inexact-feasible IPM(IF-IPM) is a workaround by which exact feasibility can be maintained despite an inexact linear system solver. For the IF-IPM, this disclosure assumes access to a basis for the null space of the feasibility constraint equations, that is, a linearly independent set of solutions to eq. (19) when the right-hand-side is zero. These basis vectors are arranged as the columns of a matrix B; since there are N+K+2 linear feasibility constraints and 2N+K+3 variables, the matrix B should have N+1 columns. In the case of portfolio optimization, a matrix B satisfying this criterion can be deduced by inspection, as discussed in section Additional Information C; however, this choice does not yield a B with orthogonal columns. Generating a B with orthonormal columns can be done by performing a QR decomposition of the matrix in eq. (19), which would incur a large one-time classical cost of ((N+K)3) operations. Better asymptotic scaling for QR decomposition can be accomplished using fast matrix multiplication. In either case, since B is a basis for the null space of the constraint equations, there is a one-to-one correspondence between vectors Δz∈, and vectors that satisfy eq. (19) via the relation (Δx; Δy; Δτ; Δθ; Δs; Δκ)=BΔz. Thus, the Newton system can be reduced to
The Newton system above can be solved by first computing Δz by inverting the quantity in brackets in the first line and applying it to the right-hand-side, and then computing (Δx; Δy; Δτ; Δθ; Δs; Δκ) by performing the multiplication BΔz. This matrix-vector product can be accomplished classically in (N2) operations. Note that matrix-matrix products where one of the matrices is an arrowhead matrix (S or X) can also be carried out in (N2) classical time, as the form of arrowhead matrices given in eq. (21) implies that the product can be computed by summing several matrix-vector products. Finally, note that since the second and fourth block columns of the first matrix in eq. (22) are zero, the second and fourth block rows of B (e.g., in eq. (C1)) can be completely omitted from the calculation.
Thus, there are three main choices for how to run the IPM when the solution to linear systems is inexact: first, by solving eqs. (19) and (22) directly and allowing intermediate solutions to be infeasible; second, by determining a matrix B by inspection as described in section Additional Information C and then solving eqs. (23) and (24); third, by determining a matrix B via QR decomposition and then solving eqs. (23) and (24). When the linear system is solved using a quantum algorithm, as discussed in section IV, this disclosure refers to the algorithm that results from each of these three options by II-QIPM, IF-QIPM, and IF-QIPM-QR, respectively. The pros and cons of each method are summarized in table II.
Prior literature establishes that if sufficiently small steps are taken (i.e., if σ is sufficiently close to 1), then each intermediate point stays within a small neighborhood of the central path. This disclosure now reviews these conclusions. For a vector u=(u0; ũ)∈, the following matrix is defined:
which, as for the arrowhead matrix, generalizes to the product of multiple cones by forming a block diagonal of matrices of the above form. This disclosure uses the following distance metric
d
F(x,τ,s,κ)=√{square root over (2)}√{square root over (∥Txs−μ(x,τ,s,κ)e∥2+(τκ−μ(x,τ,s,κ))2)}. (26)
The distance metric induces a neighborhood , which includes both feasible and infeasible points, as well as the neighborhood F, which includes only feasible points
(γ)={(x;y;τ;θ;s;κ): dF(x,τ,s,κ)≤γμ(x,τ,s,κ)} (27)
F(γ)=(γ)∩F, (28)
where F denotes the set of feasible points for the self-dual SOCP. Note that the vector Txs can be computed classically in (N) time given access to the entries of x and s. Thus, whether or not a point lies in (γ) can be determined in (N) time.
So long as 0≤γ≤⅓ and (x; y; τ; θ; s; κ)∈F(γ), then the following may be true:
(x+Δx;y+Δy;τ+Δτ;θ+Δθ;s+Δs;κ+Δκ)∈F(Γ), (29)
where
Thus, if Γ≤γ, and assuming the Newton system is solved exactly, every intermediate point will lie in F(γ). This condition is met, for example, if γ= 1/10 and σ=1−(20√{square root over (2)}√{square root over ((r+1))})−1. Since each iteration reduces the duality gap by a factor σ, the duality gap can be reduced to ∈ after roughly only 20√{square root over (2(r+1))} ln(1/∈) iterations. If the Newton system is solved inexactly, but such that feasibility is preserved (e.g., by solving inexactly for Δz and then multiplying by B, as described above), then an error δ on the vector (x; τ; s; κ) can be tolerated, and the resulting vector can still be within the neighborhood at each iteration.
On the other hand, if the Newton system is not solved exactly, then the resulting vector may not be feasible. Thus, the II-QIPM version of the QIPM does not enjoy the theoretical guarantee of convergence in (√{square root over (r)}) iterations that the IF-QIPM and IF-QIPM-QR versions do (see table II). The best guarantees for the II-QIPM would imply convergence only after (r2) iterations. Nevertheless, it is unclear if a small amount of infeasibility makes a substantial difference in practice: multiple versions of the QIPM were simulated and similar overall performance was observed when intermediate solutions were allowed to be infeasible, despite an inferior theoretical guarantee of success. Thus, in sections V and VI, where the QIPM implementation, resource count, and numerical analysis are described, this disclosure focuses on the II-QIPM. This disclosure presents some of the results of the numerical simulations of the IF-QIPM and IF-QIPM-QR results in the section Additional Information.
As discussed in section III, each iteration of an IPM SOCP solver involves forming and solving a linear system of equations that depends on the intermediate point at the current iteration. For classical IPM implementations for SOCP, the linear systems of equations are typically solved exactly; for example, the numerical SOCP solving package ECOS solves linear systems with a sparse LDL (Cholesky) factorization. For arbitrary dense systems, the runtime of solving an L×L system this way is (L3), but by exploiting sparsity the actual runtime in practice could be much faster, by an amount that is hard to assess. Alternatively, it would, in principle, be possible to employ classical iterative approximate linear system solvers such as conjugate gradient descent or the randomized Kaczmarz method. The choice of the linear system solver thereby determines the overall complexity of the IPM SOCP solver. An idea of QIPM is to use a quantum subroutine to solve the linear system of equations. Other steps of QIPMs may be classical and may remain the same as described in section III. As a quantum linear system solver (QLSS) does not solve the exact same mathematical problem as classical linear system solvers and, moreover, a QLSS uses coherent (quantum) access to the classical data as given by the entries of the relevant matrices, there are various additional tools (discussed herein) that allow embedded QLSS subroutines as a step of IPM SOCP solvers.
First, this disclosure discusses in section IV B the input and output model of QLSSs and present the complexity of state-of-the-art QLSSs. Then, section IV C provides constructions based on quantum random access memory (QRAM) to load classical data as input into a QLSS and discusses the complexity overhead arising from that step. Subsequently, section IV D presents so-called pure state quantum tomography that allows to convert the output of the QLSS into an estimate of the classical solution vector of the linear system of equations. Section IV E puts all the steps together and states the overall classical and quantum complexities of using QLSSs as a subroutine in IPM SOCP solvers. The idea is to compare these costs to the complexities of classical IPM SOCP solvers and point out regimes where quantum methods can potentially scale better that any purely classical methods (e.g., in terms of the SOCP size N, the matrix condition number κ, etc.)
For current purposes, a linear system of equations is given by a real invertible L×L matrix G together with a real vector h=(h1, . . . , hL), and one is looking to give an estimate of the unknown solution vector u=(u1, . . . , uL) defined by Gu=h. The (Frobenius) condition number is defined as
κF(G):=∥G∥F∥G−1∥, (31)
where ∥⋅∥F denotes the Frobenius norm and ∥⋅∥ for a matrix argument denotes the spectral norm.
For this setting, the input to a QLSS is then comprised of: (i) a preparation unitary Uh that creates the :=┌log L┐ qubit quantum state
|h>:=∥h∥−1·Σi=1Lhi|i via|h=Uh|, (32)
where ∥⋅∥ for a vector argument denotes the vector two-norm (standard Euclidean norm), (ii) a block encoding unitary UG in the form
on +G qubits for some G ∈, and (iii) an approximation parameter εQLSP ∈(0,1]. The quantum linear system problem (QLSP) is stated as follows: For a triple (G, h, εQLSP) as above, the goal is to create an -qubit quantum state |{tilde over (v)} such that
defined by Gu=h with u=(u1, . . . , uL), by employing as few times as possible the unitary operators UG, Uh, their inverses UG†, Ut†, controlled versions of UG, Uh, and additional quantum gates on potentially additional ancilla qubits. The state-of-the-art QLSS using the fewest calls to UG, Uh and their variants, is based on ideas from discrete adiabatic evolution. This disclosure notes the following explicit complexities. In this formulation, the quantum state |v corresponds to the normalized solution vector of the normalized linear system Gu=h. Thus, the state |v does not carry information on the norm of the solution ∥u∥. This norm is related to v by the relationship ∥u∥=∥h∥/∥Gv∥.
Proposition 1. The QLSP for (G, h, ε1) can be solved with a quantum algorithm on ┌log2(L)┐+4 qubits for
for some constant C≤15307 using Q≥κF(G) controlled queries to each of UG and UG†, and 2Q queries to each of Uh and Uh†, and constant quantum gate overhead. If G is positive semi-definite, then C≤5632 instead.
Note that a stronger version of above proposition works with the (regular) condition number κ(G):=∥G∥∥G−1∥, but it uses a block-encoding of the form eq. (33) in which the normalization factor is ∥G∥ rather than ∥G∥F. In the current case, the Frobenius version κF(G) is used, since there may not a straightforward method to perform UG with normalization factor ∥G∥F, described in section IV C. It is then sufficient to give upper bounds for the remaining κF(G) to run the algorithm from proposition 1. In practice, such upper bounds are given by using appropriate heuristics (cf. section V on implementations).
Note that proposition 1 implies a solution to the QLSP in eq. (34) with an asymptotic query complexity of (κF/εQLSP) to UG, Uh and their variants and under standard complexity-theoretic assumptions this is optimal in terms of the scaling (κ), but not in terms of the scaling (εQLSP). To get to an improved (log(1/εQLSP)) scaling, an eigenstate filtering method may be used that additionally invokes a quantum singular value transform based on a minimax polynomial. The following overall complexities are provided.
Proposition 2. The QLSP problem for (G, h, ε2) can be solved with a quantum algorithm on ┌log2(L)┐+5 qubits that produces a quantum state
√{square root over (p)}|05|{tilde over (v)}+√{square root over (1−p)}|⊥|fail (36)
with 05|⊥|=0 and success probability p≥½. With that, the sought-after ε2-approximate solution quantum state |{tilde over (v)} can be prepared using Q+d controlled queries to each of UG and UG†, and 2Q+2d queries to each of Uh and Uh†, where
Q=2CκF(G)+(√{square root over (κF(G))}) (37)
d=2κF(G)ln(2/ε2). (38)
Here, C≤15307 is the same constant as in proposition 1.
This version of the algorithm basically uses proposition 1 with constant choice of ε1≤¼, and then uses eigenstate filtering to measure whether the final state is the correct solution state. On average the algorithm is repeated no more than twice to produce the desired state |{tilde over (v)}. The resulting scaling that proposition 2 implies for the QLSP problem in eq. (34) is (κ log(1/εQLSP)), which under standard complexity-theoretic assumptions is optimal in both κ and εQLSP. In practice the Q=2CκF(G) dominates over d and all other terms can be safely neglected for typical settings—even for finite scale analyses. Moreover, the constant C is typically an order of magnitude smaller than the estimates given; for positive semi-definite G the constant is estimated as 638. No direct estimates for general matrices G are available, but this disclosure henceforth assumes C=2000 for the numerical estimates. Additionally, note that for the eigenstate filtering step via QSVT, the minimax polynomial and its corresponding quantum signal processing angles have to be computed. This is done as part of classical pre-processing.
Note that the implementation of the QLSS in each of proposition 1 and proposition 2 assume perfect implementation of the underlying circuits, without additional gate synthesis errors. In practice, however, these circuits will not be implemented perfectly, and hence additional sources of error are later included (e.g., block-encoding error, imperfect rotation gates, etc.) that also contribute to εQLSP. These additional contributions are in section IV D, for example.
The following continues by laying out the additional classical and quantum resources used to employ QLSS for estimating in an end-to-end fashion the classical solution vector v=(v1, . . . , v1) instead of the quantum state |v.
Many quantum algorithms (and in particular for the current use case), use coherent access to classical data for use in the algorithm. Block-encodings of matrices provide a commonly used access model for the classical data by encoding matrices into unitary operators, thereby providing oracular access to the data. As mentioned above, for a matrix G∈, a unitary matrix UG block-encodes G when the top-left block of UG is proportional to G, i.e.
where α≥∥G∥ is a normalization constant, chosen as α=∥G∥F for the use case. The other blocks in UG are irrelevant, but they are encoded such that UG is unitary. For current real matrices G are focused on, but the extension to complex matrices is straightforward. A block-encoding makes use of unitaries that implement (controlled) state preparation, as well as quantum random access memory (QRAM) data structures for loading the classical data. Specifically, QRAM is referred to as the quantum circuit that allows query access to classical data in superposition:
where j is the address in superposition with amplitude ψj and |aj is the classical data loaded into a quantum state. There are several models of QRAM one can use that differ in the way in which the data is loaded. The two most notable QRAM models are the select-swap (SS) model, which is particularly efficient in terms of T-gate utilization, and the bucket-brigade (BB) model, which has reduced susceptibility to errors when operated on potentially faulty hardware.
The block-encoding unitary UG acts on +G qubits, where =┌log2(L)┐ and, in our construction, G=. To build it, UG may be formed as the product of a pair of controlled-state preparation unitaries UL and UR. Specifically,
U
G
=U
R
†
U
L, (41)
U
R:|0>⊗l|j>|ψj|j (42)
U
L:|0>⊗l|k>|k|ϕk (43)
where the -qubit states |ψj and |ϕk are determined from the matrix elements Gjk of G, as follows:
where Gj, denotes the jth row of G. That is, controlled on the second -qubit register in the state |j, UR prepares the -qubit state |ψj into the first -qubit register, and UL performs the same operation for the states |ϕk modulo a swap of the two registers. Both UL and UR utilize an additional QRAM ancilla qubits that begin and end in the state |0. These controlled-state preparation unitaries UR and UL are implemented by combining a QRAM-like data-loading step with a protocol for state preparation of -qubit states. There are several combinations of state preparation procedure and QRAM model one can choose with varying benefits and resource requirements. Reference arXiv:2206.03505 (hereafter “Clader”) studies the resources used to implement these block-encodings and provides explicit circuits for their implementation. This disclosure simply imports the relevant resource estimates from that work in table III. In the current setting, the matrices to block encode are typically dense, which is why the general constructions from Clader are sufficient. For the current purposes, this disclosure works with the minimum depth circuits that achieve a T-gate depth of (log L), at the price of using a total number of (L2) many qubits for the data structure implementing the block encoding unitary UG. Additionally, the -qubit unitary Uh defined by |h=Uh| corresponds to the special case of quantum state preparation and is directly treated by the methods outlined in Clader. The resources used to synthesize Uh up to error εh are also reported in table III.
The minimum-depth block encodings of Clader also incur some classical costs. Specifically, the quoted depth values are only achievable assuming a number of angles have been classically pre-computed and for each angle a gate sequence of single-qubit Clifford and T gates that synthesizes a single-qubit rotation by that angle up to small error. Calculating one of the angles can be done by summing a subset of the entries of G and computing an arcsin. Meanwhile, circuit synthesis uses applying a version of the Solovay-Kitaev algorithm. For the block-encoding procedure, L(L−1) angles and their corresponding gate sequences are computed, which uses a total runtime of L2 poly log(1/εG), although this computation is amenable to parallelization. For the state preparation procedure, L−1 angles and their sequences are used.
This disclosure described how a quantum state |{tilde over (v)} approximating the (real-valued) solution |v of a linear system up to precision εQLSP can be produced. As mentioned in section IV B, in the actual circuit implementation, the approximation error εQLSP accounts for both the inherent error from eigenstate filtering captured in proposition 2 as well as additional gate synthesis error arising from imperfect implementation of block-encoding unitaries and single-qubit rotations. The next step is to approximately read out the amplitudes of |{tilde over (v)} into classical form. To start out, this disclosure proves the following proposition, which communicates how many copies of a quantum state are used to provide a good enough classical description of it, up to a phase on each amplitude.
Proposition 3. Let 0<ε,δ<1 and |ψ=Σj∈[L]αj|j be a quantum state. Then,
ln(2L/δ)<3.1942ε−2 ln(2L/δ) measurements of |ψ in the computational basis suffice to learn an ε∞-norm estimate |{tilde over (α)}| of |α|, with success probability at least 1−δ.
The proof is provided in the section Additional Information B1. Recall that proposition 2 gives a unitary U such that
U|05=√{square root over (p)}|05{tilde over (v)}+√{square root over (1−p)}|⊥|fail (46)
with |{tilde over (v)}:=Σi=1N{tilde over (v)}i|i, <05|⊥=0, and p≥½. The vector {tilde over (v)} may have complex coefficients, but it approximates a real vector v up to some error εQLSP in 2 norm. A goal is to obtain an estimate {tilde over (v)}′=(v1′, . . . , vN′) such that
∥v−{tilde over (v)}′∥≤ξ for an error parameter ξ∈[0,1]. (47)
where ξ captures other (e.g., all) sources of error. Proposition 3 is not quite sufficient because it only gives us an estimate of the absolute value of V. However, the following procedure is sufficient:
which by applying a Hadamard can be mapped to
Here |⊥′ is an arbitrary state orthogonal to |05and |fail′ and |fail″ are arbitrary unnormalized states. The quantities √{square root over (pi′)} are (possibly complex) amplitudes that satisfy |√{square root over (pi′)}−√{square root over (pi)}|εtsp for all i; they arise because the state Σi=1L√{square root over (pi′)}|i can only be prepared up to some error. Next, measure this state in the computational basis, denoting the measurement count of the result 06i as ki+ and the result 05 1i as ki−.
Output the estimate
Proposition 4. Suppose that ∥{tilde over (v)}−v∥≤εQLSP and that v is a real-valued vector. Let ε and εtsp be constants that satisfy ε+√{square root over (2L)}εtsp+√{square root over (2)}εQLSP≤½. Then the algorithm above outputs an estimate {tilde over (v)}′ such that ∥{tilde over (v)}′−v∥<ε+1.58√{square root over (L)}εtsp+1.58εQLSP with probability 1−δ.
The proof is provided in the section Additional Information B1. The statement is used to bound the total error parameter ξ by the quantity ε+1.58√{square root over (L)}εtsp+1.58εQLSP. Proposition 4, together with proposition 2, produces with high probability an (ε) good estimate {tilde over (v)}′ of v by using (Lln(L)/ε2) many samples. If a goal is to resolve the initial linear system Gu=h, then the vector {tilde over (v)}′ produced as in Section IV D as an estimate for the normalized vector v=u/∥u∥, gives an estimate for u via
for which the following is found:
There are other methods in the literature that allow to perform pure state quantum tomography with comparable query complexities, but this disclosure favors the above method because of its computational simplicity, and the fact that it does not require solving any potentially costly additional optimization problems. The sample complexity has been improved to (Lln(L)/ε2), which comes at the cost of more complicated quantum circuits and higher constant overheads (See Reference arXiv:2206.03505). It would be interesting to work out the more involved finite complexity of this result, and further comment on the potential impact of this are provided in section VII.
Putting things together, the steps of our QLSS for given real L×L matrix G and real vector h of size L may be:
The QLSS can then be used for each iteration of an IPM SOCP solver, which involves forming and solving a linear system of equations, resulting in the QIPM SOCP solver. This disclosure provides the quantum circuits used to implement the solver in the section IV D.
The following are the quantum circuits used for the QLSS of proposition 1. The QLSS includes applying a unitary U[s] for many different values of s, where U[s] is a block-encoding of a certain Hamiltonian related to G and h, as specified below. The unitary acts on 4++G total qubits, where the final G qubits are ancillas associated with UG. The four single-qubit registers are referred to with labels a1, a2, a3, a4, the -qubit register with label L, and the G-qubit register with label G. These labels are used as subscripts on bras, kets, and operators to clarify the register to which they apply. The circuit for U[s] is depicted in
and where IL denotes the identity operation on subsystem L, and the four rows and columns correspond to the sectors with qubits a4, a1 set to (0,0), (0,1), (1,0), (1,1).
where H denotes the single-qubit Hadamard gate, and R(s) is given by
The normalization factor of R(s) above combines with a factor of 1/√{square root over (2)} introduced by the Hadamard gate to give an overall normalization factor for (s) of
c(s)=(2((1−f(s))2+f(s)2))−1/2∈[2−1/2,1] (60)
and scheduling function ƒ(s) with ƒ(0)=0 and ƒ(1)=1. Note the self-inverse property U[s]2=1 ∀s ∈[0,1]. The overall quantum circuit U for the quantum algorithm of proposition 1 is then given as:
with the walk operator
P[s]:=WU[s],
where W is the operator that acts as identity on registers a4a1L (which host the Hamiltonian [s]) while performing the reflection (2|00−) on the remaining qubits. The unitary U makes Q controlled queries to each of UG and UG†, and 2Q queries to each of Uh and Uh†, and it has constant quantum gate overhead.
Next, give the remaining QSVT eigenstate filtering quantum circuit for the refined quantum linear system solver of proposition 2. The null space of c(1)·[1] is of interest, which has ground-state energy equal to zero and spectral gap at least c(1)κF−1(G)=(√{square root over (2)}κF)−1. As such, employ the Chebyshev minimax polynomial:
where Tl(⋅) is l-th Chebyshev polynomial of the first kind, as part of the corresponding QSVT quantum circuit. Rl has even degree d equal to
d:=2l=2┌κF(G)ln(2/εqsp))for some εqsp∈(0,1] (63)
where εqsp is the precision to which Rl approximates the optimal filter operator. The QSP subscript stands for “quantum signal processing.”
The circuit for the eigenstate filtering step is depicted in
A quantum tomography routine may also include performing controlled versions of the above circuits as described in eq. (49) and illustrated in
Any QSVT circuit can be made controlled by simply controlling the application of the z rotation gates, since the rest of the circuit contains only symmetric applications of unitary gates and their inverses. Thus, a controlled version of
Controlling the linear system portion is not enough to implement eq. (49). This may (e.g., must) be followed up with a controlled state-preparation routine, controlled on the value of the qubit c being in the |1 state. The resource analysis for controlled state-preparation is reported in Ref. Clader. The resource counts are reported here in table IV.
The previous section reviewed the ingredients used to implement the QIPM, namely, QLSS, block-encoding, and tomography. Here, combine those ingredients to describe how the QIPM is actually implemented, making several observations that go beyond prior literature. Also perform a full resource analysis of the entire protocol and report resources used to run the algorithm.
A QIPM is formed from an IPM by performing the step of solving a linear system with a quantum algorithm; the rest of the steps are classical. Example Algorithm 1 presents pseudocode for the interior point method where the single quantum subroutine—approximately solving a linear system—appears in blue text. The input to Algorithm 1 is an SOCP instance with N variables, K linear constraints, and r second-order cone constraints, along with a tolerance parameter ∈. Here note that K=(N) in the case of the formulation of the PO problem simulated in section VI. The output of the QIPM is a vector x that is (∈) close to feasible, and (∈) close to optimal.
A description of the QIPM algorithm is provided herein along with the following new observations:
This expression is chosen such that the duality gap of the new point is (e.g., exactly) a factor of σ smaller than the old point, up to deviations that are second order in the step length. Note that if the old point is feasible and the solution to the linear system is exact, the second and higher order contributions vanish anyway.
The pseudocode in Algorithm 1 illustrates the “infeasible” version of the example algorithm (II-QIPM from table II). The numbers on the left refer to steps of Algorithm 1. Text in Algorithm 1 bookended by “/*” and “*/” are comments on one or more of the steps. To implement the feasible versions (IF-QIPM and IF-QIPM-QR), minor modifications are made to reflect the process described in section III.
The QIPM described in the pseudocode takes 20√{square root over (2)}√{square root over (r)}ln(∈−1) iterations to reduce the duality gap to ∈, where r is the number of second-order cone constraints. In the case of the portfolio optimization problem, r=3n+1, where n is the number of stocks in the portfolio. Choosing the constant pre-factor to be 20√{square root over (2)} allows us to utilize theoretical guarantees of convergence (modulo the issue of infeasibility discussed in section III C 5); however, it would not be surprising if additional optimization of the parameters or heuristic changes to the implementation of the algorithm (e.g. adaptive step size during each iteration) would lead to constant-factor speedups in the number of iterations. Since the number of iterations would be the same for both the quantum and classical IPM, these sorts of improvements would not impact the performance of the QIPM relative to its classical counterpart.
1. Quantum Circuit Compilation and Resource Estimate for Quantum Circuits Appearing within QIPM
The QIPM includes repeatedly performing a quantum circuit associated with the QLSS and measuring in the computational basis. The costs of each of these individual quantum circuits is accounted for herein. There are two kinds of circuits that are used: first, the circuit that creates the output of the QLSS subroutine, given by the state in eq. (36), and second, the circuit that creates the state used to determine the signs of the amplitudes during the tomography subroutine corresponding to a controlled-QLSS subroutine, given in eq. (49).
To simplify the analysis, first compile the circuits from the previous section into a primitive gateset that includes Toffoli gates (and multi-controlled versions of them), rotation gates, block-encoding unitaries, state-preparation and controlled state-preparation unitaries. This compilation allows combining previous in-depth resource analysis for these primitive routines (See Ref. Clader) with the additional circuits shown here.
From left to right in the U[s] circuit shown in
With these decompositions in place, table V reports the resources used to perform each of the two kinds of quantum circuits involved in the QIPM (which are each performed many times over the course of the whole algorithm). The resource quantities are reported in terms of the number of calls Q to the block-encoding (which scales linearly with the condition number), as well as the controlled-block-encoding and state-preparation resources given previously in tables III and IV. The expressions also depend on various error parameters which must be specified to obtain a concrete numerical value.
In section VI, after observing empirical scaling of certain algorithmic parameters, error parameters are described to arrive at a concrete number for a specific problem size.
The resource estimates described above capture the quantum resources used for a single coherent quantum circuit that appears during the algorithm. The output of this quantum circuit is a quantum state, but the QIPM includes a classical estimate of the amplitudes of this quantum state. This classical estimate is produced through tomography, as described in section IV D, by performing k=57.5 L ln(6L/δ)/(ε2(1−ε2/4)) repetitions each of the QLSS and controlled-QLSS circuits, where ε is the desired tomography precision and δ is the probability that the tomography succeeds. In the example implementation given in Algorithm 1, fix δ=0.1. Thus, to estimate the quantum resources of a single iteration of the QIPM, the previous resource estimates reported in table V should each be multiplied by k. With P processors large enough to prepare the output of the QLSS, these k copies may be prepared in k/P parallel steps, saving a factor of P in the runtime at the expense of a factor of P additional space. The resources and scaling estimates do not account for any parallelization, and completely serial execution and runtime is assumed.
After multiplication by k, these expressions give the quantum resources used to perform the single quantum line of the QIPM, ApprSolve. This subroutine has both classical input and output and can thus be compared to classical approaches for approximately solving linear systems.
where ƒ(s) given in eq. (59). The CR1(s) gate is identical but with the control bit sign flipped. Note that the Ry(±π/4) gates are Clifford conjugate to a single T or T† gate.
TABLE V shows quantum resources used to create the state output by the quantum linear system solver, given in eq. (36) (QLSS, left) or the state used to compute the signs during the tomography subroutine, given in eq. (49) (Controlled QLSS, right) for a square linear system of size L=. Note that these resource quantities do not yet account for the k classical repetitions used in order to perform tomography on the output state. The parameters Q and d each scale linearly with the condition number of the linear system, as defined in proposition 2. The symbols NQcbe, TDcbe, and TCcbe denote the number of logical qubits, the T-depth, and the T-count, respectively, for performing a controlled-block-encoding, as reported in table III. The symbols TDSp, and TCsp are analogous quantities for state-preparation, as reported in table IV. The parameters εar, εtsp and εz ∈(0,1] are error parameters corresponding to the gate synthesis precision used for the CR0(s) and CR1(s) rotations, controlled state-preparation step used by tomography, and the QSVT phases, respectively.
Recall that the full QIPM algorithm is an iterative algorithm, where each iteration involves approximately solving a linear system by preparing many copies of the same quantum states. The duality gap μ, which measures the proximity of the current interior point to the optimal point, begins at 1 and decreases by a constant factor σ with each iteration. Thus, the number of iterations to reach a final duality gap ∈ is given by:
Recall from the discussion in section III C 3 that the output of the QIPM will achieve an (∈) approximation to the optimal value of the objective function.
Pulling this all together, the resources to perform the full QIPM algorithm can be estimated, including the multiplicative factors used to perform tomography as well as the number of iterations to converge to the optimal solution. Note that the relevant condition number κF(G) and linear-system precision ξ may vary from iteration-to-iteration as the Newton matrix G changes. The overall runtime can be upper bounded using the maximum observed value of κF(G), which is denoted by κF, and minimum observed value of ξ across all iterations. At each iteration, to achieve overall precision ξ, the tomography precision ε is chosen to be just smaller than ξ (e.g., choose ε=0.9ξ), while all other error parameters (εar, εtsp, εz, etc.) are chosen to be small constant fractions of ξ, such that a total error budget of ξ is not exceeded. As the non-tomographic error parameters all appear underneath logarithms, these small constant factors will drop out of a leading order analysis, and it suffices to replace all of these error parameters with ξ.
The overall runtime may then be expressed in terms of κF, ξ, L (the size of the Newton system), and r (the number of second-order cone constraints) up to leading order and including (e.g., all) constant factors, which are reported in table VI. Recall that for the infeasible version of the QIPM acting on the self-dual embedding, L=2N+K+3, where N is the number of SOCP variables and K is the number of linear constraints. Note that in the leading order expression, it is assumed that the contributions proportional to Q=(κF) dominate over terms proportional to Q=(κF) at practical choices of ξ due to the large constant prefactor in the definition of Q (see proposition 2 and surrounding discussion). The left column of table I from the introduction is formed using the expressions in table VI, and substituting the corresponding relations between L and n, where n is the number of stocks in the portfolio optimization problem given in eq. (10). That is, substitute r=3n+1 and L=2N+K+3=8n+3m+6=14n+6 when m=2n is taken, where N is the number of SOCP variables, K is the number of SOCP constraints, n is the number of stocks, and m is the number of time epochs used to create the matrix M as described in section II.
TABLE VI shows leading order contribution to the logical qubit count, T-depth, and T-count for the entire QIPM, including constant factors. The parameter L denotes the size of the Newton linear system and r denotes the number of second-order cone constraints, while ∈ denotes the final duality gap that determines when the algorithm is terminated. For the infeasible QIPM running on an n-asset instance of portfolio optimization, as given in eq. (10), L=14n+6 and r=3n+1; these substitutions yield the results in table I. The parameter κF denotes the maximum observed Frobenius condition number and ξ denotes the minimum observed tomographic precision parameter across all iterations.
The resource expressions in table VI include constant factors but leave parameters κF and ξ unspecified. These parameters depend on the specific SOCP being solved. As a final step, numerical simulations of small PO problems can be used to study the size of these parameters for different PO problem sizes. This information enables us to give concrete estimates for the resources used to solve realistic PO problems with our implementation of the QIPM and sheds light on whether there could be an asymptotic quantum advantage.
The numerical experiments simulate the entirety of Algorithm 1. The (e.g., only) quantum part of the algorithm is to carry out the subroutine ApprSolve (G, h, ξ). The quantum algorithm for this subroutine can be simulated by solving the linear system exactly using a classical solver and then adding noise to the resulting estimated values to simulate the output of tomography. Since the tomography scheme illustrated in section IV D repeatedly prepares the same state and draws k samples from measurements in the computational basis, the result is a sample from the multinomial distribution. In the numerical simulations herein, samples from this same multinomial distribution are drawn, thus capturing tomographic noise in a more precise way than by simply adding uniform Gaussian noise, as was done in conventional work. For simplicity, this disclosure assumes that the part of the tomography protocol that calculates the signs of each amplitude correctly computes each sign. To numerically estimate resource counts, it is helpful to understand what level of precision ξ is used to stay close enough to the central path throughout the algorithm, as well as how large the Frobenius condition number κF of the Newton system is. Noteworthy, it may be desirable to know how these quantities scale with system size and duality gap μ, which decreases by a constant factor with each iteration of the QIPM.
Section III C 5 discussed three formulations of the QIPM (see table II). The first (II-QIPM) is closely related to the original formulation, which does not guarantee that the intermediate points generated by the IPM are feasible. The other two are instantiations of the inexact-feasible formulation, which uses pre-computing a basis for the null-space of the SOCP constraint matrix. The first of these computes a valid basis by hand (IF-QIPM), while the second uses a QR decomposition to determine the basis (IF-QIPM-QR). All three versions were simulated, and it was determined that the II-QIPM stayed close to the central path, despite the lack of a theoretical guarantee that this would be the case. Here the results of the II-QIPM are presented. For comparison, in the section Additional Information E, some numerical results are presented for the feasible QIPMs, which do benefit from a theoretical convergence guarantee, but have other drawbacks.
As discussed in section V A, a simple preconditioner is implemented that reduces the condition number by about at least an order of magnitude with negligible additional classical cost. Resources estimates are reported assuming a preconditioned matrix.
The infeasible QIPM acting on the corresponding SOCP in eq. (10) was also simulated. The figure illustrates how the simulation successfully follows the central path to the optimal solution after many iterations. The duality gap decreases with each step, and, notably, the infeasibility and distance to the central path also decrease (exponentially) with iteration. Also plotted is the tomography precision that was used to ensure that each iteration stayed sufficiently close to the central path (determined adaptively as described in the pseudocode in Algorithm 1). The plot exemplifies how, despite the lack of theoretical convergence guarantees, our simulations suggest that in practice the II-QIPM acting on the PO SOCP will yield valid solutions.
Remarkably, for this instance, observe that both the Frobenius condition number κF and the inverse tomography precision ξ−1 initially increase but ultimately plateau with the iteration number, even as the duality gap gets arbitrarily small (see
To understand the problem scaling with portfolio size, example problem instances are generated by randomly sampling n stocks from the DWCF, using returns over m=2n time epochs (days) to construct our SOCP as in eq. (10). Parameters q, ζ,
To help understand the asymptotic scaling of the quantum algorithm it may be helpful is to determine how the condition number scales as a function of the number of assets, as the runtime of the QLSS algorithm grows linearly with the condition number.
While the depth of the individual quantum circuits that compose the QIPM scales only with the Frobenius condition number, the QIPM also includes a number of repetitions of this circuit for tomography that scales as 1/ξ2, the inverse of the tomography precision squared. To see how this scales with problem size, an analysis for ξ−2 is performed similar to the analysis performed for κF. These results are presented in
(n0.60±0.02)
(n0.94±0.04)
(n0.92±0.04)
(n0.91±0.05)
(n−0.19±0.05)
(n1.10±0.06)
(n0.79±0.11)
(n1.16±0.10)
Above fits for κF and ξ−2 as a function of n on the range n ∈[10, 120] are provided. Here the quantity n1.5κF/ξ2 is studied, which determines the asymptotic scaling of the runtime of the QIPM.
Ultimately, it is not essential to pin down the asymptotic scaling of the algorithm, because a determination of this work is that, even if a slight asymptotic polynomial speedup exists, the size of the constant prefactors involved in the algorithm preclude an actual practical speedup, barring significant improvements to multiple aspects of the algorithm. The next subsection elaborates on this point in a more quantitative fashion.
TABLE IX shows exponent parameter estimates from the fits to the line generated by plotting n1.5κF/ξ2 in
(n2.01±0.05)
(n3.56±0.07)
(n3.36±0.14)
(n3.75±0.12)
Rather than examine algorithmic scaling, actual resource counts for the QIPM applied to PO are computed. It is these resource counts that may matter most from a practical perspective. The total circuit size is estimated in terms of the number of qubits, T-depth, and T-count for a portfolio of 100 assets. This size is chosed because it is small enough that the entire quantum algorithm can be simulated classically. However, at this size, solving the PO problem is not classically hard; generally speaking, the PO problem becomes challenging to solve with classical methods when n is on the order of 103 to 104. A similar concrete calculation can be performed at larger n by extrapolating trends observed in the numerical simulations herein, but there is not confidence that the fits on n∈[10, 120] reported above are reliable predictors for larger n.
Recall that the only step in the QIPM performed by a quantum computer is the task of producing a classical estimate to the solution of a linear system to error ξ. The complexity of this task as it is performed within the QIPM depends on ξ as well as the Frobenius condition number κF. The first step of the calculation is to fix values for ξ and κF at n=100. They are choosen by taking the median over the 128 samples in the numerical simulation at duality gap μ=10−7.
Once κF and ξ are fixed, concrete values for the various other error parameters can be determined that appear in the algorithm such that overall error ξ can be achieved. Tomography dominates the complexity and overall error, but there are a number of other factors that contribute to the error in the final solution. The sources of error are enumerated and labeled here for completeness:
Section IV described a quantum circuit that prepares a state |{tilde over (v)} (after postselection) for which ∥|{tilde over (v)}−|v∥≤εQLSP. If the block-encoding unitaries, state-preparation unitaries, and single-qubit rotations were perfect, then the only contribution to εQLSP would be from eigenstate filtering and we may have εQLSP≤εqsp. Note the relationship d=2κF ln(2/εqsp) from proposition 2. Since the block-encoding unitary UG, the state-preparation unitary Uh, and the single-qubit rotations are implemented imperfectly, there is additional error. In preparing the state, the unitary UG is called 2Q+2d times and the unitary Uh is called 4Q+4d times, where Q is given in proposition 2. Additionally, there are 2Q combined appearances of CR0(s) and CR1(s) gates, where each appearance includes two single-qubit rotations. Note that the appearances of CR0(s) and CR1(s) within the eigenstate filtering portion of the circuit do not contribute to the error because at s=1 these gates can be implemented exactly. Additionally, there are another d single-qubit rotations used to implement the eigenstate filtering step. Since operator norm errors add sublinearly, one can thus say that
εQLSP≤εqsp+(2Q+2d)εG+(4Q+4d)εh+4Qεar+2dεz. (66)
Now, the result of proposition 4 implies that, in order to assert that the classical estimate {tilde over (v)}′ output by tomography satisfies ∥{tilde over (v)}′−v∥≤ξ, it suffices to have
ξ≥ε+1.58√{square root over (L)}εtsp+1.58[εqsp+(2Q+2d)εG+(4Q+4d)εh+4Qεar+dεz], (67)
where for convenience recall the definitions (ignoring the (√{square root over (κF)}) term) of Q and d as
Q=2CκF (68)
d=2κF ln(2/εqsp) (69)
Recalling that the dominant term in the complexity of the algorithm scales as ε−2 but logarithmically in the other error parameters, to minimize the complexity assign the majority of the error budget to ε: let ε=0.9ξ, and split the remaining 0.1ξ across the remaining six terms of eq. (67). There is room for optimizing this error budget allocation, but the savings would be at most a small constant factor in the overall complexity.
Note that elsewhere in the draft, ξ has been referred to as “tomography precision” since ε will dominate the contribution to ξ. Here, the resource calculation uses differentiating ε from ξ, but when speaking conceptually about the algorithm, one can focus on ξ as it is the more fundamental parameter: it represents the precision at which the classical-input-classical-output linear system problem is solved, allowing apples-to-apples comparisons between classical and quantum approaches.
With values for κF, εG, εh, εqsp, εz, and εtsp now fixed, the resource count can be completed using the expressions in table V. Note that for gate synthesis error, the following formula is used Ry=3 log2(1/εr), where Ry is the number of T gates used to achieve an εr-precise Clifford+T gate decomposition of the rotation gate. Putting this all together yields the resource estimates for a single run of the (uncontrolled) quantum linear system solver in table X at n=100. We report these estimates both in terms of primitive block-encoding and state-preparation resources, as well as the raw numerical estimates. For the total runtime, the resources used for the controlled state-preparation routine may also be estimated. These quantities are estimated, but to the precision of the reported estimates, the numbers are the same as the controlled version, so they are excluded for brevity.
To estimate total runtime, our estimates are multiplied by the tomography factor k (for controlled and for uncontrolled) as well as the number of iterations Nit=┌ ln(∈)/ln(σ)┐, where ∈ is the target duality gap (which is taken to be ∈=10−7), and σ=1.0−1/(20 √{square root over (2r)}). While k will vary from iteration to iteration, in the calculation it is assumed that the total number of repetitions is given by the simple product (2k)Nit, which, noting that the value of ξ plateaus after a certain number of iterations, will give a roughly accurate estimate. Note that these 2kNit repetitions need not be done coherently, in the sense that the entire system is measured and reprepared in between each repetition. One can bound the tomography factor k to be k≤57.5 L ln(L)/ξ2, where ξ is determined empirically. However, our numerical simulations of the algorithm yield an associated value of k used to generate the estimate to precision ξ, so this numerically determined value can be used directly. The observed median value of k=3.3×108 from simulation is multiple orders of magnitude smaller than the theoretical bound. Using this substitution for k and Nit, the results are shown in the right column of table I in the introduction.
To aid in understanding which portions of the algorithm dominate the complexity, a breakdown of the resources is shown in
TABLE X shows estimated number of logical qubits NQ, T-depth TD, and T-count TC used to perform the quantum linear system solver (QLSS) subroutine within the QIPM running on a PO instance with n=100 stocks. This calculation uses the empirically observed median value for the condition number at duality gap μ=10−7, which was κF=1.6×104. The full QIPM repeats this circuit k=(nln(n)ξ−2) times in each iteration to generate a classical estimate of the output of the QLSS, and also performs Nit=(n0.5) iterations, where the linear system being solved changes from iteration to iteration. In the left column, the resources are written as numerical prefactors times the resources used to perform the controlled-block-encoding of the matrix G (denoted by a subscript cbe), and the state-preparation of the vector |h (denoted by a subscript sp), defined in tables III and IV. Written this way, one can see the large prefactors occurring from the linear system solver portion of the algorithm. In the right column the exact resources are computed, including those coming from the block-encoding. The notation AeB is short for A×10B.
The resource quantities reported are large, even for the classically easy problem size of n=100 assets in the portfolio optimization instance. Our detailed analysis allows us to see exactly how this large number arises, which is helpful for understanding how to improve it. Several independent factors leading to the large resource estimates are outlined below.
Remarkably, the five factors described above contribute roughly equally to the overall T-depth calculation; the exception being the number of copies used to do tomography, which is a much larger number than the others. Another comment regarding tomography is that, in principle, the k tomographic samples can be taken in parallel rather than in series. Running in parallel leads to a huge overhead in memory: one can reduce the tomographic depth by a multiplicative factor P at the cost of a multiplicative factor P additional qubits. Note that even preparing a single copy uses the large number of nearly ten million logical qubits at n=100. Moreover, it is unlikely that improvements to tomography alone may make the algorithm practical, as the other four factors still contribute roughly 1016 to the T-depth.
Besides the rather large constant factors pointed out above for tomography and especially for the QLSS, note that the multiplicative “log factors” that are typically hidden underneath notation in asymptotic do tomography, which is a much larger number than the analyses contribute meaningfully here. For instance, the entire block-encoding depth is (log(n/εG)), which, in practice, is as large as 1000. Moreover, there is an additional ln(∈−1)≈16 coming from the iteration count, and a ln(L)≈7 from tomography.
This quantitative analysis of bottlenecks for QIPMs can inform likely bottlenecks in other applications where QLSS, tomography, and QRAM subroutines are used. While some parameters such as κF and ξ are specific to the application considered here, other observations such as the numerical size of various constant and logarithmic factors (e.g., block-encoding depth) would apply more generally in other situations.
The bottlenecks above focused mainly on the T-depth and did not take into account the total T-count or the number of logical qubits, which are also large. Indeed, the estimate of 8 million logical qubits, as reported in table I, is drastically larger than estimates for other quantum algorithms, such as Shor's algorithm and algorithms for quantum chemistry, both of which can be on the order of 103 logical qubits. By contrast, the current generation of quantum processors have tens to hundreds of physical qubits, and no logical qubits; a long way from the resources used for this QIPM.
However, note that, as for other algorithms using repeated access to classical data, the vast majority of the gates and qubits in the QIPM arise in the block-encoding circuits, which are themselves dominated by QRAM-like data-loading subcircuits. These QRAM-like sub-circuits have several special features. Firstly, they are largely composed of controlled-swap gates, each of which can be decomposed into four T gates that can even be performed in a single layer, given one additional ancilla and classical feed-forward capability. Furthermore, in some cases, the ancilla qubits can be “dirty”, i.e., initialized to any quantum state, and, if designed correctly, the QRAM circuits can possess a natural noise resilience that may reduce the resources used for error correction. Implementing these circuits with full-blown universal and fault-tolerant hardware could be unnecessary given their special structure. Just as classical computers have dedicated hardware for RAM, quantum computers may have dedicated hardware optimized for performing the QRAM operation. Preliminary work on hardware based QRAM data structures (as opposed to QRAM implemented via quantum circuits acting on logical qubits) shows promise in this direction.
Our estimates suggest that the size of the QRAM used to solve an n=100 instance of PO is one megabyte, and the QRAM size for n=104 (i.e., sufficiently large to potentially be challenging by classical standards) is roughly 10 gigabytes, which is comparable to the size of classical RAM one might use on a modern laptop. These numbers could perhaps be reduced by exploiting the structure of the Newton matrix, as certain blocks are repeated multiple times in the matrix, and many of the entries are zero (see eqs. (10) and (19)). Exploiting the sparsity of the matrix can lead to reduced logical qubit count and T-count, but not reduced T-depth. In fact, it may lead to non-negligible increases in the T-depth, since the shallowest block-encoding constructions can be hyper-optimized for low-depth, and are explicitly not compatible with exploiting sparsity.
With this in mind, one can ask the following hypothetical question: suppose access to a sufficiently large dedicated QRAM element in our quantum computer, and furthermore that the QRAM ran at a 4 GHz clock speed (which is comparable to modern classical RAM); would the algorithm become more practical in this case? Under the conservative simplifying assumption that each block-encoding and state-preparation unitary uses just a single call to QRAM and the rest of the gates are free, a rough answer can be given by referring to the expression in table X, which states that 4×108 total block-encoding and state-preparation queries are used. Thus, even if the rest of the estimates stay the same, the number of QRAM calls involved in just a single QLSS circuit for n=100 would be 4×108. Accounting for the fact that the QIPM involves an estimated 6×1012 repetitions of similarly sized circuits, the overall number of QRAM calls used to solve the PO problem would be larger than 1021, and the total evaluation time would be on the order of ten thousand years. Thus, even at 4 GHz speed for the QRAM, the problem remains decidedly intractable. Nonetheless, if the QIPM is made practical, it may involve specialized QRAM hardware in combination with improvements to the algorithm itself. Separate from the QLSS, a relatively small number of state preparation queries is used in tomography to create the state in eq. (50), but this number does not scale with K and it is neglected in this back-of-the-envelope analysis.
The discussion above suggests that the current outlook for practicality with a QIPM is pessimistic, but simultaneously highlights several avenues by which to improve the results. Even with such improvements, if QIPMs are to one day be practical, they may need to at least have an asymptotic speedup over CIPMs. Here we comment on this possibility. A step of both QIPMs and CIPMs is the problem of computing a classical estimate of the solution to a linear system, a task that is also of broad use beyond interior point methods. Thus, we can compare different approaches to solving linear systems, and our conclusions are relevant in any application where linear systems are solved. Accordingly, in table XI, the asymptotic runtime of several approaches to solving an L×L linear system to precision are given, including the (QLSS+tomography) approach utilized by QIPMs, as well as two classical approaches. Whereas prior literature primarily compared against Gaussian elimination (which scales as (L3)), this disclosure notes a comparison against the randomized Kaczmarz method, which scales as (LκF2 ln(ξ−1)). This scaling comes from the fact that 2κF2 ln(ξ−1) iterations are used, and each iteration involves computing several inner products at cost (L). It is observed that the worst-case cost of an iteration is 4L floating point multiplications, meaning all the constant prefactors involved are more-or-less mild. Thus, the asymptotic quantum advantage of the QIPM is limited to an amount equal to (min(ξ2κF, ξ2L2/κF)), which is at most (L) when κF ∝L and ξ=O(1). Encouragingly, our numerical results are consistent with κF ∝L. However, our results are not consistent with ξ=(1), suggesting instead that ξ is decreasing with L.
TABLE XI shows a comparison of time complexities of different approaches for exactly or approximately solving an L×L linear system with Frobenius condition number κF to precision ξ. The comparison highlights how a quantum advantage only persists when κF is neither too large nor too small. The constant pre-factor roughly captures the T-depth determined for the quantum case (the same pre-factor from Tab. VI after discounting the 20√{square root over (2)} IPM iteration factor) and the number of multiplications in the classical case.
If κF ∝L and ξ=(1), it is determined a total QIPM runtime of (n2.5), improving over classical (n3.5) for a portfolio with n stocks. This speedup is a material asymptotic improvement over the classical complexity, but leveraging this speedup for a practical advantage might still be difficult. Firstly, the difference in the constant prefactor between the quantum and classical algorithms would likely negate the speedup unless n is taken to be very large. Secondly, the speedup would necessarily be sub-quadratic. In the context of combinatorial optimization, where quadratic speedups can be obtained easily via Grover's algorithm, even a quadratic speedup is unlikely to exhibit actual quantum advantage after factoring in slower quantum clock speeds and error-correction overheads.
Our results suggest that determining a practical quantum advantage for portfolio optimization might require structural improvements to the QIPM itself. In particular, it may be helpful to explore whether additional components of the IPM can be quantized, and whether the costly contribution of quantum state tomography could be completely circumvented. Naively, circumventing tomography entirely is challenging, as it is useful (e.g., important) to retrieve a classical estimate of the solution to the linear system at each iteration in order to update the interior point and construct the linear system at the next iteration. Nevertheless, tomography represents a formidable bottleneck.
While our results are pessimistic on the question of whether quantum interior point methods will deliver quantum advantage for portfolio optimization (and other applications), it is our hope that by highlighting the precise issues leading to daunting resource counts, our work can inspire innovations that render quantum algorithms for optimization more practical. Finally, we conclude by noting that detailed, end-to-end resource estimations of the kind performed here may be helpful (e.g., important) for commercial viability of quantum algorithms and quantum applications. While it is helpful to discover and prove asymptotic speedups of quantum algorithms over classical, an asymptotic speedup alone does not imply practicality. For this, a detailed, end-to-end resource estimate is used, as the quantum algorithm may nevertheless be far from practical to implement.
Here we list the symbols that appear in this disclosure for reference.
Symbols related to portfolio optimization
Symbols related to second-order cone programs
Symbols related to second-order cone programs for portfolio optimization
Symbols related to interior point methods (IPMs)
Relations Between Parameters
Symbols related to quantum linear system solvers
Symbols related to block encoding and state preparation
Symbols related to tomography
Proof of proposition 3. Consider a single coordinate αj with associated probability pj=|αj|2, and suppose k samples are taken to determine an estimate {tilde over (p)}j of pj. By Bernstein's inequality,
and so for a given component-wise target deviation in the probability εj, choosing
guarantees that Pr[|{tilde over (p)}j−pj|>εj]≤δ′.
Now pick εj=√{square root over (3γ)}|αj|ε+γε2 for some yet undetermined γ>0. With this choice
and hence it suffices to choose
Letting δ′=δ/L, the union bound implies that for
all estimates {tilde over (p)}j satisfy |{tilde over (p)}j−pj|≤εj. Now bound the distance between |{tilde over (α)}j| and |αj|. First,
Next, bound |αj|−|{tilde over (α)}j|. If pj≤εj then
which follows because the function ƒ(x)=x−√{square root over (x2−√{square root over (3x)}−1)} has its maximum at
Therefore with the choice
we can guarantee that ∥{tilde over (α)}j|−|αj∥≤ε, which corresponds to
measurements.
Proof of proposition 4. Define
Then k 2.875ε′−2 ln(6L/δ). Consider the following three assertions:
and the estimates pi−=ki−/k satisfy
for all i.
From proposition 3, we know that Assertion 1 holds with probability at least 1-δ/3, and Assertion 2 holds with probability at least 1−2δ/3. Therefore both assertions hold with probability at least 1−δ. Moreover, Assertion 3 holds by assumption. From here on it is assumed that all three assertions hold.
Let ai be the real part and bi be the imaginary part of the quantity √{square root over (p)}{tilde over (v)}i. Let ri+=|√{square root over (p)}{tilde over (v)}i+√{square root over (pi)}|, and ri−=|√{square root over (p)}{tilde over (v)}i−√{square root over (pi)}|. Note that ri+ and ri− are proportional to the absolute value of the ideal amplitudes of the state created in eq. (50). One can show that
Define ƒi(x, y)=(x2−yz)/√{square root over (pi)}; then ai=ƒi(ri+/2, ri−/2). Note that the estimates √{square root over (pi±)} give good approximations of ri+/2 and ri−/2:
which follows from Assertions 2 and 3. The amplitudes di that define the estimate output by the tomography algorithm are given in eq. (52), which can now be rewritten as
We prove that the ãi values approximate the ai values, specifically
|ãi−ai|≤ε′+εtsp+|bi|. (B11)
We will prove the claim above using a case-by-case analysis. Assume that ai≥0; the case ai<0 will proceed similarly.
First, consider the case
In this case ãi=0 and
Second, consider the case ƒi(√{square root over (pi+)}, √{square root over (pi−)})≥ai. From the definition of ãi and Assertion 1, we have
and thus
We also have (again invoking Assertion 1)
Additionally, consider the case ƒi(√{square root over (pi+)}, √{square root over (pi−)})<ai. Defining
we can lower bound ƒi(√{square root over (pi+)}, √{square root over (pi−)}):
Here in the second line we used eq. (B9) and the fact that
We now upper bound ri++ri−:
where in the fourth line we used ai≤√{square root over (ai2+bi2)}=√{square root over (p)}|{tilde over (v)}i|≤√{square root over (pi)}+ε′/3 (Assertion 1).
where in the fourth line we used {tilde over (ε)}/√{square root over (pi)}≤1. This implies
Here, we used ai−√{square root over (pi)}≤√{square root over (p)}|{tilde over (v)}i|−√{square root over (pi)}≤ε′/3.
We've shown that |ãi−ai|≤ε′+εtsp+|bi| for all cases. Therefore,
where we used Σi ((vi−ai/√{square root over (p)})2+bi2/p)≤εQLSP2. Since {tilde over (v)}′∝ã, for some proportionality factor λ we have ∥λ{tilde over (v)}′−v∥√{square root over (2L)}(ε′+εtsp)+√{square root over (2)}εQLSP, where we used p≤½. A bit of geometry will show that if
Applying this with c=λ{tilde over (v)}′ and d=v we obtain
as claimed. In the second inequality we used the convexity of g; in the third inequality we used the fact that g(√{square root over (2L)}ε′)=ε, √{square root over (2L)}(ε′+εtsp)+√{square root over (2)}εQLSP<ε+√{square root over (2)}εtsp+E√{square root over (2)}εQLSP≤½, and √{square root over (2)}g′(½)<1.58.
In section III C, an inexact-feasible interior point method was described that uses as input a matrix B with columns that form a basis for the null space of the feasibility equations for the self-dual SOCP that appear in eq. (19). A straightforward way to find such a B in general may be to perform a QR decomposition of the constraint matrix, costing classical O(N3) runtime (or, using techniques for fast matrix multiplication, between O(N2) and O(N3) time). The upshot is that B can only be computed once and does not change with each iteration of the algorithm, but depending on other parameters of the problem, this classical runtime may dominate the overall complexity. Alternatively, in many specific cases including ours, a valid matrix B can be determined by inspection. For example, suppose that we have a (N−K)×N matrix QA with full column rank for which AQA=0, a K×(K−1) matrix P with full column rank for which
The leftmost column in the above block matrix corresponds to K−1 basis vectors formed by choosing y to be a vector perpendicular to
The third column corresponds to the vector formed by choosing x=e, τ=θ=1, and then
The final column corresponds to choosing x=x0, τ=1, θ=0, and
In each case, the choices of x, y, τ, and θ uniquely determine the values of s and κ. Note that in practice the second and fourth block rows of B can be ignored because in eq. (22) they are left-multiplied by a matrix whose second and fourth block columns are zero.
What remains is to specify P, QA, and x0 for the case of portfolio optimization, given in eq. (10). Finding a valid matrix P is straightforward. Note that from eq. (10), we have b=(1;
The solution (Δx; Δy; Δτ; Δθ; Δs; Δκ) to the Newton systems in eqs. (19) and (22) is one possible search direction for the interior point method. Alternative search directions can be found by applying a scale transformation to the convex set. For the k-dimensional second-order cone , define the set
For the product of multiple cones, let the set include of direct sums of entries from . This definition implies that the matrices G∈ map the set onto itself. Thus for a fixed choice G∈, consider a change of variables x′=GTx, s′=G−1s, y′=y. Let X′ and S′ be the arrowhead matrices for x′ and s′, and, following the same logic as above, the following Newton system is arrived at:
The solution to this linear set of equations (along with the feasibility equations of eq. (19)) is distinct for different choices of G. The choice G=I recovers eq. (22) and is called the Alizadeh-Haeberly-Overton (AHO) direction. Note that the IPM can reduce the duality gap by a constant factor after O(√{square root over (r)}) iterations for any choice of G. However, some choices of G can yield additional potentially desirable properties; for example, the Nesterov-Todd search direction scales the cone such that x′=s′. However, in the numerical simulations of the QIPM, no obvious benefits of choosing a search direction were observed other than the AHO direction.
Section VI presented numerical results for the “II-QIPM,” for which intermediate points could be infeasible. Here we also present some results for two variants of the “feasible” QIPM, denoted by “IF-QIPM” and “IF-QIPM-QR,” as summarized in table II. The IF-QIPM uses the null space basis B outlined in Additional Information C, whereas the IF-QIPM-QR version uses a null space basis B determined using a QR decomposition. In all cases, the algorithm was simulated for enough iterations to reduce the duality gap to 10−3, whereas for the II-QIPM it was simulated down to 10−7.
In
TABLE XII sjpws fit parameters for the Frobenius condition number for the four horizontal-axis locations considered on the scaling plot of
TABLE XIII shows fit parameters for the square of the inverse of the required tomography precision to stay near the central path, corresponding to
TABLE XIV shows estimated scaling of the quantum algorithm as a function of portfolio size for the two feasible versions of the quantum algorithm, corresponding to
(n1.41±0.01)
(n2.07±0.15)
(n1.23±0.40)
(n1.77±0.15)
(n2.87±0.91)
(n3.13±0.18)
(n3.54±0.64)
(n3.50±0.10)
The above sections describe example embodiments for purposes of illustration only. Any features that are described as essential, necessessary, required, important, or otherwise implied to be required should be interpreted as only being required for that embodiment and are not necessarily included in other embodiments.
Additionally, the above sections often use the phrase “we” (and other similar phases) to reference an entity that is performing an operation (e.g., a step in an algorithm). These phrases are used for convenience. These phrases may refer to a computing system (e.g., computing system 1700) that is performing the described operations.
In the example of
At step 1610, the computing system receives the SOCP instance (e.g., see input of Algorithm 1). The SOCP instance may include quantities (A, b, c) as described with respect to eq. 10. The computing system may also receive a list of cone sizes (N1, . . . , Nr) and a tolerance ∈ (also referred to as precision ∈). In some embodiments, computing system receives the SOCP instance with N>0 variables, K>0 linear constraints, r>0 second-order cone constraints, and the tolerance parameter E, where matrix A of the SOCP instance is a K×N matrix encoding linear constraints, vector b is a length-K vector encoding right-hand side of linear constraints, and vector c is a length-Nvector encoding an (e.g., objective) function (e.g., see eq. (5)).
At step 1620, the computing system defines a Newton system for the SOCP instance by constructing matrix G and vector h (e.g., steps 4 and 5 of Algorithm 1). Matrix G and vector h describe constrains for a linear system Gu=h based on the SOCP instance. Defining the system in this way enables the benefits of the quantum approach over classical approaches to be realized.
At step 1630, the computing system preconditiones matrix G and vector h via row normalization to reduce a condition number of matrix G (e.g., steps 6-10 of Algorithm 1). Among other advantages, preconditioning matrix G and vector h reduces their computational complexity (due to the reduced condition number), which reduces computational time.
Preconditioning matrix G and vector h may include determining a diagonal matrix D where at least one (e.g., each or every) entry Dii is equal to the norm of row i of matrix G. After determining the the matrix D, matrix G and vector h may be redefined according to: G=D−1G and h=D−1h, where D−1G has a condition number less that the condition number of previous/original matrix G. More information on preconditioning can be found above at, for example, sections I B and V A.
At step 1640, the computing system iteratively determines u until a predetermined iteration condition is met. For example, see steps 13-18 of Algorithm 1, where in this context u is (x′; y′; τ′; θ′; s′; ) and the predetermine iteration condition is (x′; y′; τ′; θ′; s′; ) ∈(γ)). The iterations may include:
The iterations of step 1640 may include causing the quantum computing system to apply matrix G and vector h to a quantum linear system solver (QLSS) to generate a quantum state (e.g., part of step 15 of Algorithm 1). Causing the quantum computing system to apply matrix G and vector h to the QLSS to generate the quantum state may include k>0 applications of the QLSS and k controlled-applications of the QLSS to generate the quantum state. Causing the quantum computing system to apply matrix G and vector h to the QLSS may include causing the quantum computing system to execute a quantum circuit. More information on the QLSS can be found above at, for example, section IV B.
The QLSS may operate on a block encoded version of matrix G and a state-prepared version of vector h. To do this, the method 1600 may include causing matrix G to be block encoded onto the quantum computing system and the vector h to be state encoded onto the quantum computing system. As previously described, block-encodings enable quantum algorithms (the QLSS in this case) to coherently access classical data (G and h in this case). More information on block encoding can be found above at, for example, sections I B and IV C.
The interations of step 1640 may include causing the quantum computing system to perform quantum state tomography on the quantum state (e.g., part of step 15 (e.g., step 27) of Algorithm 1). Causing the quantum computing system to perform quantum state tomography may include causing the quantum computing system to execute a quantum circuit. Conceptually, the quantum state tomography “reads out” the results of calculations “stored” in the quantum state.
The output of the of the quantum state tomography may be a unit vector {tilde over (v)}′ indicating an iteration direction for u (e.g., {tilde over (v)}′ is (Δx; Δy; Δτ; Δθ; Δs; Δ) of Algorithm 1). The unit vector {tilde over (v)}′ may be characterized by ∥{tilde over (v)}′−v∥≤ξ with probability equal to or greater than a predetermined probability threshold (e.g., probability at least 1−δ, where δ is, for example, 0.1), where v∝G−1h and ξ is a tomography precision parameter (e.g., see steps 24 and 28 of Algorithm 1).
In some embodiments, the quantum state tomography is performed according to a tomography precision parameter ξ that decreases with each new iteration (e.g., see ξ in Algorithm 1). The tomography precision parameter ξ may decrease by ξ/2 with each new iteration (see e.g., step 14 of Algorithm 1). More information on quantum tomography can be found above at, for example, sections IV D and V A.
The interations of step 1640 may include updating a value of u based on a current value of u and the output of the quantum state tomography (e.g., steps 16-17 of Algorithm 1). For example, updating the value of u may include determining an updated step length σ≠0 based on the unit vector {tilde over (v)}′, and adding the current value of u to the product of: (1) the updated step length σ and (2) the unit vector {tilde over (v)}′ (e.g., see steps 16-17 of Algorithm 1, where “step length” refers to the updated step length, σ refers to the current step length, the current value of u is (x; y; τ; θ; s; ), and the updated value of u is (x′; y′; τ′; θ′; s′; ). As used herein, a value of a vector (e.g., u or {tilde over (v)}′) may refer to the value of a single component or of multiple components of that vector.
In some embodiments, the updated step length σ is given by by:
where Δx, Δs, Δ, and Δτ are components of the unit vector {tilde over (v)}′; x, s, , and τ are components of current u; μ(x, τ, s, ) is a duality gap; σ is the current step length; and r is the number of second-order cone constraints of the SOCP instance. The updated step length may be determined such that a duality gap μ of the updated value of u is a factor of σ (the current step length) smaller than the current value of u within a second order deviation in the step length σ. The duality gap μ may describe a difference between the current value of u and an exact solution to the linear system Gu=h. More information on step length can be found above at, for example, sections I B, V A, and IV A.
At step 1650, the computing system determines a solution to the SOCP instance based on the updated value of u (e.g., steps 19-21 of Algorithm 1). In the example of Algorithm 1, the solution to the SOCP instance (within precision ∈) is vector x of (x; y; τ; θ; s; )).
In some embodiments, steps 1620-1640 are repeated iteratively (e.g., see loop of steps 3-21 of Algorithm 1). For example, a target precision F for the solution to the SOCP instance is received (referred to as “tolerance” or “precision” in Algorithm 1), and steps 1620-1640 are repeated and a duality gap μ is iteratively updated based on the output of the quantum state tomography (e.g., the new duality gap μ is the product of current duality gap μ and the updated step length σ (e.g., see also step 20 of Algorithm 1) until the duality gap μ is less than the target precision F (e.g., see condition at step 3 of Algorithm 1), where the duality gap μ describes a difference between the current value of u and an exact solution to the linear system Gu=h. In these embodiments, matrix G and vector h may be constructed (in step 1620) further based on the (e.g., updated) value of u (e.g., u is (x; y; τ; θ; s; ) and G and h depend at least on one of these quantities in steps 4 and 5 of Algorithm 1).
Embodiments described above may be implemented using one or more computing systems. Example computing systems are described below.
The classical computing system 1710 may operate or control the quantum computing system 1720. For example, the classical computing system 1710 causes the quantum computing system 1720 to perform one or more operations, such as to execute a quantum algorithm or quantum circuit (e.g., the classical computing system 1710 generates and transmits instructions for the quantum computing system 1720 to execute a quantum algorithm or quantum circuit). For example, the computing system 1710 causes the quantum computing system 1720 to perform one or more steps of method 1600. Although only one classical computing system 1710 is illustrated in
The quantum computing system 1720 exploits the laws of quantum mechanics in order to perform computations. A quantum processing device, a quantum computer, a quantum processor system, and a quantum processing unit (QPU) are each examples of a quantum computing system. The quantum computing system 1700 can be a universal or a non-universal quantum computing system (a universal quantum computing system can execute any possible quantum circuit (subject to the constraint that the circuit doesn't use more qubits than the quantum computing system)). In some embodiments, the quantum computing system 1700 is a gate model quantum computer. As previously described, quantum computing systems use so-called qubits, or quantum bits (e.g., 1750A). While a classical bit has a value of either 0 or 1, a qubit is a quantum mechanical system that can have a value of 0, 1, or a superposition of both values. Example physical implementations of qubits include superconducting qubits, spin qubits, trapped ions, arrays of neutral atoms, and photonic systems (e.g., photons in waveguides). Additionally, the disclosure is not specific to qubits. The disclosure may be generalized to apply to quantum computing systems 1720 whose building blocks are qudits (d-level quantum systems, where d>2) or quantum continuous variables, rather than qubits.
A quantum circuit is an ordered collection of one or more gates. A sub-circuit may refer to a circuit that is a part of a larger circuit. A gate represents a unitary operation performed on one or more qubits. Quantum gates may be described using unitary matrices. The depth of a quantum circuit is the least number of steps used to execute the circuit on a quantum computing system. The depth of a quantum circuit may be smaller than the total number of gates because gates acting on non-overlapping subsets of qubits may be executed in parallel. A layer of a quantum circuit may refer to a step of the circuit, during which multiple gates may be executed in parallel. In some embodiments, a quantum circuit is executed by a quantum computing system. In this sense, a quantum circuit can be thought of as comprising a set of instructions or operations that a quantum computing system can execute. To execute a quantum circuit on a quantum computing system, a user may inform the quantum computing system what circuit is to be executed. A quantum computing system may include both a core quantum device and a classical peripheral/control device (e.g., a qubit controller 1740) that is used to orchestrate the control of the quantum device. It is to this classical control device that the description of a quantum circuit may be sent when one seeks to have a quantum computer execute a circuit.
The parameters of a parameterized quantum circuit may refer to parameters of the gates. For example, a gate that performs a rotation about the y axis may be parameterized by a real number that describes the angle of the rotation.
The description of a quantum circuit to be executed on one or more quantum computing systems may be stored in a non-transitory computer-readable storage medium. The term “computer-readable storage medium” should be taken to include a single medium or multiple media (e.g., a centralized or distributed database, or associated caches and servers) able to store instructions. The term “computer-readable medium” shall also be taken to include any medium that is capable of storing instructions for execution by the quantum computing system and that cause the quantum computing system to perform any one or more of the methodologies disclosed herein. The term “computer-readable medium” includes, but is not limited to, data repositories in the form of solid-state memories, optical media, and magnetic media.
Illustrated in
The storage device 1808 is any non-transitory computer-readable storage medium, such as a hard drive, compact disk read-only memory (CD-ROM), DVD, or a solid-state memory device or other optical storage, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices, optical disk storage devices, flash memory devices, or other non-volatile solid state storage devices. Such a storage device 1808 can also be referred to as persistent memory. The pointing device 1814 may be a mouse, track ball, or other type of pointing device, and is used in combination with the keyboard 1810 to input data into the computer 1800. The graphics adapter 1812 displays images and other information on display 1818. The network adapter 1816 couples the computer 1800 to a local or wide area network.
The memory 1806 holds instructions and data used by the processor 1802. The memory 1806 can be non-persistent memory, examples of which include high-speed random access memory, such as DRAM, SRAM, DDR RAM, ROM, EEPROM, flash memory.
As is known in the art, a computer 1800 can have different or other components than those shown in
As is known in the art, the computer 1800 is adapted to execute computer program modules for providing functionality described herein. As used herein, the term “module” refers to computer program logic utilized to provide the specified functionality. Thus, a module can be implemented in hardware, firmware, or software. In one embodiment, program modules are stored on storage device 1808, loaded into the memory 1806, and executed, individually or together, by one or more processors (e.g., 1802).
Some portions of the above disclosure describe the embodiments in terms of algorithmic processes or operations. These algorithmic descriptions and representations are commonly used by those skilled in the computing arts to convey the substance of their work effectively to others skilled in the art. These operations, while described functionally, computationally, or logically, are understood to be implemented by computer programs comprising instructions for execution, individually or together, by one or more processors, equivalent electrical circuits, microcodes, or the like. Furthermore, it has also proven convenient at times, to refer to these arrangements of functional operations as modules, without loss of generality. In some cases, a module can be implemented in hardware, firmware, or software.
As used herein, any reference to “one embodiment” or “an embodiment” means that a particular element, feature, structure, or characteristic described in connection with the embodiment is included in at least one embodiment. The appearances of the phrase “in one embodiment” in various places in the specification are not necessarily all referring to the same embodiment. Similarly, use of “a” or “an” preceding an element or component is done merely for convenience. This description should be understood to mean that one or more of the elements or components are present unless it is obvious that it is meant otherwise. As used herein, the terms “comprises,” “comprising,” “includes,” “including,” “has,” “having” or any other variation thereof, are intended to cover a non-exclusive inclusion. For example, a process, method, article, or apparatus that comprises a list of elements is not necessarily limited to only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Further, unless expressly stated to the contrary, “or” refers to an inclusive or and not to an exclusive or. For example, a condition A or B is satisfied by any one of the following: A is true (or present) and B is false (or not present), A is false (or not present) and B is true (or present), and both A and B are true (or present).
In addition, use of the “a” or “an” are employed to describe elements and components of the embodiments. This is done merely for convenience and to give a general sense of the disclosure. This description should be read to include one or at least one and the singular also includes the plural unless it is obvious that it is meant otherwise. Where values are described as “approximate” or “substantially” (or their derivatives), such values should be construed as accurate +/−10% unless another meaning is apparent from the context. From example, “approximately ten” should be understood to mean “in a range from nine to eleven.”
Alternative embodiments are implemented in computer hardware, firmware, software, and/or combinations thereof. Implementations can be implemented in a computer program product tangibly embodied in a machine-readable storage device for execution by a programmable processor system including one or more processors that can act individually or together; and method steps can be performed by a programmable processor system executing a program of instructions to perform functions by operating on input data and generating output. Embodiments can be implemented advantageously in one or more computer programs that are executable on a programmable system including one or more programmable processors coupled to receive data and instructions from, and to transmit data and instructions to, a data storage system, at least one input device, and at least one output device. Each computer program can be implemented in a high-level procedural or object-oriented programming language, or in assembly or machine language if desired; and in any case, the language can be a compiled or interpreted language. Suitable processors include, by way of example, both general and special purpose microprocessors. Generally, a processor will receive instructions and data from a read-only memory and/or a random-access memory. Generally, a computer will include one or more mass storage devices for storing data files; such devices include magnetic disks, such as internal hard disks and removable disks; magneto-optical disks; and optical disks. Storage devices suitable for tangibly embodying computer program instructions and data include all forms of non-volatile memory, including by way of example semiconductor memory devices, such as EPROM, EEPROM, and flash memory devices; magnetic disks such as internal hard disks and removable disks; magneto-optical disks; and CD-ROM disks. Any of the foregoing can be supplemented by, or incorporated in, ASICs (application-specific integrated circuits) and other forms of hardware.
Although the above description contains many specifics, these should not be construed as limiting the scope of the disclosure but merely as illustrating different examples. It should be appreciated that the scope of the disclosure includes other embodiments not discussed in detail above. Various other modifications, changes, and variations which will be apparent to those skilled in the art may be made in the arrangement, operation, and details of the methods and apparatuses disclosed herein without departing from the spirit and scope of the disclosure.
This application claims priority to U.S. Provisional Patent Application Ser. No. 63/413,230, “End-to End Analysis for Quantum Interior Point Methods with Improved Block-Encodings,” filed on Oct. 4, 2022, which is incorporated herein by reference in its entirety.
Number | Date | Country | |
---|---|---|---|
63413230 | Oct 2022 | US |