SYSTEMS AND METHODS FOR FORMULATING A PREDICTION MODEL AND FOR USING THE SAME

Information

  • Patent Application
  • 20240061902
  • Publication Number
    20240061902
  • Date Filed
    August 14, 2023
    9 months ago
  • Date Published
    February 22, 2024
    3 months ago
Abstract
Systems and methods for formulating a prediction model. A linear prediction of an expert model is received, wherein given point xi, the linear prediction is gi:=g(xi)=gTxi+g0, the expert model having an expert model feature list. New data (xi,yi)∀i∈[1,N] is received, wherein the expert model feature list is a subset of a new feature list of the new data. The prediction model is formulated as
Description
TECHNICAL FIELD

The present disclosure relates to modeling, and, in particular, to systems and methods for formulating a prediction model and for using the same.


BACKGROUND

Many companies have collected data that reflects the accumulated expertise. Meanwhile, there is a newly accumulated set of data that reflects modern phenomena. Many organizations seek to find good prediction models that incorporate both the accumulated expertise as well as the modern phenomena.


However, the decision maker has an incomplete knowledge on the model parameters and only knows ranges (bounds) of the uncertain parameters lie in. Furthermore, once the prediction model is built, every realization of the uncertain parameter must satisfy certain restrictions imposed by the decision maker. This gives rise to a model with infinitely many constraints. This causes intractability issues of the model and the model cannot be solved using a standard software without further engineering.


SUMMARY

In accordance with a first aspect of the present disclosure, there is provided a computer-implemented method for formulating a prediction model, comprising: one or more processors; a memory storing computer-executable instructions that, when executed by the one or more processors, cause the computing system to: receive a linear prediction of an expert model wherein given point xi, the linear prediction is gi:=g(xi)=gTxi+g0, the expert model having an expert model feature list; receive new data (xi,yi)∀i∈[1,N], wherein the expert model feature list is a subset of a new feature list of the new data; and formulate a prediction model as:









min
w





i
=
1

N



(



f
i

(
w
)

-

y
i


)

2



+


μ

(



f
i

(
w
)

-

g
i


)

2


,




wherein: fx(w)≤c1, ∀x∈X, and fx(w)≤c3, ∀x∈X∩H, and wherein μ is a positive number assigning weight to the linear prediction.


In some or all example embodiments of the first aspect, the computer-executable instructions, when executed by the one or more processors, cause the computing system to formulate the prediction model as a linear prediction model, fi(w), set to fi(w)=wTxi+w0, fx(w)=wTx+w0, with an objective function of: Σi=1N(wTxi+w0−yi)2+μ(wTxi+w0−gi)2, and with the following constraints: type 1: wTx+w0≤c+βTx, ∀x∈X, and type 2: wTx+w0≤c, ∀x∈X∩H.


In some or all example embodiments of the first aspect, the computer-executable instructions, when executed by the one or more processors, cause the computing system to formulate the prediction model as a quadratic prediction model, fi(γ,q,Q), set to fi(γ,q,Q)=γ+2qTxi+(xi)TQxi, fx(γ,q,Q)=γ+2qTx+(x)TQx, with an isometric realization










x
T


Qx

+

2


q
T


x

+
γ

=





[



γ



q
T





q


Q



]

,

[



1



x
T





x



x


x
T





]




=




W
,

Y
x




=




w
˜

,

x
˜







,




wherein {tilde over (w)}=svec(W), {tilde over (x)}=svec(Yx), wherein an objective function can be written as Σi=1N(custom-character{tilde over (w)},{tilde over (x)}icustom-character−yi)2+μ(custom-character{tilde over (w)},{tilde over (x)}icustom-character−gi)2, and wherein the constraints can be written as type 1: custom-character{tilde over (w)},{tilde over (x)}custom-character≤c1Tx, ∀x∈X, and type 2: custom-character{tilde over (w)},{tilde over (x)}custom-character≤c3, ∀x∈X∩H.


In a second aspect of the present disclosure, there is provided a computer-implemented method for formulating a prediction model, comprising: receiving a linear prediction of an expert model wherein given point xi, the linear prediction is gi:=g(xi)=gTxi+g0, the expert model having an expert model feature list; receiving new data (xi,yi)∀i∈[1,N], wherein the expert model feature list is a subset of a new feature list of the new data; and formulating a prediction model as:









min
w





i
=
1

N



(



f
i

(
w
)

-

y
i


)

2



+


μ

(



f
i

(
w
)

-

g
i


)

2


,




wherein: fx(w)≤c1, ∀x∈X, and fx(w)≤c3, ∀x∈X∩H, and wherein μ is a positive number assigning weight to the linear prediction.


In some or all example embodiments of the second aspect, the prediction model is formulated as a linear prediction model, fi(w), set to fi(w)=wTxi+w0, fx(w)=wTx+w0, with an objective function of: Σi=1N(wTxi+w0−yi)2+μ(wTxi+w0−gi)2, and with the following constraints: type 1: wTx+w0≤c+βTx, ∀x∈X, and type 2: wTx+w0≤c, ∀x∈X∩H.


In some or all example embodiments of the second aspect, the prediction model is formulated as a quadratic prediction model, fi(γ,q,Q), set to fi(γ,q,Q)=γ+2qTxi+(xi)TQxi, fx(γ,q,Q)=γ+2qTx+(x)TQx, with an isometric realization










x
T


Qx

+

2


q
T


x

+
γ

=





[



γ



q
T





q


Q



]

,

[



1



x
T





x



x


x
T





]




=




W
,

Y
x




=




w
˜

,

x
˜







,




wherein {tilde over (w)}=svec(W), {tilde over (x)}=svec(Yx), wherein an objective function can be written as Σi=1N(custom-character{tilde over (w)},{tilde over (x)}icustom-character−yi)2+μ(custom-character{tilde over (w)},{tilde over (x)}icustom-character−gi)2, and wherein the constraints can be written as type 1: custom-character{tilde over (w)},{tilde over (x)}custom-character≤c1Tx, ∀x∈X, and type 2: custom-character{tilde over (w)},{tilde over (x)}custom-character≤c3, ∀x∈X∩H.


In a third aspect of the present disclosure, there is provided a non-transitory machine-readable medium having tangibly stored thereon computer-executable instructions for execution by one or more processors, wherein the computer-executable instructions, in response to execution by the one or more processors, cause the one or more processors to: receive a linear prediction of an expert model wherein given point xi, the linear prediction is gi:=g(xi)=gTxi+g0, the expert model having an expert model feature list; receive new data (xi,yi)∀i∈[1,N], wherein the expert model feature list is a subset of a new feature list of the new data; and formulate a prediction model as:









min
w





i
=
1

N



(



f
i

(
w
)

-

y
i


)

2



+


μ

(



f
i

(
w
)

-

g
i


)

2


,




wherein: fx(w)≤c1, ∀x∈X, and fx(w)≤c3, ∀x∈X∩H, and wherein μ is a positive number assigning weight to the linear prediction.


In some or all example embodiments of the third aspect, the computer-executable instructions, when executed by the one or more processors, cause the computing system to formulate the prediction model as a linear prediction model, fi(w), set to fi(w)=wTxi+w0, fx(w)=wTx+w0, with an objective function of: Σi=1N(wTxi+w0−yi)2+μ(wTxi+w0−gi)2, and with the following constraints: type 1: wTx+w0≤c+βTx, ∀x∈X, and type 2: wTx+w0≤c, ∀x∈X∩H.


In some or all example embodiments of the third aspect, the computer-executable instructions, when executed by the one or more processors, cause the computing system to formulate the prediction model as a quadratic prediction model, , fi(γ,q,Q), set to fi(γ,q,Q)=γ+2qTxi+(xi)TQxi, fx(γ,q,Q)=γ+2qTx+(x)TQx, with an isometric realization










x
T


Qx

+

2


q
T


x

+
γ

=





[



γ



q
T





q


Q



]

,

[



1



x
T





x



x


x
T





]




=




W
,

Y
x




=




w
˜

,

x
˜







,




wherein {tilde over (w)}=svec(W), {tilde over (x)}=svec(Yx), wherein an objective function can be written as Σi=1N(custom-character{tilde over (w)},{tilde over (x)}icustom-character−yi)2+μ(custom-character{tilde over (w)},{tilde over (x)}icustom-character−gi)2, and wherein the constraints can be written as type 1: custom-character{tilde over (w)},{tilde over (x)}custom-character≤c1Tx, ∀x∈X, and type 2: custom-character{tilde over (w)},{tilde over (x)}custom-character≤c3, ∀x∈X∩H.


Other aspects and features of the present disclosure will become apparent to those of ordinary skill in the art upon review of the following description of specific implementations of the application in conjunction with the accompanying figures.





BRIEF DESCRIPTION OF THE DRAWINGS


FIG. 1 is a schematic diagram of a general approach to formulating a prediction model in accordance with example embodiments described herein.



FIG. 2 shows a high level overview of a problem form in accordance with example embodiments described herein.



FIG. 3 is a schematic diagram showing various physical and logical components of a computing system for formulating a prediction model and using the same in accordance with example embodiments described herein.





DETAILED DESCRIPTION OF EXAMPLE EMBODIMENTS

The present disclosure is made with reference to the accompanying drawings, in which embodiments are shown. However, many different embodiments may be used, and thus the description should not be construed as limited to the embodiments set forth herein. Rather, these embodiments are provided so that this application will be thorough and complete. Wherever possible, the same reference numbers are used in the drawings and the following description to refer to the same elements, and prime notation is used to indicate similar elements, operations or steps in alternative embodiments. Separate boxes or illustrated separation of functional elements of illustrated systems and devices does not necessarily require physical separation of such functions, as communication between such elements may occur by way of messaging, function calls, shared memory space, and so on, without any such physical separation. As such, functions need not be implemented in physically or logically separated platforms, although such functions are illustrated separately for ease of explanation herein. Different devices may have different designs, such that although some devices implement some functions in fixed function hardware, other devices may implement such functions in a programmable processor with code obtained from a machine-readable medium. Lastly, elements referred to in the singular may be plural and vice versa, except wherein indicated otherwise either explicitly or inherently by context.


Models with two objectives with no constraints can incorporate complicated prediction models of higher orders. However, it is complicated to incorporate hard constraints imposed on the models. Furthermore, these models cannot incorporate uncertainty in the prediction model effectively.


The present disclosure describes example embodiments of methods, systems, and computer-executable media for modeling a problem by formulating a prediction model. In some embodiments, described is a strong tractable prediction model reformulation framework that avoids the intractability issue in modeling the problem. For example, it may represent the infinitely many restraints by using a finitely many number of restraints.



FIG. 1 shows a general approach for formulating a prediction model in accordance with an embodiment. An expert model 20 represents the knowledge of experts. Modern data 24 represents modern phenomena. A conventional prediction model 28 can be formulated from the expert model 20 and the modern data 24. There are, however, infinitely many scenarios or constraints 32 are imposed by a decision maker. The embodiments described herein may reformulate such prediction models using an effective modelling framework 36, after which the problem can be solved using a conventional solver 40.


Modelling techniques involving two objectives often appear in many classes in mathematical programming. They often have the form “min[f(x)+g(x)]”, as shown in FIG. 2. The terms f(x) and g(x) are responsible for the first objective and the second objective, respectively. The goal of formulation is to achieve the middle ground of the two objectives.


An example of such a problem that can be modeled using the approach described herein is the combining of federated learning with a mechanism model. The first objective engages the federated learning. In federated learning, agents work collectively without exchanging the entire data in order to keep the data privacy. This can be modelled using a function f(x). The second objective engages the mechanism model. The mechanism model engages the knowledge collected by local experts. This can be modelled using a function g(x).


Semi-infinite programs (SIP) form a class of mathematical programs that handle problems with an infinitely many constraints and a finite number of variables. An approach to modeling such problems starts with the detection of the intractability of the model where the infinitely many constraints appear as the form of an uncertainty set. Second, the duality theory of mathematical programs is used to transform the intractable model into a solvable model and achieve tractability.


In order to demonstrate the application of this approach, an example of a tractable transformation is provided. Suppose that variables x1 and x2 must satisfy






c
1
x
1
+c
2
x
2≤40, for all c1,c2 in range 0≤c1≤2, 0≤c2≤10   (1)


This is intractable because there are infinitely many choices for c.


Equivalently, equation (1) has representation












max


c
1

,

c
2




{





c
1



x
1


+


c
2



x
2



subject


to


0




c
1


2

,

0


c
2


10


}



40

,




(
2
)







for fixed x1, x2.


Using the duality theory, equation (2) can be reworked as a set of inequalities with new variables v1, v2.





2v1+10v2≤40, v1≥0, v2≥0, v1=x1, v2=x2   (3)


Three inputs are used in modeling the problem: a mechanism (or expert) model, modern data, and constraints on the prediction.


The mechanism model is assumed to be a linear prediction model given by experts for simplicity purposes. Given point xi, the expert's prediction is gi=g(xi)=gTxi+g0.


In modern data, a new set of data can have more features than the traditional mechanism model. Modern data is represented by (xi,yi), where i runs from 1 to N, meaning that there are N data points. The feature list of modern data contains the feature list of the mechanistic model. There is an unknown relationship between xi and yi.


Constraints on the prediction may be imposed by experts with infinitely many parameters in a closed range.


The general formulation of the problem may take the form










solve


min
w





i
=
1

N



(



f
i

(
w
)

-

y
i


)

2



+


μ

(



f
i

(
w
)

-

g
i


)

2





(
4
)







while satisfying (i.e., subject to)






f
x(w)≤c1, ∀x∈X   (5)






f
x(w)≤c3, ∀x∈X∩H   (6)


The quantity μ is a positive number that assigns the weight to the mechanism model.


Two prediction models are discussed with respect to embodiments: a linear prediction model and a quadratic prediction model.


For the linear prediction model, fi(w) is set to






f
i(w)=wTxi+w0, fx(w)=wTx+w0   (7)


The objective function is written as:





Σi=1N(wTxi+w0−yi)2+μ(wTxi+w0−gi)2   (8)


with the following constraints:





Type 1: wTx+w0≤c+βTx, ∀x∈X   (9)





Type 2: wTx+w0≤c, ∀x∈X∩H   (10)


For the quadratic prediction model, fi(γ,q,Q) is set to:






f
i(γ,q,Q)=γ+2qTxi+(xi)TQxi, fx(γ,q,Q)=γ+2qTx+(x)TQx   (11)


The isometric realization:













x
T


Qx

+

2


q
T


x

+
γ

=





[



γ



q
T





q


Q



]

,

[



1



x
T





x



x


x
T





]




=




W
,

Y
x




=




w
˜

,

x
˜







,




(
12
)







where





{tilde over (w)}=svec(W), {tilde over (x)}=svec(Yx)   (13)


This is described herein.


The objective function can be written as:





Σi=1N(custom-character{tilde over (w)},{tilde over (x)}icustom-characteryi)2+μ(custom-character{tilde over (w)},{tilde over (x)}icustom-character−gi)2   (14)


and the constraints as:





Type 1: custom-character{tilde over (w)},{tilde over (x)}custom-character≤c1Tx, ∀x∈X   (15)





Type 2: custom-character{tilde over (w)},{tilde over (x)}custom-character≤c3, ∀x∈X∩H   (16)


Modern computers with standard software cannot accommodate mathematical models with infinitely many constraints. The prediction model reformulation framework handles such difficulties effectively by utilizing duality theory. In particular, it uses the duality theory to convert the problem with infinitely many constraints with a finite number of variables and a finite number of constraints. As a consequence, the prediction model reformulation framework converts the intractable model into a quadratic program that is efficiently solvable in theory and practice.


The prediction model reformulation framework described herein amy be applicable to scenarios when the exact model parameters are unknown but they lie in a certain range. For example, it can be used where a decision maker wishes to build a prediction function where the predicted value must be less than equal to a given value. Moreover, the predicted value generated by the prediction function must be less than equal to a given value for all realization of the uncertain model parameters.


In order to illustrate the approach described herein, it will now be discussed with respect to a real-life example. An owner of an all-you-can-eat (AYCE) sushi restaurant wants to expand the menu with Chinese cuisine. The owner, by his/her expertise, knows the cost of ingredients that each customer consumes. The decision maker needs some reference data to evaluate different Chinese entrees, and there is a set of available data collected by Chinese restaurant community. As the decision maker runs an AYCE restaurant, customers do not consume an exact set amount of ingredients. For example, each diner consumes a different amount of rice, from 0 g to 50 g. Hence, uncertain model parameters appear in this model.


Uncertainty arises in many real-life situations. The proposed approach helps the user adapt to the unpredictable future.


Modern computers cannot accommodate mathematical models with infinitely many constraints. The proposed approach may handle such difficulties effectively.


The linear and quadratic prediction models generated by the prediction model reformulation framework can be solved using any quadratic programming (QP) solvers; i.e., they are efficiently solvable.


Linear Prediction Model

In a first embodiment, the manner by which how the prediction model reformulation framework converts the intractable from into a tractable form for the linear prediction model will now be described.


The data (xi, yi) and the mechanism model gi:=g(xi)=gTxi+g0 are given. They are described previously herein.


The linear prediction, fi(w), is:






f
i(w)=wTxi+w0, fx(w)=wTx+w0   (17)


The term w0 represents the bias of the prediction.


Let A be the matrix such that each row of A is [1;(xi)T], ŷ be a vector of N entries where the i-th element is yi, and ĝ be a vector of N entries where the i-th element is gi. The objective function can be rewritten in the form:















i
=
1

N



(



w
T



x
i


+

w
0

-

y
i


)

2


+


μ

(



w
T



x
i


+

w
0

-

g
i


)

2


=






A
[


w
0

;
w

]

-

y
^




2

+

μ






A
[


w
0

;
w

]

-

g
^




2




,




(
18
)







which can be rewritten as:





(1+μ)[w0;w]TATA[w0;w]+2custom-characterAT(ŷ+μĝ),[w0;w]custom-character+∥ŷ∥2+μ∥ĝ∥2   (19)


The objective function is now in the form of a conventional convex quadratic function.


The Type 1 constraint is written as:






w
T
x+w
0
≤c
1Tx, ∀x∈X   (20)


Here, the decision maker provides the following parameters that suits their own purpose:





β is a vector,   (21)





c1 is a constant (a number), and   (22)





X is a hypercube of the form X={x∈Rn:li≤xi≤ui, ∀i=1, . . . , n} and li<ui   (23)


The following equivalence is obtained; i.e., equation (24) is equivalent to equation (25):













w
T


x

+

w
0





c
1

+


β
T


x



,



x

X






(
24
)








max

x

X



{



w
T


x

-


β
T


x


}





c
1

-

w
0






(
25
)







It is noted that:











max

x

X



{



w
T


x

-


β
T


x


}


=


max
x



{






w
-
β

,
x



:
l


x

u

}

.






(
26
)







By the duality theory,











max
x


{






w
-
β

,
x



:
l


x

u

}


=


min
v



{







u
-
l

,
v



+




l
,

w
-
β




:
v

-
w
+
β


0

,

v

0


}

.






(
27
)







Thus, equation (24) is equivalent to equation (28):






w
0
+
custom-character
u−l,v
custom-character
+
custom-character
l,w−β
custom-character
≤c
1
,v−w+β≥0,v≥0  (28)


The inequalities in equation (28) have the variables w,v, whereas equation (24) has the single variable w. It is noted that an additional variable v was introduced in equation (28) in order to handle the infinitely many model parameters.


Equation (28) is used in order to obtain a tractable model.


The type 2 constraint can be written as:






w
T
x≤c
3
, ∀x∈X∩H   (29)


Here, the decision maker provides the following parameters that suit their own purpose.





c3 is a constant (a number), (30)





X is a hypercube of the form given in constraint type 1, and (31)





H is a halfspace of the form H={x∈Rn:custom-characterd,xcustom-character≤c4}, d≥0   (32)


A similar approach introduced in constraint type 1 above is used. The following equivalence is obtained; i.e., equation (33) is equivalent to equation (34):






w
T
x+w
0
≤c
4
, ∀x∈X∩H   (33)











max

x


X

H




{


w
T


x

}





c
4

-

w
0






(
34
)







By the duality theory, it is observed that:











max

x


X

H




{


w
T


x

}


=



max
x


{




w
T


x
:



d
,
x






c
4


;

l

x

u


}


=


min
v



{







u
-
l

,
v



+

λ

(


c
4

-


d
T


l


)

+




l
,
w






:
[

I


d

]

[

v
;
λ

]




w

,

v

0

,

λ

0


}

.







(
35
)







Thus, equation (33) is equivalent to equation (36):






w
0
+
custom-character
u−l,v
custom-character
(c4−dTl)+custom-characterl,wcustom-character≤c3, [I d][v;λ]≥w, v≥0, λ≥0   (36)


The inequalities in equation (36) have the variables w, v, λ, whereas equation (33) has the single variable w. It is noted that the additional variable v was introduced in order to handle the parameters in X. The additional variable λ is introduced to capture the parameters associated with H.


Equation (36) is used in order to obtain a tractable model.


The final tractable reformulated model is as follows:














min


w
0

,
w
,

v
1

,

v
2



(

1
+
μ

)

[


w
0

;
w

]

T



A
T



A
[


w
0

;
w

]


+

2






A
T

(


y
^

+

μ


g
^



)

,

[


w
0

,
w

]





+




y
^



2

+

μ





g
^



2






(
37
)







subject to:






w
0
+
custom-character
u−1,v
1
custom-character
+
custom-character
l,w−β
custom-character
≤c
1   (38)






v
1
−w+β≥0   (39)






v
1≥0   (40)






w
0
+
custom-character
u−l,v
2
custom-character+λ(c4−dTl)+custom-characterl,wcustom-character≤c3   (41)





[I d][v2;λ]≥w,   (42)






v
2≥0, and (43)





λ≥0   (44)


The reformulated model is now a conventional convex quadratic program and a standard QP solver can be used to find a solution. Of note is that, for each constraint, an individual variable v must be introduced. In the above model, v1,v2 are responsible for the individual v that appear in the derivation of the constraints. Once the model is solved, set the prediction function as





prediction(x)=wTx+w0   (45)


Quadratic Prediction Model

In another embodiment, it is elaborated how the prediction model reformulation framework converts the intractable form into a tractable form for the quadratic prediction model.


The data (xi,yi) and the mechanism model gi:=g(xi)=gTxi+g0 is given. They are described herein above and they are the same data used in the previously described linear equation embodiment.


The quadratic prediction can be written as:






f
i(γ,q,Q)=γ+2qTxi+(xi)TQxi, fx(γ,q,Q)=γ+2qTx+(x)TQx   (46)


The term γ is responsible for the bias of the prediction.


The general conversion scheme used for the quadratic prediction model is the same as the one in the linear model. However, before adopting the theory described with reference to the linear equation embodiment above, a sophisticated transformation is performed. That is, a mathematical conversion is performed that allows for the interpretation of the quadratic form as a linear form.


First, a function, svec(.), is defined. The term Sn is used to denote the vector space of n-by-n symmetric matrices. Then,





svec:Sn→Rn(n+1)/2 by svec(M)=[diagonal of M; √2 strict upper triangular part of M].   (47)


The svec(.) operation appears in numerous literature reference, such as “On the Kronecker Product”, Kathrin Schäcke, Aug. 1, 2013. The svec(.) mapping is often used to carry the minimum number of variables in mathematical operations. By using svec(.), the inner product in Sn as the inner product in Rn(n+1)/2 can be realized as follows:






custom-character
M
1
,M
2
custom-character=custom-charactersvec(M1), svec(M2)custom-character=custom-character{tilde over (m)}1,{tilde over (m)}2custom-character, where   (48)






M
1
,M
2
∈S
n   (49)






{tilde over (m)}
1=svec(M1), and   (50)






{tilde over (m)}
2=svec(M2)   (51)


Additionally, a function is defined that discards the first entry of svec(.):












svec
0

:

S
n





R



n

(

n
+
1

)

2

-
1




by




svec
0

(
M
)



=



svec

(
M
)

[

2
:


n

(

n
+
1

)

2


]

.





(
52
)







Let à be the matrix that each row of à is [1,(svec(xi))T], ŷ be a vector of N entries where i-th element is yi, and ĝ be a vector of N entries where i-th element is gi. Then the objective function is of the form













i
=
1

N



(





svec

(

W
~

)

,

svec

(

Y

x
i


)




-

y
i


)

2


+


μ

(





svec

(

W
~

)

,

svec

(

Y

x
i


)




-

g
i


)

2





(
53
)










where













W
~

=

[



γ



q
T





q


Q



]


,





(
54
)










w
~

=

svec

(

W
~

)


,
and





(
55
)









Y

x
i


=


[



1




(

x
i

)

T






x
i






x
i

(

x
i

)

T




]

.






(
56
)







Then the objective function can be rewritten as





(1+μ)svec({tilde over (W)})TÃTÃsvec({tilde over (W)})+2custom-characterÃT(ŷ+μĝ),svec({tilde over (W)})custom-character+∥ŷ∥2+μ∥ĝ∥2   (57)


The objective function is a conventional quadratic function.


The type 1 constraint is in the form:





γ+2qTx+(x)TQx≤c1Tx, ∀x∈X   (58)


Here, the decision maker must provide the following parameters that suits their own purpose: β, c1, and X. They are identical to the ones described herein above.


The svec(.) operation is applied:











γ
+

2


q
T


x

+


x
T


Qx


=





[



γ



q
T





q


Q



]

,

[



1



x
T





x



xx
T




]




=




svec

(

W
~

)

,

svec

(

Y
x

)






,


where



Y
x


=


[



1



x
T





x



xx
T




]

.






(
59
)







Then the type 1 constraint has the following representation:






custom-charactersvec({tilde over (W)}),svec(Yx)custom-character+c1Tx, ∀x∈X   (60)


We represent equation (60) as











max



svec

-
1


(
y
)



conv

(

{



Y
x

:
x


X

}

)




{




svec

(

W
~

)

T


y

-




svec

(

[



0



0.5

β
T







0.5
β



0



]

)

,
y




}





c
1

.





(
61
)







The two sets Y, {tilde over (Y)}, the lifted cube and the set after the isoemtric realization are defined, respectively, as follows:











Y
_

=

{



Y



S

n
+
1


:



:

i
,
j





Y

i
,
j




U

i
,
j




,


i

,
j

}


,


where


L

=

[



1



l
T





l



ll
T




]


,

U
=

[



1



u
T





u



uu
T




]


,




(
62.1
)








Y
~

=

{

y




R




(

n
+
1

)



(

n
+
2

)


2

-
1


:

l
~



y


u
~



}


,


where



l
~


=


svec
0

(
L
)


,


u
~

=



svec
0

(
U
)

.






(
62.2
)







We note the containment relation:





conv({Yx:x∈X})⊆Y,   (62.3)


where cony denotes the convex hull. Since the geometry of conv({Yx:x∈X}) is complex, the set conv({Yx:x∈X}) is replaced with Y. Equation (61) can be reformulated using this approximation as











max



svec

-
1


(
y
)



Y
~




{




svec

(

W
~

)

T


y

-




svec

(

[



0



0.5

β
T







0.5
β



0



]

)

,
y




}





c
1

.





(
63
)







It is noted that equation (63) is a linear program. Thus, the same operation performed herein above can be applied. Then, equation (63) can be replace with the inequalities listed in equation (64) below:













w
~

0

+




l
~

,

w
~




+





u
~

-

l
~


,

s
~








c
1

-




l
~

,


svec
0

(

[



0



0.5

β
T







0.5
β



0



]

)






,




(
64
)









-

w
~


+

s
~

-


svec
0



(

[



0



0.5

β
T







0.5
β



0



]

)




0

,
and




(
65
)







s
~


0.




(
66
)







In equations (64) to (66), the variable dimensions are:












w
~

0


R

,


w
~



R




(

n
+
1

)



(

n
+
2

)


2

-
1



,


s
~




R
+




(

n
+
1

)



(

n
+
2

)


2

-
1


.






(
67
)







The type 2 constraint is written as:





γ+2qTx+xTQx≤c3, ∀x∈X∩H   (68)


The decision maker provides the following parameters c3,X,H (described above) that suit the decision maker's own purpose.


For the type 2 constraint for the quadratic prediction model, an approximation of the set intersection X∩H may be performed. This is due to the NP-hardness of the general nonconvex quadratic programs. In approximating the intersection xnH, a decision by the decision maker has to be involved.


A scheme is provided for approximating the set X∩H using two hypercubes. That is, the desire is to obtain two sets:






B
1
⊆X∩H⊆B
2,   (69)


where B1, B2 are hypercubes.


A hypercube B2 that contains X∩H may be obtained as below:






B
2
={x∈R
n
:l≤x≤ū}  (70)


where:












l
_

i

=


min
x


{


x
i


:


x


X



H



}



,


i

,




(
71
)















u
_

i

=


max
x


{


x
i


:


x


X



H



}



,



i
.






(
72
)







It is now desired to obtain a hypercube B1 that is contained in X∩H:






custom-character
d,x
custom-character
≤c
4 is eqivalent to Σi:di>0dixi23 c4   (73)


Let ĉ be a number that holds:






ĉ=Σ
i:d

i

>0
c
i,   (74)


for some ci's. For each i such that di>0, a halfspace













e
i

,
x






c
i


d
i






(
75
)







is generated, where ei is the i-th column of the identity matrix. Then, by setting B1 as below,











B
1

=

X




{


x
:


x
i




c
i


d
i




,



for


i


such


that



d
i


>
0


}



,




(
76
)







B1 that is contained in X∩H is obtained.


B2 is chosen for an approximation and replaces X∩H. A similar approach introduced for constraint type 1 and the type 2 constraint above.


Let











l
^

=


svec
0

(

[



1




l
_



T







l
_





l
_




l
_



T






]

)


,

and




(
77
)












û
=



svec
0

(

[



1




u
¯

T






u
¯





u
¯




u
¯

T





]

)

.





(
78
)







The inequalities below are used to replace the type 2 constraint for the quadratic prediction model.






{tilde over (w)}
0
+
custom-character
{circumflex over (l)},{tilde over (w)}
custom-character
+
custom-character
û−{circumflex over (l)},{tilde over (s)}
custom-character
≤c
3   (79)





{tilde over (w)}+{tilde over (s)}≥0, and   (80)






{tilde over (s)}≥0   (81)


The variable dimensions are












w
˜

0


R

,




(
82
)














w
˜



R




(

n
+
1

)



(

n
+
2

)


2

-
1



,

and




(
83
)













s
˜





R
+





(

n
+
1

)



(

n
+
2

)


2

-
1


.





(
84
)







The final tractable reformulated model is as follows:














min




w
˜

0





,
w

˜

0


,


s
~

1

,


s
~

2



(

1
+
μ

)

[



w
˜

0

;

w
˜


]

T



Ã
T



Ã
[



w
˜

0

;

w
˜


]


+

2






Ã
T

(


y
ˆ

+

μ


g
ˆ



)

,

[



w
˜

0

;

w
˜


]





+




y
ˆ



2

+

μ





g
ˆ



2






(
85
)







subject to













w
˜

0

+




l
~

,

w
~




+





u
~

-

l
~


,


s
˜

1








c
1

-




l
~

,


svec
0

(

[



0




0
.
5



β
T








0
.
5


β



0



]

)






,




(
86
)















-

w
˜


+


s
˜

1

-


svec
0

(

[



0




0
.
5



β
T








0
.
5


β



0



]

)



0

,



s
˜

1


0

,




(
86
)
















w
˜

0

+




l
~

,

w
~




+





u
~

-

l
~


,


s
˜

2







c
3


,




(
87
)















-

w
˜


+


s
˜

2



0

,

and




(
88
)














s
˜

2


0.




(
89
)







Of note is that for each constraint, individual variable {tilde over (s)} must be introduced, as discussed herein above. Once the model is solved, the original form is recovered by taking the inverse of the svec(.) mapping:





W*=svec−1([{tilde over (w)}0;{tilde over (w)}])   (90)


The following assignment is made:











[



γ



q
T





q


Q



]

=

W
*


,




(
91
)







and the prediction function is set as





prediction(x)=γ+2qTx+(x)TQx   (92)


An alternative approximation for the set conv({Yx:x∈X}) is now provided. It is noted that the containment relation is:





conv({Yx:x∈X})⊆Y∩S+n+1Y  (93)


Equation (63) in the type 1 constraint can be replaced by











max


s

v

e



c

-
1


(
y
)





Y
¯





S
+

n
+
1






{




svec

(

W
~

)

T


y

-




svec

(

[



0




0
.
5



β
T








0
.
5


β



0



]

)

,

y




}





c
1

.





(
94
)







That is, Y is replaced by Y∩S+n+1. The duality theory can be employed to reformulate this model. This approximation yields the reformulated model to lie in the class of the semidefinite programs. There are semidefinite programming solvers available.


The approaches described above may provide an efficient and effective reformulation framework for addressing problems with infinitely many model parameters that are not conventionally solvable via standard software. The prediction made by every realized data point from the uncertain range satisfies the required upper bound. The disclosed prediction model reformulation framework may convert an intractable model to a tractable model. The tractable model does not require a specific solver. Any QP solver available in the market can be used.


Two models, linear and quadratic, are proposed herein above under the assumption that the lower and upper bounds of the uncertain bounds do not agree; i.e., li<ui. If there are features that hold li=ui, higher order models may be applied to in conjunction with the model proposed. That is, for the features that hold li<ui, the proposed method may also be applied, and for the features that hold li=ui, any arbitrary prediction function can be applied. The logistic function is an example and appears in many disciplines of machine learning.



FIG. 3 shows various physical and logical components of an exemplary computing system 100 for formulating a prediction model and for using a formulated prediction model in accordance with an embodiment of the present disclosure. Although an example embodiment of the computing system 100 is shown and discussed below, other embodiments may be used to implement examples disclosed herein, which may include components different from those shown. Although FIG. 3 shows a single instance of each component of the computing system 100, there may be multiple instances of each component shown.


The computing system 100 includes one or more processors 104, such as a central processing unit, a microprocessor, an application-specific integrated circuit (ASIC), a field-programmable gate array (FPGA), a dedicated logic circuitry, a tensor processing unit, a neural processing unit, a dedicated artificial intelligence processing unit, or combinations thereof. The one or more processors 104 may collectively be referred to as a processor 104. The computing system 100 may include a display 108 for outputting data and/or information in some applications, but may not in some other applications.


The computing system 100 includes one or more memories 112 (collectively referred to as “memory 112”), which may include a volatile or non-volatile memory (e.g., a flash memory, a random access memory (RAM), and/or a read-only memory (ROM)). The non-transitory memory 112 may store machine-executable instructions for execution by the processor 104. A set of machine-executable instructions 116 defining a prediction model formulation process (described herein) and a process for using the formulated prediction model is shown stored in the memory 112, which may be executed by the processor 104 to perform the steps of the methods for reformulating a prediction model and for using the same described herein. The memory 112 may include other machine-executable instructions for execution by the processor 104, such as machine-executable instructions for implementing an operating system and other applications or functions.


The memory 112 stores mechanistic model data 120 and modern data 124 as described herein.


The memory 108 may also store other data, information, rules, policies, and machine-executable instructions described herein.


In some examples, the computing system 100 may also include one or more electronic storage units (not shown), such as a solid state drive, a hard disk drive, a magnetic disk drive and/or an optical disk drive. In some examples, one or more datasets and/or modules may be provided by an external memory (e.g., an external drive in wired or wireless communication with the computing system 100) or may be provided by a transitory or non-transitory computer-executable medium. Examples of non-transitory computer readable media include a RAM, a ROM, an erasable programmable ROM (EPROM), an electrically erasable programmable ROM (EEPROM), a flash memory, a CD-ROM, or other portable memory storage. The storage units and/or external memory may be used in conjunction with memory 112 to implement data storage, retrieval, and caching functions of the computing system 100.


The components of the computing system 100 may communicate with each other via a bus, for example. In some embodiments, the computing system 100 is a distributed computing system and may include multiple computing devices in communication with each other over a network, as well as optionally one or more additional components. The various operations described herein may be performed by different computing devices of a distributed system in some embodiments. In some embodiments, the computing system 100 is a virtual machine provided by a cloud computing platform.


Although the components for both formulating a prediction model and for using the same are shown as part of the computing system 100, it will be understood that separate computing devices can be used for formulating a prediction model and for using the same.


General

Through the descriptions of the preceding embodiments, the present invention may be implemented by using hardware only, or by using software and a necessary universal hardware platform, or by a combination of hardware and software. The coding of software for carrying out the above-described methods described is within the scope of a person of ordinary skill in the art having regard to the present disclosure. Based on such understandings, the technical solution of the present invention may be embodied in the form of a software product. The software product may be stored in a non-volatile or non-transitory storage medium, which can be an optical storage medium, flash drive or hard disk. The software product includes a number of instructions that enable a computing device (personal computer, server, or network device) to execute the methods provided in the embodiments of the present disclosure.


All values and sub-ranges within disclosed ranges are also disclosed. Also, although the systems, devices and processes disclosed and shown herein may comprise a specific plurality of elements, the systems, devices and assemblies may be modified to comprise additional or fewer of such elements. Although several example embodiments are described herein, modifications, adaptations, and other implementations are possible. For example, substitutions, additions, or modifications may be made to the elements illustrated in the drawings, and the example methods described herein may be modified by substituting, reordering, or adding steps to the disclosed methods.


Features from one or more of the above-described embodiments may be selected to create alternate embodiments comprised of a sub-combination of features which may not be explicitly described above. In addition, features from one or more of the above-described embodiments may be selected and combined to create alternate embodiments comprised of a combination of features which may not be explicitly described above. Features suitable for such combinations and sub-combinations would be readily apparent to persons skilled in the art upon review of the present disclosure as a whole.


In addition, numerous specific details are set forth to provide a thorough understanding of the example embodiments described herein. It will, however, be understood by those of ordinary skill in the art that the example embodiments described herein may be practiced without these specific details. Furthermore, well-known methods, procedures, and elements have not been described in detail so as not to obscure the example embodiments described herein. The subject matter described herein and in the recited claims intends to cover and embrace all suitable changes in technology.


Although the present invention and its advantages have been described in detail, it should be understood that various changes, substitutions and alterations can be made herein without departing from the invention as defined by the appended claims.


The present invention may be embodied in other specific forms without departing from the subject matter of the claims. The described example embodiments are to be considered in all respects as being only illustrative and not restrictive. The present disclosure intends to cover and embrace all suitable changes in technology. The scope of the present disclosure is, therefore, described by the appended claims rather than by the foregoing description. The scope of the claims should not be limited by the embodiments set forth in the examples, but should be given the broadest interpretation consistent with the description as a whole.

Claims
  • 1. A computing system for formulating a prediction model, comprising: one or more processors;a memory storing computer-executable instructions that, when executed 5 by the one or more processors, cause the computing system to: receive a linear prediction of an expert model wherein given point xi, the linear prediction is gi:=g(xi)=gTxi+g0, the expert model having an expert model feature list;receive new data (xi, yi) ∀i∈[1, N], wherein the expert model feature list is a subset of a new feature list of the new data; andformulate a prediction model as:
  • 2. The computing system of claim 1, wherein the computer-executable instructions, when executed by the one or more processors, cause the computing system to formulate the prediction model as a linear prediction model, fi(w), set to: fi(w)=wTxi+w0, fx(w)=wTx+w0
  • 3. The computing system of claim 1, wherein the computer-executable instructions, when executed by the one or more processors, cause the computing system to formulate the prediction model as a quadratic prediction model, fi(γ,q,Q), set to: fi(γ,q,Q)=γ+2qTxi+(xi)TQxi, fx(γ,q,Q)=γ+2qTx+(x)TQx
  • 4. A computer-implemented method for formulating a prediction model, comprising: receiving a linear prediction of an expert model wherein given point xi, the linear prediction is gi:=g(xi)=gTxi+g0, the expert model having an expert model feature list;receiving new data (xi,yi)∀i∈[1,N], wherein the expert model feature list is a subset of a new feature list of the new data; andformulating a prediction model as:
  • 5. The method of claim 4, wherein the prediction model as formulated as a linear prediction model, fi(w), set to: fi(w)=wTxi+w0, fx(w)=wTx+w0
  • 6. The method of claim 4, wherein the prediction model is formulated as a quadratic prediction model, fi(γ,q,Q), set to: fi(γ,q,Q)=γ+2qTxi+(xi)TQxi, fx(γ,q,Q)=γ+2qTx+(x)TQx
  • 7. A non-transitory machine-readable medium having tangibly stored thereon computer-executable instructions for execution by one or more processors, wherein the computer-executable instructions, in response to execution by the one or more processors, cause the one or more processors to: receive a linear prediction of an expert model wherein given point xi, the linear prediction is gi:=g(xi)=gTxi+g0, the expert model having an expert model feature list;receive new data (xi,yi)∀i∈[1,N], wherein the expert model feature list is a subset of a new feature list of the new data; andformulate a prediction model as:
  • 8. The non-transitory machine-readable medium of claim 7, wherein the computer-executable instructions, when executed by the one or more processors, cause the computing system to formulate the prediction model as a linear prediction model, fi(w), set to: fi(w)=wTxi+w0, fx(w)=wTx+w0
  • 9. The non-transitory machine-readable medium of claim 7, wherein the computer-executable instructions, when executed by the one or more processors, cause the computing system to formulate the prediction model as a quadratic prediction model, fi(γ,q,Q), set to: fi(γ,q,Q)=γ+2qTxi+(xi)TQxi, fx(γ,q,Q)=γ+2qTx+(x)TQx
CROSS-REFERENCE TO RELATED APPLICATIONS

This application claims the benefit of U.S. Provisional Application No. 63/397,717 filed on Aug. 12, 2022, the entire contents of which are incorporated herein by reference.

Provisional Applications (1)
Number Date Country
63397717 Aug 2022 US