This application claims the benefit of Taiwan application Serial No. 107144359, filed Dec. 10, 2018, the disclosure of which is incorporated by reference herein in its entirety.
The disclosure relates in general to a dynamic prediction model establishment method, an electric device and a user interface.
In various production fields of products, a prediction model is often used to predict process results. The prediction model could be established through the target data. Although the target data has high reality, it takes a large amount of cost to implement the target data. Obtaining the target data with large data volume is unavailable, so a prediction model with high accuracy cannot be established (that is, a prediction model with small bias and high uncertainty).
Alternatively, the prediction model could be established using a large amount of auxiliary data. For example, the auxiliary data could be obtained through the simulation of a simulator established according to the theories of physics or chemistry or could be obtained through the estimation of the historical data of similar products. The acquisition of auxiliary data takes low cost. However, the simulator cannot simulate the realities to 100%, and has the problem of incompleteness. Furthermore, there are differences between similar products and the target product. Therefore, the prediction result still has a certain degree of error (that is, the prediction model has large bias and low uncertainty).
Therefore, it has become a prominent task for the research personnel to provide a prediction model with lower cost and higher prediction accuracy to meet the requirements of the industries.
The disclosure is directed to a dynamic prediction model establishment method, an electric device and a user interface.
According to one embodiment, a dynamic prediction model establishment method is provided. The dynamic prediction model establishment method includes the following steps. An integration model is established by a processing device according to at least one auxiliary data set. The integration model is modified as a dynamic prediction model by the processing device according to a target data set. A sampling point recommendation information is provided by the processing device according to an error degree or an uncertainty degree between the at least one auxiliary data set and the target data set.
According to another embodiment, an electric device is provided. The electric device includes a processing device configured to perform a dynamic prediction model establishment method. The dynamic prediction model establishment method includes the following steps. An integration model is established by a processing device according to at least one auxiliary data set. The integration model is modified as a dynamic prediction model by the processing device according to a target data set. A sampling point recommendation information is provided by the processing device according to an error degree or an uncertainty degree between the at least one auxiliary data set and the target data set.
According to an alternative embodiment, a user interface is provided. The user interface is configured to display at least one auxiliary data set and a target data set, and an error degree and an uncertainty degree of a dynamic prediction model.
The above and other aspects of the disclosure will become better understood with regard to the following detailed description of the preferred but non-limiting embodiment(s). The following description is made with reference to the accompanying drawings.
In the following detailed description, for purposes of explanation, numerous specific details are set forth in order to provide a thorough understanding of the disclosed embodiments. It will be apparent, however, that one or more embodiments may be practiced without these specific details. In other instances, well-known structures and devices are schematically shown in order to simplify the drawing.
In various production fields of products, a prediction model is often used to predict process results. The prediction model could be established through the target data. Although the target data has high reality, it takes a large amount of cost to implement the target data. Obtaining the target data with large data volume is unavailable. The prediction model with small target data volume will have high uncertainty degree.
Alternatively, a prediction model could be established using a large amount of auxiliary ata. For example, the auxiliary data could be obtained through the simulation of a simulator established according to the theories of physics or chemistry or could be obtained through the estimation of the historical data of similar products. The acquisition of auxiliary data takes low cost. However, the simulator cannot simulate the realities to 100%, and has the problem of incompleteness. Furthermore, there are differences between similar products and the target product. Therefore, the prediction model established by the auxiliary data has a large error degree.
In this disclosure, various embodiments of a dynamic prediction model establishment method are disclosed below. The dynamic prediction model establishment method combines at least one auxiliary data set and a target data set via a two-stage model stacking technology and provides suitable sampling strategies, such that both of the error degree and the uncertainty degree can be reduced. More specifically, the auxiliary data set includes at least one input parameter IT received by a target system 300 and at least one output response OT generated by the target system 300. Namely, both of the input parameter IT and the output response OT are disclosed in detail as below.
Referring to
Refer to
Refer to Table 1. Table 1 illustrates examples of the input parameter IT and the output response OT of the target system 300.
Let an MOCVD process be taken for example. When the model or the brand changes, the present disclosure could fully utilize the original information of the MOCVD process. The input parameter IT could be sensing information such as the pressure or temperature of the MOCVD process, and the present disclosure is not limited thereto. The output response OT could be the health status of the MOCVD process, and the present disclosure is not limited thereto.
Let an LED process be taken for example. When product specifications change (for example, the wavelength requirement is adjusted), the prediction performance of the prediction model could be increased with the assistance of the original product information. The input parameter IT could be the flow of various gases required in the LED process, and the present disclosure is not limited thereto. The output response OT could be product quality (such as the film thickness, the magnitude of stress, and so on), and the present disclosure is not limited thereto.
Let a hyperparameter auto-tuning software system be taken for example. When a hyperparameter auto-tuning process is performed to a complicated model, excellent pre-emptive information could be provided to the auto-tuning system according to the results of an adjusted model or a simple strategy. The input parameter IT could be the hyperparameter of the hyperparameter auto-tuning software system, and the present disclosure is not limited thereto. The output response OT could be the prediction accuracy of the hyperparameter auto-tuning software system, and the present disclosure is not limited thereto.
The target data TD is stored in target database 150 to form a target data set TDS.
Then, the method proceeds to step S131, an integration model IM is established by the integration model calculation module 131 according to at least one auxiliary data set. In the descriptions below, two auxiliary data sets ADS1 and ADS2 are used as an exemplification. However, the amount of auxiliary data sets is not limited in the present disclosure.
Refer to
The auxiliary data sets ADS1 and ADS2 could be simulation data sets provided by a simulation software or approximate data sets provided by similar machinery (similar processes, similar systems or historical products), and the present disclosure is not limited thereto. The auxiliary data sets ADS1 and ADS2 include an input parameter IT and an output response OT.
Let an LED process be taken for example. When developing a new product, the simulation data obtained by a physical simulator or a chemical simulator (such as the CVDSim of the STR Group) could provide sufficient information to the model.
Refer to Table 2. Table 2 illustrate examples of the auxiliary data set ADS1.
Refer to Table 3. Table 3 illustrate examples of the auxiliary data set ADS2.
In sub-step S1311, a suitable model could be selected for each of the auxiliary data sets ADS1 and ADS2 according to respective data features of the auxiliary data sets ADS1 and ADS2. For example, the random forest (RF) model, the Gaussian process (GP) model and the logistic regression (LR) model could be used as candidate models, and a suitable model could be selected according to a mean absolute error (MAE) score in a leave-one-out verification method, and the present disclosure is not limited thereto.
Then, the method proceeds to sub-step S1312, respective confidence weights W1 and W2 of the individual models M1 and M2 are calculated by the integration model calculation module 131. As indicated in Table 4, in response to the input parameter IT (that is, −10.0, −7.50, −5.00) inputted to the individual model M1, the target system 300 outputs a predicted output response OT (that is, 6.58, 4.15, 4.15). A mean absolute error E1 (that is, 6.17) is calculated according to the predicted output response OT (that is, 6.58, 4.15, 4.15) and the output response OT (that is, 12.08, 3.13, 14.50) of the target system 300. Since the auxiliary data sets ADS1 and ADS2 are provided by the simulation software or similar systems or products and the target data set TDS is provided by the target system 300, there are errors existing between the auxiliary data sets ADS1 and ADS2 and the target data set TDS. The smaller the errors are, the better the auxiliary data sets ADS1 and ADS2 represent the target data set TDS.
Refer to Table 4. In response to the input parameter IT (that is, −10.0, −7.50, −5.00) inputted to the individual model M2, the target system 300 outputs a predicted output response OT (that is, 6.13, 5.34, 5.34). A mean absolute error E2 (that is, 5.95) is calculated according to the predicted output response OT (that is, 6.13, 5.34, 5.34) and the output response OT (that is, 12.08, 3.13, 14.50).
Refer to equations (1) and (2). The confidence weights W1 and W2 could be calculated according to the mean absolute errors E1 and E2 of the individual models M1 and M2. The confidence weights W1 and W2 are negatively correlated with respective mean absolute errors E1 and E2 of the individual models M1 and M2.
Then, the method proceeds to sub-step S1313, the individual models M1 and M2 are integrated to obtain the integration model IM by the integration model calculation module 131 according to the confidence weights W1 and W2 of the individual models M1 and M2.
Through sub-steps S1311 to S131, the integration model IM could be obtained according to the auxiliary data sets ADS1 and ADS2. The integration model IM integrates the simulation data provided by a simulator or the historical data of similar products, and therefore could provide better prediction result.
Then, the method proceeds to step S132 of
Refer to
A residual model RM could be trained according to the input parameter IT and the residual value RD of the target system 300. In step S1321, the random forest (RF) model, the Gaussian process (GP) model and the logistic regression (LR) model could be used as candidate models, and a suitable residual model RM could be selected by the model performance improvement module 132 according to the MAE score of the LOO verification method.
Then, the method proceeds to step S1322, the integration model IM and the residual model RM are combined by the model performance improvement module 132 to obtain the dynamic prediction model AM.
Through sub-steps S1321 to S1322, the integration model IM obtained from the auxiliary data sets ADS1 and ADS2 is modified as the dynamic prediction model AM according to the target data set TDS. After the residual model RM is trained using the two-stage model stacking technology adopted in the present disclosure, the residual model RM is further combined with the integration model IM, such that the dynamic prediction model AM, having combined the integration model IM and the target data set TDS, could provide better prediction results.
Then, the method proceeds to step S133 of
Referring to
In the first sampling strategy as indicated in sub-step S1331, a sampling point recommendation information SL is provided by the sampling point recommendation module 133 according to the uncertainty degree UC only. In an early stage of the establishment of the dynamic prediction model AM, the target data TD may not be available yet or may have only a tiny amount, so the first sampling strategy could be used. Referring to
In the second sampling strategy as indicated in sub-step S1332, a sampling point recommendation information SL is provided by the sampling point recommendation module 133 according to the error degree ER and the uncertainty degree UC. In a middle stage of the establishment of the dynamic prediction model AM, only a small amount of target data TD is available, so the second sampling strategy could be used. Refer to
UCB=α*ER+β*UC (3)
wherein, α, β are weighting coefficients.
Referring to
In the third sampling strategy as indicated in sub-step S1333, a sampling point recommendation information SL is provided by the sampling point recommendation module 133 according to a sampling point distribution information DC. Referring to
According to the dynamic prediction model establishment method and the electric device 100 disclosed in above embodiments, the integration model IM is established according to the similarity information between the target data set TDS and the auxiliary data sets ADS1 and ADS2, wherein the similarity information is such as mean absolute errors E1 and E2, and the present disclosure is not limited thereto. The target data set TDS and the integration model IM are combined (the residual model RM) to obtain a dynamic prediction model AM with higher accuracy. Besides, under different sampling strategies, sampling regions with high prediction risk could be found according to the error degree ER and/or the uncertainty degree UC, and the sampling point recommendation information SL could be provided.
In comparison to the prediction model established using Borehole, Park91a, Park91b, EF, Alpine, the dynamic prediction model AM of the present disclosure could improve accuracy by more than 39%.
It will be apparent to those skilled in the art that various modifications and variations could be made to the disclosed embodiments. It is intended that the specification and examples be considered as exemplary only, with a true scope of the disclosure being indicated by the following claims and their equivalents.
| Number | Date | Country | Kind |
|---|---|---|---|
| 107144359 | Dec 2018 | TW | national |
| Number | Name | Date | Kind |
|---|---|---|---|
| 4858147 | Conwell | Aug 1989 | A |
| 4874963 | Alspector | Oct 1989 | A |
| 4943931 | Allen | Jul 1990 | A |
| 5004932 | Nejime | Apr 1991 | A |
| 5016204 | Simoudis et al. | May 1991 | A |
| 5056037 | Eberhardt | Oct 1991 | A |
| 5063492 | Yoda et al. | Nov 1991 | A |
| 5142612 | Skeirik | Aug 1992 | A |
| 5161110 | Dorchak | Nov 1992 | A |
| 5349541 | Alexandro, Jr. et al. | Sep 1994 | A |
| 5418710 | Ono et al. | May 1995 | A |
| 7904397 | Fan et al. | Mar 2011 | B2 |
| 8363819 | Zgardovski et al. | Jan 2013 | B2 |
| 8412563 | Taintor et al. | Apr 2013 | B2 |
| 8473415 | Siegel et al. | Jun 2013 | B2 |
| 8738549 | Lingenfelder et al. | May 2014 | B2 |
| 9256838 | Sow et al. | Feb 2016 | B2 |
| 20080154664 | Kuo et al. | Jun 2008 | A1 |
| 20140114840 | Arnold et al. | Apr 2014 | A1 |
| 20170193400 | Bhaskar et al. | Jul 2017 | A1 |
| 20180121814 | Yu | May 2018 | A1 |
| 20190087685 | Wu | Mar 2019 | A1 |
| Number | Date | Country |
|---|---|---|
| 106127366 | Nov 2016 | CN |
| 108021985 | May 2018 | CN |
| 0 432 267 | Jun 1991 | EP |
| H3-14002 | Jan 1991 | JP |
| 2015-95219 | May 2015 | JP |
| 10-1492733 | Feb 2015 | KR |
| 10-1828503 | Mar 2018 | KR |
| 200745952 | Dec 2007 | TW |
| I509550 | Nov 2015 | TW |
| I539298 | Jun 2016 | TW |
| WO 2007038713 | Apr 2007 | WO |
| Entry |
|---|
| Chuang, Yao-Chen, et al. “Transfer learning for efficient meta-modeling of process simulations.” Chemical Engineering Research and Design 138 (2018): 546-553. (Year: 2018). |
| Vehtari, Aki, Andrew Gelman, and Jonah Gabry. “Practical Bayesian model evaluation using leave-one-out cross-validation and WAIC.” Statistics and computing 27.5 (2017): 1413-1432. (Year: 2017). |
| Taiwanese Notice of Allowance and Search Report for Taiwanese Application No. 107144359, dated May 31, 2021. |
| Author Unknown “Transfer learning.”, Wikipedia, 2 Pages. |
| Currin et al. “A Bayesian Approach to the Design and Analysis of Computer Experiments.”, Source Unknown, Sep. 1, 1988, 1 Page. |
| Golovin et al. “Google Vizier: A Service for Black-Box Optimization.”, Google Research, Pittsburgh, PA, USA, 10 Pages, (2017). |
| Morris et al. “Bayesian Design and Analysis of Computer Experiments: Use of Derivatives in Surface Prediction.”, Technometrics, 35(3), 15 Pages (1993). |
| Pan et al. “A Survey on Transfer Learning.”, IEEE, 15 Pages (2009). |
| Park “Turning complex computer codes to data and optimal designs.”, Thesis (Ph.D.) University of Illinois at Urbana-Champaign, 1 Page (1991). |
| Rahnamayan et al. “A novel population initialization method for accelerating evolutionary algorithms.”, Computers and Mathematics With Applications, 53, 1605-1614 (2007). |
| Xiong et al. “Sequential Design and Analysis of High-Accuracy and Low-Accuracy Computer Codes.”, Technometrics, 55(1), 1 Page (2013). |
| Zhou et al. “A sequential multi-fidelity metamodeling approach for data regression.”, Knowledge-Based Systems, 134, 199-212 (2017). |
| Number | Date | Country | |
|---|---|---|---|
| 20200184346 A1 | Jun 2020 | US |