METHOD AND DEVICE FOR PREDICTING AND CONTROLLING TIME SERIES DATA BASED ON AUTOMATIC LEARNING

Information

  • Patent Application
  • 20230162050
  • Publication Number
    20230162050
  • Date Filed
    December 28, 2021
    2 years ago
  • Date Published
    May 25, 2023
    a year ago
Abstract
A method and device for predicting and controlling time series data based on automatic learning are disclosed. According to an example embodiment, the method of predicting and controlling the time series data based on automatic learning includes training a plurality of time series data prediction models according to conditions respective for the models, determining, among the trained time series data prediction models, one or more optimal models that meet a predetermined condition, and generating a final model by combining the one or more optimal models, wherein the plurality of time series data prediction models includes at least one of statistical-based prediction models and deep learning-based prediction models.
Description
TECHNICAL FIELD

The following description relates to a method and device for predicting and controlling time series data based on automatic learning, and more particularly, to an automatic learning-based artificial intelligence learning and verifying technique that may precisely predict future only with an appropriate amount of training time series data.


BACKGROUND ART

In fields such as finance and manufacturing, people often have to recognize changes in time series and sequential data and make appropriate judgments. For example, a professional investor in a securities company monitors changes in market values such as exchange rates and interest rates and predicts timing and amount of investment, and an operator of factory equipment checks temperature, pressure, and flow rate information and predicts the conditions of the facilities to perform optimal control. However, since analysis of the time series data such as stocks and exchange rates involves complex factors, it is difficult to pinpoint which factors have an effect.


Recent advances in artificial intelligence technology have shown superior prediction performance compared to traditional statistical analysis in forecasting. However, the existing artificial intelligence model has a problem in that the model trained based on data lacks retraining over time, and accordingly the consistency thereof decreases over time. In addition, the existing artificial intelligence model focuses only on the diagnosis of abnormal data, so it is not suitable for automatically learning and providing optimized facility control and investment techniques.


DISCLOSURE OF THE INVENTION
Technical Goals

Example embodiments are not only to perform learning and prediction based on a machine learning model, but also to select an optimal model by automatically learning a deep learning model.


The example embodiments are to provide an automatic learning function for optimally controlling a target variable.


The example embodiments are to provide a description of deep learning model training and a time series deep learning model.


Technical goals of the present disclosure are not limited to what is described in the foregoing, and other technical goals that are not described above may also be clearly understood by those skilled in the art from the following description.


Technical Solutions

According to an aspect, there is provided a method of predicting and controlling time series data based on automatic learning, the method including training a plurality of time series data prediction models according to conditions for the respective models, determining, among the trained time series data prediction models, one or more optimal models that meet a predetermined condition, and generating a final model by combining the one or more optimal models, wherein the plurality of time series data prediction models includes at least one of statistical-based prediction models and deep learning-based prediction models.


The method may further include receiving target variable data for predicting time series data, inputting the target variable data to the final model and outputting target variable prediction data that corresponds to the target variable data.


The method may further include receiving control variable data that determines a direction of a change in the target variable prediction data, inputting the control variable data to the final model and outputting control variable prediction data that corresponds to the control variable data.


The method may further include providing a prediction result and a control method of the time series data based on the target variable prediction data and the control variable prediction data.


The method may further include adjusting the control variable data based on a correlation between the target variable prediction data and the control variable prediction data.


The adjusting of the control variable data may include training a reinforcement learning model according to a reward function that is determined based on the target variable prediction data and the control variable prediction data.


The outputting of the control variable prediction data may include determining a moving direction of the control variable data, and determining an optimal search time for the control variable data.


The outputting of the target variable prediction data may include outputting the target variable prediction data based on the moving direction and the optimal search time for the control variable data.


The training may include training the plurality of time series data prediction models a predetermined number of times according to the conditions for the respective models.


The method may further include evaluating prediction performance of the final model, updating the final model, when the prediction performance of the final model decreases below a predetermined threshold.


The method may further include updating the final model according to a predetermined interval.


According to another aspect, there is provided a device for predicting and controlling time series data based on automatic learning, the device including a processor configured to train a plurality of time series data prediction models according to conditions for the respective models, determine, among the trained time series prediction models, one or more optimal models that meet a predetermined condition, and generate a final model by combining the one or more optimal models, wherein the plurality of time series data prediction models includes at least one of statistical-based prediction models and deep learning-based prediction models.


The processor may be further configured to receive target variable data for predicting time series data, input the target variable data to the final model, and output target variable prediction data that corresponds to the target variable data.


The processor may be further configured to receive control variable data that determines a direction of a change in the target variable prediction data, input the control variable data to the final model, and output control variable prediction data that corresponds to the control variable data.


The processor may be further configured to provide a prediction result and a control method for the time series data based on the target variable prediction data and the control variable prediction data.


The processor may be further configured to adjust the control variable data based on a correlation between the target variable prediction data and the control variable prediction data.


The processor may be further configured to train a reinforcement learning model according to a reward function that is determined based on the target variable prediction data and the control variable prediction data.


The processor may be further configured to train the plurality of time series data prediction models a predetermined number of times according to the conditions for the respective models.


The processor may be further configured to evaluate prediction performance of the final model and update the final model when the prediction performance of the final model decreases below a predetermined threshold.


The processor may be further configured to update the final model according to a predetermined interval.


Advantageous Effects

Example embodiments may not only perform learning and prediction based on a machine learning model but also select an optimal model by automatically learning a deep learning model.


The example embodiments may provide an automatic learning function for optimally controlling a target variable.


The example embodiments may provide a description of deep learning model learning and a time series deep learning model.


Effects of the present disclosure are not limited to what is described in the foregoing, and other effects that are not described above may also be clearly understood by those skilled in the art from the scope of the claims.





BRIEF DESCRIPTION OF DRAWINGS


FIG. 1 is a diagram illustrating a method of predicting and controlling time series data based on automatic learning according to an example embodiment.



FIG. 2 is a diagram illustrating a relationship between a training device and a prediction device according to an example embodiment.



FIG. 3 is a diagram illustrating a method of predicting and controlling time series data based on automatic learning according to an example embodiment.



FIG. 4 is a diagram illustrating a training method according to an example embodiment.



FIGS. 5A and 5B are diagrams illustrating a method of predicting and controlling time series data according to an example embodiment.



FIG. 6 is a block diagram of an artificial intelligence device according to an example embodiment.





BEST MODE FOR CARRYING OUT THE INVENTION

The following structural or functional descriptions are exemplary to merely describe the example embodiments, and the scope of the example embodiments is not limited to the descriptions provided in the present specification.


Although terms of “first” or “second” are used to explain various components, the components are not limited to the terms. These terms should be used only to distinguish one component from another component. For example, a “first” component may be referred to as a “second” component, or similarly, and the “second” component may be referred to as the “first” component within the scope of the right according to the concept of the present disclosure.


It should be noted that if it is described that one component is “connected”, “coupled”, or “joined” to another component, a third component may be “connected”, “coupled”, and “joined” between the first and second components, although the first component may be directly connected, coupled, or joined to the second component. On the contrary, it should be noted that if it is described that one component is “directly connected”, “directly coupled”, or “directly joined” to another component, a third component may be absent. Expressions describing a relationship between components, for example, “between”, directly between”, or “directly neighboring”, etc., should be interpreted to be alike.


The singular forms “a”, “an”, and “the” are intended to include the plural forms as well, unless the context clearly indicates otherwise. It should be further understood that the terms “comprises” and/or “comprising,” when used in this specification, specify the presence of stated features, integers, steps, operations, elements, components or a combination thereof, but do not preclude the presence or addition of one or more other features, integers, steps, operations, elements, components, and/or groups thereof.


Unless otherwise defined, all terms, including technical and scientific terms, used herein have the same meaning as commonly understood by one of ordinary skill in the art to which this disclosure pertains. It will be further understood that terms, such as those defined in commonly-used dictionaries, should be interpreted as having a meaning that is consistent with their meaning in the context of the relevant art and will not be interpreted in an idealized or overly formal sense unless expressly so defined herein.


The example embodiments may be implemented as various types of products, such as, for example, a personal computer (PC), a laptop computer, a tablet computer, a smartphone, a television (TV), a smart home appliance, an intelligent vehicle, a kiosk, and a wearable device. Hereinafter, example embodiments will be described in detail with reference to the accompanying drawings. In the drawings, like reference numerals are used for like elements.



FIG. 1 is a diagram illustrating a method of predicting and controlling time series data based on automatic learning according to an example embodiment.


Since analysis of time series data such as stocks and exchange rates involves complex factors, it is difficult to pinpoint which factors have an effect on the analysis. Recent advances in artificial intelligence technology have shown superior prediction performance compared to traditional statistical analysis in forecasting. Specifically, a deep learning technology may accurately recognize various behavior patterns of a user appearing in an image and signal big data to be comparable to humans. An artificial intelligence model based on the deep learning technology may make more accurate predictions than humans by recognizing patterns of a personalized user with a cognitive function comparable to that of humans. However, for accurate learning, it is necessary to compare accuracy with numerous artificial intelligence (e.g., machine learning and deep learning) models.


According to an example embodiment, a device for predicting and controlling time series data based on automatic learning (hereinafter referred to as a control and prediction device) 100 may provide prediction, automatic control, and description services based on federated learning that is a combination of advantages of various artificial intelligence technologies.


Referring to FIG. 1, a control and prediction device 250 may receive input time series data 110 and output prediction time series data 120 that corresponds to the input time series data 110. Furthermore, the control and prediction device 250 may output a prediction description 130 that includes at least one of a time series prediction result, a reason for the prediction, and an optimal control suggestion together with the prediction time series data 120.


For example, the control and prediction device 250 may output the prediction description 130 that includes the time series prediction result saying “Data is likely to decrease linearly from now on.” and the optimal control suggestion saying “To adjust the data to an appropriate level, lowering an input A by X % is recommended.” together with the prediction time series data 120.



FIG. 2 is a diagram illustrating a relationship between a training device and a prediction device according to an example embodiment.


Referring to FIG. 2, a training device 200 may correspond to a computing device having various processing functions, for example, functions of generating a neural network, training (or learning) a neural network, or retraining a neural network. For example, the training device 200 may be implemented as various types of devices, for example, a personal computer (PC), a server device, or a mobile device.


The training device 200 may generate a trained neural network 210 by repetitively training (or learning) a given initial neural network. Generating of the trained neural network 210 may refer to determining neural network parameters. Here, the parameters may include various types of data, for example, input/output activations, weights, and biases that are input to and output from the neural network. When the neural network is repetitively trained, the parameters of the neural network may be tuned to calculate a more accurate output for a given input.


The training device 250 may transmit the trained neural network 210 to a prediction device 250. The prediction device 250 may be included in a mobile device or an embedded device. The prediction device 250 may be dedicated hardware for operating the neural network.


The prediction device 250 may operate the trained neural network 210 without a change, or may operate a neural network 260 obtained by processing (for example, quantizing) the trained neural network 210. The prediction device 250 for operating the processed neural network 160 may be implemented in a separate device independent of the generative model training device 200. However, example embodiments are not limited thereto, and the prediction device 250 and the training device 200 may also be implemented in a same device. Hereinafter, a device including both the prediction device 250 and the training device 200 will be referred to as an artificial intelligence device.



FIG. 3 is a diagram illustrating a method of predicting and controlling time series data based on automatic learning according to an example embodiment.


Referring to FIG. 3, operations 310 to 330 may be performed by the training device described above with reference to FIG. 2. The training device may be implemented by one or more hardware modules, one or more software modules, or various combinations thereof. Furthermore, the operations of FIG. 3 may be performed in a shown order and manner. However, the order of some operations may be changed, or some operations may be omitted, without departing from the spirit and scope of the example embodiment shown in FIG. 3. The operations in FIG. 3 may be performed in parallel or simultaneously.


In operation 310, the training device trains a plurality of time series data prediction models according to conditions for the respective models. The training device may train the models a predetermined number of times (e.g., three times) under the different conditions for the respective models.


In operation 320, the training device determines, among trained time series data prediction models, one or more optimal models that meet a predetermined condition. For example, the training device may determine, among the plurality of time series data prediction models, top three models of which prediction performance is good and does not significantly change according to a model change as optimal models.


In operation 330, the training device generates a final model by combining the one or more optimal models.


When given prediction performance decreases or a specific interval passes after training, operations 310 to 330 may be automatically repeated. For example, the training device may evaluate prediction performance of the final model, and when the prediction performance of the final model decreases below a predetermined threshold, the training device may repeat operations 310 to 330 to update the final model. Alternatively, the training device may update the final model according to a predetermined interval.



FIG. 4 is a diagram illustrating a training method according to an example embodiment.


Referring to FIG. 4, a training device may train a plurality of time series data prediction models according to conditions for the respective models.


The plurality of time series data prediction models may include at least one of statistical-based prediction models and deep learning-based prediction models. For example, the statistical-based prediction models may include LASSO, ARIMA, XGBoost, and the like, and the deep learning-based prediction models may include FCN/CNN, LSTM, LSTM-CNN, STGCN, DARNN, DSANet, and the like. However, the above-described plurality of time series data prediction models is exemplary to merely describe the example embodiments according to technical concepts, and the plurality of time series data prediction models may include various different models and are not limited to the examples described in the present specification.


The training device may train each of the plurality of time series data prediction models three times and may determine an optimal model among the trained models. For example, the training device may determine M2XGB, which is a second trained model among XGBoost models, as a first optimal model, M3LSTM-CNN, which is a third trained model among LSTM-CNN models, as a second optimal model, and M1DARNN, which is a first trained model among DARNN models, as a third optimal model.


Furthermore, the training device may generate a final model (e.g., ModelMIX) by combining M2XGB, M3LSTM-CNN, and M1DARNN. When the final model is generated, a prediction device may use the final model that is optimally trained and stored to provide a predicted value for input data in real time without retraining. A detailed method of predicting and controlling time series data will be described below with reference to FIGS. 5A to 5B.



FIG. 5A is a diagram illustrating a method of predicting and controlling time series data according to an example embodiment.


Referring to FIG. 5A, a prediction device may receive a final model from a training device. However, as described above, the prediction device and the training device may be implemented in a same device.


The prediction device may receive target variable data for predicting the time series data. Furthermore, the prediction device may input the target variable data to the final model and output target variable prediction data that corresponds to the target variable data. The target variable data may be input time series data to be predicted, and the prediction target variable data may be predicted data that corresponds to the target variable data. The target variable data may be, for example, process yield data or return on investment data over time.


The prediction device may receive control variable data that determines a direction of a change in the target variable prediction data. Furthermore, the prediction device may input the control variable data to the final model and output control variable prediction data that corresponds to the control variable data. The control variable data may be data that determines a direction of a change in the target variable prediction data. For example, if target data is return on investment data, the control variable data may be international oil price data or exchange rate data that may affect the return on investment data.


The prediction device may provide a prediction result and a control method of the time series data based on the target variable prediction data and the control variable prediction data. The prediction device may adjust the control variable data based on a correlation between the target variable prediction data and the control variable prediction data. As an example, the prediction device may train a reinforcement learning model according to a reward function that is determined based on the target variable prediction data and the control variable prediction data.


For example, the prediction device may adjust a control variable based on an optimal prediction model to determine the direction of the change in the target variable data such as process yield and production improvement, a return on investment increase, and a stability increase, and learn control variable data that optimizes the same through the reinforcement learning. Furthermore, the prediction device may provide a user with an adjustment direction of the optimized control variable. With reference to FIG. 5B below, a guide method will be described in more detail.


Referring to FIG. 5B, the prediction device may guide the user on an optimal value search direction and an optimal value search time of control variables (e.g., process parameters or process independent variables). That is, variables of an optimization function may be an optimization target parameter Y, process parameters (x1, x2, . . . , xn), and the optimal value search time (or a number of searches).


The prediction device may further include a black box model as well as the final model. The black box model may be a model that outputs a result value corresponding to independent variables. The final model may be used to predict the result value, and the black box model may be used to determine an optimal search time and an optimal value based on an optimal value search result predicted by the final model.


More particularly, the prediction device may have to select a moving direction (e.g., increasing or decreasing) from the origin (e.g., an initial x) of process independent variables to search a process optimal value. For example, the prediction device may determine the moving direction according to a correlation (e.g., a gradient) between the independent variables and a dependent variable y that the final model (e.g., an interpretable model) learned. For example, the prediction device may search the process independent variables one time (xt=xt−1+dx) in an arbitrary direction from the origin and then, store a corresponding search value and perform a next optimal value search when a response (yt=f(xt)) of the interpretable model is improved (yt>yt−1), or dismiss the corresponding search value and return the independent variables to a previous position (xt−1) when the response is not improved.


The prediction device may input the stored optimal value search result (xt) to the black box model when a preset search end time is reached. The prediction device may find a search time (a number of searches) at which a response of the black box model is optimal (argmax f(x)) to determine the corresponding time as the optimal search time and determine the optimal value search result at the optimal search time as a guide (an optimal x). Thus, this is a system capable of providing an accurate process guide and a description thereof.



FIG. 6 is a block diagram of an artificial intelligence device according to an example embodiment.


Referring to FIG. 6, an artificial intelligence device includes a processor 610. The artificial intelligence device 600 may further include a communication interface 630 and a memory 620. The processor 610, the memory 620, and the communication interface may communicate with each other via a communication bus.


The processor 610 may train a plurality of time series data prediction models according to conditions for the respective models, determine, among the trained time series data prediction models, one or more optimal models that meet a predetermined condition, and generate a final model by combining the one or more optimal models.


The memory 620 may store a variety of information generated during the processing process of the processor 610. In addition, the memory 620 may store various data and programs. The memory 620 may include a volatile memory or a non-volatile memory. The memory 620 may include a large-capacity storage medium such as a hard disk to store the various data.


In addition, the processor 610 may perform the at least one method described above with reference to FIGS. 1 to 5 or an algorithm corresponding to the at least one method. The processor 610 may execute a program and control the artificial intelligence device 600. A program code to be executed by the processor 610 may be stored in the memory 620. The artificial intelligence device 600 may be connected to an external device (e.g., a PC or a network) through an input/output device (not shown) to exchange data therewith.


The examples described herein may be implemented using a hardware component, a software component and/or a combination thereof. A processing device may be implemented using one or more general-purpose or special-purpose computers, such as, for example, a processor, a controller and an arithmetic logic unit (ALU), a digital signal processor (DSP), a microcomputer, a field programmable gate array (FPGA), a programmable logic unit (PLU), a microprocessor or any other device capable of responding to and executing instructions in a defined manner. The processing device may run an operating system (OS) and one or more software applications that run on the OS. The processing device also may access, store, manipulate, process, and create data in response to execution of the software. For purpose of simplicity, the description of a processing device is used as singular; however, one skilled in the art will appreciate that a processing device may include multiple processing elements and multiple types of processing elements. For example, a processing device may include multiple processors or a processor and a controller. In addition, different processing configurations are possible, such as parallel processors.


The software may include a computer program, a piece of code, an instruction, or some combination thereof, to independently or uniformly instruct or configure the processing device to operate as desired. Software and data may be embodied permanently or temporarily in any type of machine, component, physical or virtual equipment, computer storage medium or device, or in a propagated signal wave capable of providing instructions or data to or being interpreted by the processing device. The software also may be distributed over network-coupled computer systems so that the software is stored and executed in a distributed fashion. The software and data may be stored by one or more non-transitory computer-readable recording mediums.


The methods according to the above-described example embodiments may be recorded in non-transitory computer-readable media including program instructions to implement various operations of the above-described example embodiments. The media may also include, alone or in combination with the program instructions, data files, data structures, and the like. The program instructions recorded on the media may be those specially designed and constructed for the purposes of example embodiments, or they may be of the kind well-known and available to those having skill in the computer software arts. Examples of non-transitory computer-readable media include magnetic media such as hard disks, floppy disks, and magnetic tape; optical media such as CD-ROM discs or DVDs; magneto-optical media such as optical discs; and hardware devices that are specially configured to store and perform program instructions, such as read-only memory (ROM), random access memory (RAM), flash memory, and the like. Examples of program instructions include both machine code, such as produced by a compiler, and files containing higher-level code that may be executed by the computer using an interpreter.


While this disclosure includes specific examples, it will be apparent to one of ordinary skill in the art that various changes in form and details may be made in these examples without departing from the spirit and scope of the claims and their equivalents. The examples described herein are to be considered in a descriptive sense only, and not for purposes of limitation. Descriptions of features or aspects in each example are to be considered as being applicable to similar features or aspects in other examples. Suitable results may be achieved if the described techniques are performed in a different order, and/or if components in a described system, architecture, device, or circuit are combined in a different manner, and/or replaced or supplemented by other components or their equivalents.


Therefore, the scope of the disclosure is defined not by the detailed description, but by the claims and their equivalents, and all variations within the scope of the claims and their equivalents are to be construed as being included in the disclosure.

Claims
  • 1. A method of predicting, controlling, and describing time series data based on automatic learning, the method comprising: training a plurality of time series data prediction models according to conditions for the respective models;determining, among the trained time series data prediction models, one or more optimal models that meet a predetermined condition; andgenerating a final model by combining the one or more optimal models,wherein the plurality of time series data prediction models comprises at least one of statistical-based prediction models and deep learning-based prediction models.
  • 2. The method of claim 1, further comprising: receiving target variable data for predicting time series data;inputting the target variable data to the final model and outputting target variable prediction data that corresponds to the target variable data.
  • 3. The method of claim 2, further comprising: receiving control variable data that determines a direction of a change in the target variable prediction data;inputting the control variable data to the final model and outputting control variable prediction data that corresponds to the control variable data.
  • 4. The method of claim 3, further comprising: providing a prediction result and a control method of the time series data based on the target variable prediction data and the control variable prediction data.
  • 5. The method of claim 3, further comprising: adjusting the control variable data based on a correlation between the target variable prediction data and the control variable prediction data.
  • 6. The method of claim 5, wherein the adjusting of the control variable data comprises training a reinforcement learning model according to a reward function that is determined based on the target variable prediction data and the control variable prediction data.
  • 7. The method of claim 3, wherein the outputting of the control variable prediction data comprises: determining a moving direction of the control variable data; anddetermining an optimal search time for the control variable data.
  • 8. The method of claim 7, wherein the outputting of the target variable prediction data comprises outputting the target variable prediction data based on the moving direction and the optimal search time for the control variable data.
  • 9. The method of claim 1, wherein the training comprises training the plurality of time series data prediction models a predetermined number of times according to the conditions for the respective models.
  • 10. The method of claim 1, further comprising: evaluating prediction performance of the final model; andupdating the final model, when the prediction performance of the final model decreases below a predetermined threshold.
  • 11. The method of claim 1, further comprising: updating the final model according to a predetermined interval.
  • 12. A computer program stored in a medium to perform the method of claim 1 in combination with hardware.
  • 13. A device for predicting, controlling, and describing time series data based on automatic learning, the device comprising: a processor configured to train a plurality of time series data prediction models according to conditions for the respective models, determine, among the trained time series prediction models, one or more optimal models that meet a predetermined condition, and generate a final model by combining the one or more optimal models,wherein the plurality of time series data prediction models comprises at least one of statistical-based prediction models and deep learning-based prediction models.
  • 14. The device of claim 13, wherein the processor is further configured to: receive target variable data for predicting time series data, input the target variable data to the final model and output target variable prediction data that corresponds to the target variable data.
  • 15. The device of claim 14, wherein the processor is further configured to: receive control variable data that determines a direction of a change in the target variable prediction data, input the control variable data to the final model and output control variable prediction data that corresponds to the control variable data.
Priority Claims (1)
Number Date Country Kind
10-2020-0185683 Dec 2020 KR national
PCT Information
Filing Document Filing Date Country Kind
PCT/KR2021/020067 12/28/2021 WO