This application claims the benefit of Korean Patent Application No. 10-2022-0094553 filed on Jul. 29, 2022, in the Korean Intellectual Property Office, the disclosure of which is incorporated herein in its entirety by reference.
The present disclosure relates to an embedding transformation method and system, and more particularly, to a method of efficiently performing various deep learning tasks using embedding transformation and a system for performing the method.
Machine translation is a task in the field of natural language processing and is one of the tasks for which deep learning technology is showing the greatest achievements. For example, neural machine translation (NMT) models (e.g., seq2seq) based on neural networks are known to significantly outperform statistical machine translation (SMT) models in English-French or English-German translations.
However, deep learning models that perform machine translation tasks usually have a very complex structure and a fairly large scale (i.e., have a very large number of weight parameters). Therefore, they require much higher learning costs than other deep learning models (e.g., models that perform classification tasks). In addition, since the deep learning models must learn a large parallel corpus in order to outperform the SMT models, enormous time and computing costs are required to build a deep learning model with a known level of translation performance.
Aspects of the present disclosure provide a method of efficiently performing various deep learning tasks (e.g., tasks related to machine translation, domain transformation, multimodal, etc.) through embedding transformation and a system for performing the method.
Aspects of the present disclosure also provide a method of accurately performing embedding transformation between a source side and a target side and a system for performing the method.
Aspects of the present disclosure also provide a method of training a deep learning model that can perform embedding transformation and a system for performing the method.
Aspects of the present disclosure also provide a method of reducing the computing cost required to train a deep learning model that can perform embedding transformation.
However, aspects of the present disclosure are not restricted to the one set forth herein. The above and other aspects of the present disclosure will become more apparent to one of ordinary skill in the art to which the present disclosure pertains by referencing the detailed description of the present disclosure given below.
According to an aspect of the inventive concept, there is provided an embedding transformation method performed by at least one computing device. The method may include obtaining a source-side embedding model, transforming source-side data into a first embedding vector through the source-side embedding model, and transforming the first embedding vector into a second embedding vector located in a target-side embedding space through a transformation model.
In some embodiments, the transformation model may include an implicit layer, and the implicit layer is configured to repeatedly perform a layer operation based on a value of a weight parameter of the implicit layer until a preset condition is satisfied.
In some embodiments, the transformation model may include an attention layer.
In some embodiments, the transformation model may be trained using the source-side embedding model and a target-side embedding model, and the source-side embedding model and the target-side embedding model are pretrained models.
In some embodiments, a training dataset for the transformation model may include source-side training data, target-side training data and target-side type information, and the type information is information for distinguishing a plurality of target-side training data corresponding to the source-side training data.
In some embodiments, the method may further include transforming target-side data into a third embedding vector through a target-side embedding model and training the transformation model based on a difference between the second embedding vector and the third embedding vector wherein the target-side data corresponds to the source-side data.
In some embodiments, the source-side embedding model and the target-side embedding model may be pretrained models, and the training of the transformation model may include updating a weight parameter of the transformation model in a state where the source-side embedding model and the target-side embedding model are frozen.
In some embodiments, the transformation model may be configured to receive the first embedding vector and the third embedding vector and output the second embedding vector.
In some embodiments, the method may further include decoding the second embedding vector through a target-side decoder and training at least one of the target-side decoder and the transformation model based on a difference between a result of the decoding and target-side data, wherein the target-side data corresponds to the source-side data.
In some embodiments, the method may further include decoding the second embedding vector through a target-side decoder, wherein the target-side decoder is trained through transforming target-side data into an embedding vector through a target-side embedding model, decoding the embedding vector through the target-side decoder, and updating a weight parameter of the target-side decoder based on a difference between a result of the decoding and the target-side data.
In some embodiments, the source-side data may be text in a source language, and the method may further include translating the text in the source language into text in a target language by decoding the second embedding vector through a target-side decoder.
In some embodiments, the source language may include a first language and a second language, and the source-side embedding model is configured to transform text in the first language and the second language into an embedding vector located in a shared embedding space.
In some embodiments, the source-side data may be data of a source modal, and the second embedding vector may be an embedding vector of data of a target modal corresponding to the data of the source modal.
In some embodiments, the source-side embedding model may include an embedding model of a first source and an embedding model of a second source, the first embedding vector is the source-side data transformed by the embedding model of the first source, and the transforming of the first embedding vector into the second embedding vector may include obtaining the second embedding vector by inputting the first embedding vector and source information indicating the first source to the transformation model.
In some embodiments, the target-side embedding space may include an embedding space of a first target and an embedding space of a second target, the second embedding vector is located in the embedding space of the first target, and the transforming of the first embedding vector into the second embedding vector may include obtaining the second embedding vector by inputting the first embedding vector and target information indicating the first target to the transformation model.
According to another aspect of the inventive concept, there is provided an embedding transformation system. The embedding transformation system may include one or more processors and a memory configured to store one or more instructions, wherein the one or more processors are configured to execute the stored one or more instructions to obtain a source-side embedding model, transform source-side data into a first embedding vector through the source-side embedding model, and transform the first embedding vector into a second embedding vector located in a target-side embedding space through a transformation model.
In some embodiments, the transformation model may be trained using the source-side embedding model and a target-side embedding model, and the source-side embedding model and the target-side embedding model may be pretrained models.
In some embodiments, the one or more processors may be further configured to transform target-side data into a third embedding vector through a target-side embedding model and train the transformation model based on a difference between the second embedding vector and the third embedding vector, wherein the target-side data corresponds to the source-side data.
In some embodiments, the source-side data may be text in a source language, and the one or more processors may be further configured to translate the text in the source language into text in a target language by decoding the second embedding vector through a target-side decoder.
According to still another aspect of the inventive concept, there is provided a non-transitory computer-readable recording medium storing computer program executable by at least one processor to execute obtaining a source-side embedding model, transforming source-side data into a first embedding vector through the source-side embedding model, and transforming the first embedding vector into a second embedding vector located in a target-side embedding space through a transformation model.
These and/or other aspects will become apparent and more readily appreciated from the following description of the embodiments, taken in conjunction with the accompanying drawings in which:
Hereinafter, example embodiments of the present disclosure will be described with reference to the attached drawings. Advantages and features of the present disclosure and methods of accomplishing the same may be understood more readily by reference to the following detailed description of example embodiments and the accompanying drawings. The present disclosure may, however, be embodied in many different forms and should not be construed as being limited to the embodiments set forth herein. Rather, these embodiments are provided so that this disclosure will be thorough and complete and will fully convey the concept of the disclosure to those skilled in the art, and the present disclosure will be defined by the appended claims and their equivalents.
In adding reference numerals to the components of each drawing, it should be noted that the same reference numerals are assigned to the same components as much as possible even though they are shown in different drawings. In addition, in describing the present disclosure, when it is determined that the detailed description of the related well-known configuration or function may obscure the gist of the present disclosure, the detailed description thereof will be omitted.
Unless otherwise defined, all terms used in the present specification (including technical and scientific terms) may be used in a sense that can be commonly understood by those skilled in the art. In addition, the terms defined in the commonly used dictionaries are not ideally or excessively interpreted unless they are specifically defined clearly. The terminology used herein is for the purpose of describing particular embodiments only and is not intended to be limiting of the disclosure. In this specification, the singular also includes the plural unless specifically stated otherwise in the phrase.
In addition, in describing the component of this disclosure, terms, such as first, second, A, B, (a), (b), can be used. These terms are only for distinguishing the components from other components, and the nature or order of the components is not limited by the terms. If a component is described as being “connected,” “coupled” or “contacted” to another component, that component may be directly connected to or contacted with that other component, but it should be understood that another component also may be “connected,” “coupled” or “contacted” between each component.
Hereinafter, embodiments of the present disclosure will be described with reference to the attached drawings.
As illustrated in
For reference, since an embedding vector is a vector representation in an embedding space, it may be used interchangeably with the term ‘embedding representation’ in some cases. In addition, in the art to which the present disclosure pertains, the embedding vector may be used interchangeably with terms such as ‘latent vector’, ‘latent code’, ‘embedding code’, and ‘latent representation’. For ease of description, the embedding transformation system 10 will hereinafter be abbreviated to a ‘transformation system 10’.
More specifically, the transformation system 10 may train a transformation model 11 using a dataset (i.e., a paired dataset) composed of pairs of source-side data and target-side data and may transform an embedding vector 12 of the source side into an embedding vector 13 of the target side using the trained transformation model 11. The detailed structure and training method of the transformation model 11 will be described in detail with reference to
The transformation system 10 may perform a target task using the trained transformation model 11. Examples of the target task may include a machine translation task, a multimodal (cross-modal) task (e.g., image captioning, visual question answering, text-to-image search, or image-to-text search), a task related to domain transformation (e.g., image domain transformation), or a combination thereof. However, the scope of the present disclosure is not limited thereto, and examples of the target task may include various types of deep learning tasks without limitation.
In a specific example, the transformation system 10 may perform a machine translation task between two languages. For example, as illustrated in
For reference, since an embedding model is a module that encodes input data into an embedding vector, it may be named ‘encoder’ in some cases (see
In another example, the transformation system 10 may perform a multimodal (cross-modal) task. For example, as illustrated in
For reference, in
In addition, in
A specific method of performing embedding transformation using the transformation system 10 will be described in more detail with reference to
The transformation system 10 may be implemented in at least one computing device. For example, all functions of the transformation system 10 may be implemented in one computing device, or a first function of the transformation system 10 may be implemented in a first computing device, and a second function may be implemented in a second computing device. Alternatively, a certain function of the transformation system may be implemented in a plurality of computing devices.
A computing device may be any device having a computing function, and an example of this device is illustrated in
Until now, the transformation system 10 according to the embodiments of the present disclosure has been roughly described with reference to
For ease of understanding, the description will be continued based on the assumption that all steps/operations of the methods to be described later are performed by the above-described transformation system 10. Therefore, when the subject of a specific step/operation is omitted, it may be understood that the step/operation is performed by the transformation system 10. However, in a real environment, some steps of the methods to be described later may also be performed by another computing device. For example, training a transformation model (e.g., 11 in
First, an embedding transformation method according to embodiments of the present disclosure will be described with reference to
As illustrated in
The obtained embedding models (i.e., the embedding models of the source side and/or the target side) may have various structures according to the type of data to be embedded. For example, a model for embedding text may be composed of a recurrent neural network (RNN) or a transformer-based neural network (e.g., BERT). For another example, a model for embedding an image may be composed of a convolutional neural network (CNN)-based neural network. However, the scope of the present disclosure is not limited by these examples, and an embedding model can have any structure as long as it can appropriately transform input data into an embedding vector.
For reference, a text embedding model may be configured to output embedding vectors on a token-by-token basis or may be configured to compress input text into a single embedding vector and output the single embedding vector. For example, as illustrated in
In addition, a transformation model to be described later may also output embedding vectors on a token-by-token basis or output a single embedding vector. For example, the transformation model may receive source-side embedding vectors on a token-by-token basis and output target-side embedding vectors on a token-by-token basis. Alternatively, the transformation model may receive a single source-side embedding vector and output a single target-side embedding vector. Alternatively, the transformation model may receive source-side embedding vectors on a token-by-token basis and output a single target-side embedding vector.
In some embodiments, the obtained embedding models (i.e., the embedding models of the source side and/or the target side) may be pretrained models. For example, the embedding model of the source side may be a model trained using a first language training dataset (i.e., corpus), and the embedding model of the target side may be a model trained using a second language training dataset. In this case, since a training cost required for an embedding model can be reduced, the overall time and computing costs required to train a transformation model (or to perform a task) can be greatly reduced.
Referring back to
A specific method of preprocessing data of the source side and the target side may vary according to the type of the data and the structure of the embedding model. For example, when the type of the data of the source side and the target side is text, the transformation system 10 may perform preprocessing such as dividing a given text into sentences and tokenizing the sentences. However, the scope of the present disclosure is not limited thereto.
At least some of the source-side data and the target-side data may have a one-to-many (or many-to-many) correspondence. In this case, if a transformation model is trained by paring the same source-side data with different target-side data, confusion may occur in training the transformation model (so-called ‘one-to-many problem’). To solve this problem, in some embodiments of the present disclosure, type information for distinguishing multiple targets may be added to a training dataset. If the source-side data and the target-side data have a many-to-one correspondence, type information for distinguishing multiple sources may be added to the training dataset. For better understanding, the current embodiments will be further described with reference to
As illustrated in
For reference, the meaning of type may vary according to which characteristic (criteria) is used to classify (distinguish) multiple targets and may be determined by a target task. For example, it is assumed that the target task is a machine translation task that provides differential translation quality. In this case, texts (e.g., 92 and 93) in a target language which correspond to a text (e.g., 91) in a source language may be classified according to translation quality (e.g., high quality, medium quality, and low quality), and the type may mean translation quality. In addition, the transformation system 10 may provide a translation service with differential quality by transforming an embedding vector of the source language into an embedding vector (i.e., an embedding vector of the target language) of a quality indicated by the type information through a trained transformation model. In another example, it is assumed that the target task is a machine translation task that provides a differential translation length (or quality). In this case, texts (e.g., 92 and 93) in the target language which correspond to a text (e.g., 91) in the source language may be classified according to text length (e.g., long sentence, short sentence, etc.), and the type may mean text length (or quality). In another example, it is assumed that the target task is a machine translation task that considers a text domain (e.g., an economic, social, or technological field). In this case, texts (e.g., 92 and 93) in the target language which correspond to a text (e.g., 91) in the source language may be classified according to domain, and the type may mean domain. In addition, the transformation system 10 may provide a high-quality translation service that considers a text domain by transforming an embedding vector of the source language into an embedding vector of a domain indicated by the type information through a trained transformation model.
Referring back to
In some embodiments, the transformation model may be trained based on a difference between an embedding vector transformed by the transformation model and an embedding vector transformed by the embedding model of the target side. The current embodiments will be described in more detail later with reference to
In some embodiments, an embedding vector transformed by the transformation model may be decoded through a decoder of the target side, and the transformation model may be trained based on a difference between the decoding result and correct answer data (i.e., target-side data). The current embodiments will be described in more detail later with reference to
In some embodiments, the transformation model may be trained based on various combinations of the above embodiments.
The detailed structure of the transformation model may also vary according to embodiments.
In some embodiments, the transformation model may be implemented (configured) based on an implicit layer. For example, as illustrated in
For reference, as illustrated in
In some embodiments, the transformation model may be implemented (configured) based on an explicit layer. For example, as illustrated in
In some embodiments, the transformation model may be configured to further include an attention layer. For example, as illustrated in
In some embodiments, the transformation model may be implemented (configured) based on various combinations of the above embodiments. For example, as illustrated in
Referring back to
As illustrated in
Until now, the embedding transformation method according to the embodiments of the present disclosure has been described with reference to
In addition, since a pretrained embedding model is used, the time and computing costs required to build a model for performing various target tasks can be significantly reduced. That is, since the cost required to train an embedding model is reduced, the overall time and computing costs required to build a model for performing target tasks can be significantly reduced.
Embodiments related to a method of training a transformation model will now be described with reference to
First, a method of training a transformation model according to embodiments of the present disclosure will be described with reference to
As illustrated in
In operation S142, the first embedding vector (i.e., the source-side embedding vector) may be transformed into a second embedding vector (i.e., a target-side embedding vector) through the transformation model. That is, the transformation model may receive the first embedding vector, perform an appropriate operation on the first embedding vector, and output the second embedding vector obtained as a result of the operation.
In operation S143, target-side data may be transformed into a third embedding vector through a target-side embedding model.
In operation S144, a weight parameter of the transformation model may be updated based on a difference between the second embedding vector and the third embedding vector. For example, the transformation system 10 may calculate the difference (e.g., loss) between the second embedding vector and the third embedding vector (e.g., through cosine similarity) and may update the weight parameter of the transformation model in a direction to reduce the calculated difference. The transformation system 10 may update only the transformation model in a state where the source-side embedding model and the target-side embedding model are frozen or may update the transformation model together with at least one embedding model.
Operations S141 through S144 described above may be repeatedly performed for other data pairs included in the training dataset. In so doing, the transformation model can have the ability to transform (map) a source-side embedding vector into a target-side embedding vector.
For better understanding, the above-described training method will be further described with reference to the example illustrated in
As illustrated in
Then, the transformation system 10 may transform the first embedding vector 152 into a second embedding vector 154 through a transformation model 153. Here, the transformation system 10 may also obtain the second embedding vector 154 by further inputting a third embedding vector 155 to the transformation model 153. The third embedding vector 155 may be an embedding vector of an English text 159 transformed through an English (target-side) embedding model 156 and may be a target-side embedding vector corresponding to a correct answer.
Next, the transformation system 10 may update a weight parameter of the transformation model 153 based on a difference (e.g., loss 157) between the second embedding vector 154 and the third embedding vector 155.
As these processes are repeated, the transformation model 153 can have the ability to accurately transform (map) a Korean (source-side) embedding vector into an English (target-side) embedding vector.
A method of training a transformation model according to embodiments of the present disclosure will now be described with reference to
As illustrated in
In operation S162, the first embedding vector may be transformed into a second embedding vector through the transformation model.
In operation S163, the second embedding vector may be decoded through a target-side decoder. Here, the target-side decoder may be a neural network configured to decode an input embedding vector and output corresponding target-side data.
In operation S164, a weight parameter of the transformation model may be updated based on a difference between the decoding result and target-side data (i.e., target-side data or correct answer data in a data pair). For example, the transformation system 10 may calculate the difference between the decoding result and the correct answer data and update the weight parameter of the transformation model in a direction to reduce the calculated difference. Here, the transformation system 10 may further update the target-side decoder. The source-side embedding model may be in a frozen state or may be updated together with the transformation model.
Operations S161 through S164 described above may be repeatedly performed for other data pairs included in a training dataset. In so doing, the transformation model can have the ability to transform (map) a source-side embedding vector into a target-side embedding vector.
For better understanding, the above-described training method will be further described with reference to the example illustrated in
As illustrated in
Then, the transformation system 10 may transform the first embedding vector 172 into a second embedding vector 174 through a transformation model 173.
Next, the transformation system 10 may decode the second embedding vector 174 through a target-side decoder 175.
Next, the transformation system 10 may update a weight parameter of the transformation model 173 based on a difference (e.g., loss 177) between the decoding result 176 (i.e., a predicted English text) and an English text 179 (i.e., a correct answer text) in the data pair. In some cases, the transformation system 10 may further update a weight parameter of the target-side decoder 175.
As these processes are repeated, the transformation model 173 can have the ability to accurately transform (map) a Korean (source-side) embedding vector into an English (target-side) embedding vector.
The target-side decoder 175 may also be trained independently from the transformation model 173. For example, the transformation system 10 may independently train the target-side decoder 175 as in the example illustrated in
As illustrated in
For reference, since the decoder 175 is a module specialized for the machine translation task, it may be changed to another module if the target task changes (e.g., see the searcher 46 in
Until now, embodiments of the method of training the transformation model have been described with reference to
Until now, for ease of understanding, the description has been made based on the assumption that the transformation model performs one-to-one embedding transformation (i.e., embedding transformation between one source and one target). However, the transformation model may also perform many-to-one, one-to-many, or many-to-many embedding transformation. Hereinafter, embodiments related to this will be described.
First, a many-to-one embedding transformation method according to embodiments of the present disclosure will be described with reference to
As illustrated in
Specifically, the transformation system 10 may train the transformation model 193 using two training datasets 191 and 192 while explicitly providing source information to the transformation model 193. The source information may be information indicating to which of a plurality of sources an embedding vector input to the transformation model 193 belongs. As illustrated, the transformation system 10 may input source information indicating the first language to the transformation model 193 and train the transformation model 193 using the first training dataset 191. In addition, source information indicating the second language may be input to the transformation model 193, and the transformation model 193 may be trained using the second training dataset 192. In so doing, the transformation model 193 can accurately learn a many-to-one embedding relationship. The method of training the transformation model 193 may be understood from the previous description.
When training is completed, as illustrated in
In other words, the transformation system 10 may transform the text 201 in the first language into an embedding vector 203 of the first language through an embedding model 202 of the first language and may accurately transform the embedding vector 203 into a target-side embedding vector 207 by inputting the embedding vector 203 to the transformation model 193 together with source information indicating the first language. Then, the embedding vector 207 may be decoded into the text 209 in the third language through a target-side decoder 208.
Similarly, the transformation system 10 may transform the text 206 in the second language into an embedding vector 204 of the second language through an embedding model 205 of the second language and may accurately transform the embedding vector 204 into the target-side embedding vector 207 by inputting the embedding vector 204 to the transformation model 193 together with source information indicating the second language. Then, the embedding vector 207 may be decoded into the text 209 in the third language through the target-side decoder 208.
When an embedding model is a model that supports multiple languages (i.e., a multilingual embedding model), source information may be omitted (or changed). For example, when the number of languages on the source side is two and the embedding model supports embedding for two source-side languages, the source information may be omitted (see the description of
In
As illustrated in
The above description can also be applied to other types of tasks without a substantial change in technical spirit. For example, if the target task is a multimodal task, a multimodal embedding model may play the same role as the multilingual embedding model 212 of
Until now, the many-to-one embedding transformation methods according to the embodiments of the present disclosure have been described with reference to
In
As illustrated in
Specifically, the transformation system 10 may train the transformation model 223 using two training datasets 221 and 222 while explicitly providing target information to the transformation model 223. The target information may be information indicating to which of a plurality of targets an embedding vector input to the transformation model 223 belongs. As illustrated, the transformation system 10 may input target information indicating the second language to the transformation model 223 and train the transformation model 223 using the first training dataset 221. In addition, target information indicating the third language may be input to the transformation model 223, and the transformation model 223 may be trained using the second training dataset 222. In so doing, the transformation model 223 can accurately learn a one-to-many embedding relationship. The method of training the transformation model 223 may be understood from the previous description.
When training is completed, as illustrated in
In other words, the transformation system 10 may transform the text 231 in the first language into an embedding vector 233 of the first language through an embedding model 232 of the first language (source side). In addition, the transformation system 10 may accurately transform the embedding vector 233 into an embedding vector 234 of a desired target (second language) (i.e., a vector located in an embedding space of the second language) by inputting the embedding vector 233 to the transformation model 223 together with target information indicating the second language. Then, the embedding vector 234 may be decoded into the text 236 in the second language through a decoder 235 of the target.
Similarly, the transformation system 10 may accurately transform the embedding vector 233 into an embedding vector 237 of a desired target (third language) (i.e., a vector located in an embedding space of the third language) by inputting the embedding vector 233 of the first language to the transformation model 223 together with target information indicating the third language. Then, the embedding vector 237 may be decoded into the text 239 in the third language through a decoder 238 of the target.
Until now, the many-to-one embedding transformation method according to the embodiments of the present disclosure has been described with reference to
Until now, the embodiments of the one-to-many, many-to-one, or many-to-many embedding transformation method have been described with reference to
Hereinafter, an example computing device 240 that can implement the transformation system 10 according to the embodiments of the present disclosure will be described with reference to
Referring to
The processors 241 may control the overall operation of each component of the computing device 240. The processors 241 may include at least one of a central processing unit (CPU), a micro-processor unit (MPU), a micro-controller unit (MCU), a graphic processing unit (GPU), and any form of processor well known in the art to which the present disclosure pertains. In addition, the processors 241 may perform an operation on at least one application or program for executing operations/methods according to embodiments of the present disclosure. The computing device 240 may include one or more processors.
Next, the memory 242 may store various data, commands and/or information. The memory 242 may load the program 246 from the storage 245 in order to execute operations/methods according to embodiments of the present disclosure. The memory 242 may be implemented as a volatile memory such as a random access memory (RAM), but the technical scope of the present disclosure is not limited thereto.
Next, the bus 243 may provide a communication function between the components of the computing device 240. The bus 243 may be implemented as various forms of buses such as an address bus, a data bus, and a control bus.
Next, the communication interface 244 may support wired and wireless Internet communication of the computing device 240. In addition, the communication interface 244 may support various communication methods other than Internet communication. To this end, the communication interface 244 may include a communication module well known in the art to which the present disclosure pertains.
Next, the storage 245 may non-temporarily store one or more programs 246. The storage 245 may include a nonvolatile memory such as a read only memory (ROM), an erasable programmable ROM (EPROM), an electrically erasable programmable ROM (EEPROM) or a flash memory, a hard disk, a removable disk, or any form of computer-readable recording medium well known in the art to which the present disclosure pertains.
Next, the computer program 246 may include one or more instructions for controlling the processors 241 to perform operations/methods according to various embodiments of the present disclosure when the computer program 246 is loaded into the memory 242. That is, the processors 241 may perform the operations/methods according to the various embodiments of the present disclosure by executing the loaded instructions.
For example, the computer program 246 may include instructions for performing an operation of obtaining a source-side embedding model, an operation of transforming source-side data into a first embedding vector through the source-side embedding model, and an operation of transforming the first embedding vector into a second embedding vector located in a target-side embedding space through a transformation model. In this case, the transformation system 10 according to the embodiments of the present disclosure may be implemented through the computing device 240.
Until now, an example computing device 240 that can implement the transformation system 10 according to the embodiments of the present disclosure has been described with reference to
So far, a variety of embodiments of the present disclosure and the effects according to embodiments thereof have been mentioned with reference to
The technical features of the present disclosure described so far may be embodied as computer readable codes on a computer readable medium. The computer readable medium may be, for example, a removable recording medium (CD, DVD, Blu-ray disc, USB storage device, removable hard disk) or a fixed recording medium (ROM, RAM, computer equipped hard disk). The computer program recorded on the computer readable medium may be transmitted to other computing device via a network such as internet and installed in the other computing device, thereby being used in the other computing device.
Although operations are shown in a specific order in the drawings, it should not be understood that desired results can be obtained when the operations must be performed in the specific order or sequential order or when all of the operations must be performed. In certain situations, multitasking and parallel processing may be advantageous. According to the above-described embodiments, it should not be understood that the separation of various configurations is necessarily required, and it should be understood that the described program components and systems may generally be integrated together into a single software product or be packaged into multiple software products.
In concluding the detailed description, those skilled in the art will appreciate that many variations and modifications can be made to the example embodiments without substantially departing from the principles of the present disclosure. Therefore, the disclosed example embodiments of the disclosure are used in a generic and descriptive sense only and not for purposes of limitation.
Number | Date | Country | Kind |
---|---|---|---|
10-2022-0094553 | Jul 2022 | KR | national |